Abstract
Background/Objectives: Predicting individual response to depression therapy prior to treatment initiation remains a critical clinical challenge, as the response rate to both selective serotonin reuptake inhibitors (SSRIs) and repetitive transcranial magnetic stimulation (rTMS) is approximately 50%, leaving treatment selection largely trial-based. This study presents a computer-aided decision (CAD) framework that predicts depression therapy outcomes from pre-treatment electroencephalogram (EEG) signals using advanced time-frequency representations and pretrained convolutional neural networks (CNNs). Methods: EEG signals from 30 SSRI patients and 46 rTMS patients are transformed into time-frequency images using Continuous Wavelet Transform (CWT), Variational Mode Decomposition (VMD), and their pixel-level fusion. Four pretrained CNN architectures, including ResNet-18, MobileNet-V3, EfficientNet-B0, and TinyViT-Hybrid, are fine-tuned and evaluated under both image-independent and subject-independent 6-fold cross-validation (CV). Results: Results reveal a clear therapy-specific pattern: CWT-based representations yield superior discrimination for SSRI outcome prediction, with ResNet-18 achieving 99.43% image-level accuracy, while VMD-based representations are statistically superior for rTMS outcome prediction, with ResNet-18 reaching 98.77%. Pixel-level fusion of CWT and VMD does not consistently improve performance over the best individual representation in either therapy context. Pairwise Wilcoxon signed-rank tests confirm a two-tier architectural hierarchy in which ResNet-18 and TinyViT-Hybrid significantly outperform MobileNet-V3 and EfficientNet-B0 across all conditions, while remaining statistically indistinguishable from each other. At the subject level, the framework achieves 82.50% and 83.53% accuracy for SSRI and rTMS, respectively, under strict subject-independent evaluation. Per-channel analysis reveals occipital dominance for SSRI under CWT and frontotemporal dominance for rTMS under VMD, consistent with known neurophysiological mechanisms. Conclusions: These findings demonstrate that the choice of time-frequency representation is therapy-specific and at least as important as architectural complexity, and that competitive performance can be achieved without recurrent or attention layers by combining well-designed spectral images with a simple pretrained residual network.
1. Introduction
Over the past two decades, the rate of depression disorder has been continuously increasing worldwide, particularly in developed countries [1]. Approximately 280 million people suffer from depression [2]. Depression is a serious mood disorder that causes various emotional and physical problems [3]. Patients with depression may engage in self-harm and, in some cases, suicide if they do not receive effective therapy. Indeed, suicide is the fourth most common cause of death among young people [4].
Psychiatrists typically initiate depression treatment by prescribing either selective serotonin reuptake inhibitors (SSRIs) or repetitive transcranial magnetic stimulation (rTMS) as the first course of treatment [5,6]. However, the response rate to depression therapy is approximately 50%, which means that doctors often prescribe therapies through trial and error [5]. The brain’s response to depression therapy varies from one patient to another due to the complex nature of the brain [7]. This means that a therapy that is effective for one patient may not be effective for another, and depression treatment should, therefore, be personalised. This represents a challenging task for medical teams. When a patient does not respond to one therapy, doctors switch to another; however, a failed treatment course wastes approximately two to ten weeks and increases the risk of self-harm and suicide.
Patients with depression who receive therapy are categorised into two groups: responders and non-responders. Patients who show improvement after depression therapy are categorised as responders, while those who do not respond to the therapy are categorised as non-responders [8,9,10].
Electroencephalography (EEG) is a low-cost, real-time, and widely accessible neuroimaging modality [11]. It is used in hospitals as well as local clinics to assess brain function [12]. Pre-treatment EEG recordings from patients with depression contain informative patterns that can differentiate responders from non-responders to various depression therapies [3]. However, clinicians cannot determine therapy outcomes through visual inspection of EEG signals, because EEG is nonlinear and non-stationary data with complex patterns, making it a difficult task for clinicians to decide on the most appropriate therapy to prescribe [6,13]. Developing EEG-based computer-aided decision (CAD) systems can, therefore, support clinicians in selecting the most appropriate treatment strategy.
EEG analysis has been widely investigated for predicting depression therapy outcomes. A review of prior studies shows that traditional machine learning approaches have produced limited performance. Although deep learning methods have reported improved results, most rely on conventional EEG image representations and focus on a single therapeutic modality rather than multiple depression treatments. Therefore, a robust and computationally efficient framework is still needed to predict depression therapy outcomes.
Table 1 summarises the most relevant prior studies on EEG-based prediction of depression therapy outcomes, including the methods employed, the therapy type, the number of patients, and the reported image-level classification accuracy.
Table 1.
A summary of previous works on predicting outcomes of depression therapy. All reported accuracy values correspond to image-level classification results.
To address these limitations, this study proposes generating informative time-frequency images from pre-treatment EEG signals using two advanced signal processing techniques and their fusion. Continuous Wavelet Transform (CWT) provides a multi-resolution time-frequency representation that captures the nonlinear and non-stationary nature of EEG signals [26]. CWT time-frequency representations capture oscillatory frequency patterns over time [27]. Variational Mode Decomposition (VMD), on the other hand, is an adaptive decomposition technique [28] that decomposes the EEG signal into a set of band-limited intrinsic mode functions (IMFs) with narrow frequency bands, which is more stable for oscillatory frequency components. In addition to analysing CWT and VMD-based images separately, we construct a fusion-based representation that combines information from both domains [29,30]. It has been shown that the fusion of time-frequency representations can yield richer information [29,30]. By transforming the outputs of CWT, VMD, and their fusion into structured time-frequency images, the information in EEG related to treatment response can be more effectively utilised.
In addition, we investigate the use of four well-known pretrained convolutional neural network (CNN) architectures: ResNet-18 [31], MobileNet-V3 [32,33], EfficientNet-B0 [34,35], and TinyViT-Hybrid [36,37,38] for classifying patients into responder (R) and non-responder (NR) groups based on the CWT, VMD, and fusion-derived EEG images. ResNet-18 is a lightweight residual model that facilitates deeper training through skip connections, while MobileNet-V3 uses depthwise convolutions and attention modules for high efficiency [31,32,33]. EfficientNet-B0 applies compound scaling to balance network width, depth, and resolution to achieve better performance [34,35]. TinyViT-Hybrid integrates residual convolutional blocks with a compact vision transformer to capture both local and global representations in a low-complexity model [36,37,38].
By comparing these architectures across three types of input representations (i.e., CWT, VMD, and their fusion), we obtain a comprehensive view of the trade-offs between accuracy and robustness for EEG-based prediction of depression therapy outcomes. The aim of this work is to develop and evaluate a robust EEG-based CAD system that integrates advanced time-frequency analysis with state-of-the-art pretrained CNN models to predict treatment response for SSRI and rTMS depression therapies.
The remainder of this paper is organised as follows. Section 2 describes the two EEG datasets used in this study. Section 3 covers the preprocessing steps applied to the EEG data. Section 4 presents the methodology, including the time-frequency representations, the pretrained CNN models, and the statistical evaluation metrics. Section 5 reports and discusses the experimental results. Section 6 presents the conclusion of the study.
2. Materials
The proposed CAD system is evaluated using two EEG datasets. This section details these datasets. Figure 1 presents representative EEG signals for R and NR patients from the SSRI and rTMS datasets.
Figure 1.
Example EEG signals from the Fp1 channel for R (left) and NR (right) patients to SSRI therapy (top row) and rTMS therapy (bottom row).
2.1. SSRI Dataset
The SSRI dataset from Universiti Teknologi PETRONAS was used to evaluate the models in predicting treatment response [14]. It contains five-minute pre-treatment EEG recordings from 30 patients diagnosed with depression. SSRI therapy was prescribed for all patients in this cohort. EEG signals were recorded in an eyes-closed resting-state condition using the 10–20 international electrode placement system. The sampling frequency was 256 Hz and the EEG data comprised 19 channels.
Depression severity was assessed before and after a six-week therapy course using the Beck Depression Inventory (BDI). Patients showing more than a 50% reduction in their BDI scores were labelled as responders, while those with 50% or less improvement were labelled as non-responders. The Universiti Teknologi Malaysia Research Ethics Committee approved the data recording protocol [14]. The dataset contains 12 EEG recordings from R and 18 recordings from NR patients to SSRI therapy, provided in EDF format.
2.2. rTMS Dataset
The rTMS dataset was collected at Atieh Hospital, Tehran, Iran, and contains pre-treatment EEG recordings from 46 patients diagnosed with depression. Five-minute resting-state EEG signals were recorded for each patient with eyes closed, using the 10–20 international electrode placement system at a sampling frequency of 500 Hz and 19 channels.
All patients underwent a six-week rTMS therapy course. Depression severity was assessed before and after treatment using the BDI. Patients showing more than 50% improvement in BDI scores were labelled as responders, while those with 50% or less improvement were labelled as non-responders. The dataset contains EEG recordings from 23 responders and 23 non-responders.
The Ethics Committee of Shahid Beheshti University of Medical Sciences approved the data collection protocol [7], and the research team obtained written informed consent from all participants prior to data collection [7,23,24].
3. Preprocessing
EEG signals were band-pass filtered within 0.5–70 Hz to attenuate slow baseline fluctuations and high-frequency muscle contamination while preserving physiologically meaningful activity for brain rhythms. A 50 Hz notch filter was used to suppress power-line noise. Motion-related artifacts were mitigated using the EEGLAB toolbox version 4.5b [39]. All filtering operations were performed using zero-phase forward and reverse filtering via MATLAB R-2023a’s filtfilt function to prevent phase distortion. The processed EEG recordings were subsequently re-referenced to the common average reference (CAR) to reduce inter-channel reference bias. Finally, each subject’s data were divided into non-overlapping 15-s segments.
The continuous recordings were segmented into non-overlapping 15-s epochs. This window length was selected to achieve a balance between spectral resolution and statistical robustness. Shorter windows below 10 s resulted in unreliable estimation of low-frequency resolution, whereas windows exceeding 20 s reduced the total number of images. Therefore, 15 s represented a practical compromise between frequency resolution and sample size. High-amplitude disturbances related to movement or environmental interference were attenuated using the artifact subspace reconstruction (ASR) function in EEGLAB toolbox.
Each 15-s EEG was denoised using Multiscale Principal Component Analysis (MSPCA). MSPCA is a hybrid signal processing approach that combines wavelet decomposition and principal component analysis (PCA) to remove noise across multiple temporal scales. In MSPCA, the EEG signal is first decomposed into multiresolution wavelet subbands, and PCA is then applied within each scale to separate structured neural activity from noise components. Reconstruction is performed after suppressing noise-dominated components, resulting in enhanced signal quality while preserving physiologically meaningful brain rhythms.
The denoising procedure was implemented in MATLAB using the wmulden function with a decomposition level of 9 and the Symlet wavelet of order 4. Soft thresholding was applied based on the universal threshold selection rule, and the number of principal components retained at both the approximation and final reconstruction stages was determined using Kaiser’s criterion. These parameter settings were selected to provide effective multiscale noise reduction while maintaining clinically relevant EEG information. The number of images generated per subject for the SSRI and rTMS datasets after preprocessing is presented in Table 2.
Table 2.
Number of images per subject for SSRI and rTMS datasets.
4. Methodology
This section describes the steps of the CAD system for predicting the outcomes of depression therapies. Figure 2 provides an overview of the entire pipeline.
Figure 2.
Workflow of the proposed computer-aided decision (CAD) system for predicting depression therapy outcomes from pre-treatment EEG signals. The pipeline comprises five stages: (1) Preprocessing: raw EEG recordings from the SSRI and rTMS datasets are band-pass filtered (0.5–70 Hz), notch filtered (50 Hz), motion-artifact corrected via artifact subspace reconstruction (ASR), re-referenced to the common average reference (CAR), denoised using Multiscale Principal Component Analysis (MSPCA), and segmented into non-overlapping 15-s epochs; (2) Image Generation: each epoch and channel (19 total) is transformed into a time-frequency image using Continuous Wavelet Transform (CWT), Variational Mode Decomposition (VMD), or their pixel-wise average (Fusion); (3) Cross-Validation: 6-fold CV applied at either the image level (image-independent) or patient level (subject-independent); (4) Classification: four ImageNet-pretrained CNNs (ResNet-18, MobileNet-V3, EfficientNet-B0, TinyViT-Hybrid) are fine-tuned to classify images as responder (R) or non-responder (NR); (5) Performance Metrics: accuracy, precision, recall, specificity, and F1-score are reported at the image level; subject-level accuracy with 95% CI is reported under subject-independent CV.
4.1. Time-Frequency Image Generation
To extract informative spectral-temporal patterns from pre-treatment EEG signals, three types of time-frequency representations are generated: CWT-based images, VMD-based images, and a fused representation combining both domains. This section describes the image generation techniques.
4.1.1. Continuous Wavelet Transform
Continuous Wavelet Transform (CWT) is used to generate high-resolution time-frequency representations that capture the nonlinear and non-stationary characteristics of EEG data [26]. The CWT provides a multi-resolution time-frequency decomposition of a signal by correlating it with scaled and translated versions of a mother wavelet [27,40]. For a continuous signal , the CWT is defined as:
where is the scale parameter, is the translation parameter, is the mother wavelet, and denotes complex conjugation. The scale a is inversely related to frequency f through , where is the centre frequency of the wavelet and is the sampling interval.
The analytic Morlet wavelet was selected as the mother wavelet due to its optimal joint time-frequency localization [26]. The Morlet wavelet is defined as:
where is the central angular frequency. The analytic form of the Morlet wavelet suppresses negative frequency components, yielding a one-sided spectrum that is more suitable for interpreting oscillatory neural activity in EEG signals. The scalogram, representing the time-frequency energy distribution, is obtained from the squared modulus of the CWT coefficients:
This representation captures how the spectral energy of the EEG signal evolves over time, enabling visualization of non-stationary oscillatory dynamics associated with different frequency bands.
In the implementation, each denoised 15-s EEG segment was transformed into a scalogram image independently for each of the 19 channels. The CWT was computed using MATLAB’s cwt function with the analytic Morlet wavelet (‘amor’), with 12 voices per octave to provide fine-grained frequency resolution across scales. The frequency range was restricted to Hz to cover all clinically relevant EEG rhythms, namely delta (: 1–4 Hz), theta (: 4–8 Hz), alpha (: 8–13 Hz), beta (: 13–30 Hz), and low gamma (: 30–60 Hz), while excluding DC drift and high-frequency noise components. The magnitude of the CWT coefficients was used as the pixel intensity in each scalogram, and the jet colormap was applied to map amplitude values to a perceptually distinguishable colour space. Each scalogram was rendered as a borderless image at a resolution of 200 DPI and saved in PNG format. The resulting CWT images constitute a structured two-dimensional representation of each EEG, with the horizontal axis encoding time and the vertical axis encoding frequency, suitable for input to pretrained convolutional neural networks. Figure 3 shows representative CWT scalogram images generated from the Fp1 channel for R and NR subjects from each dataset.
Figure 3.
CWT-based time-frequency images generated from the Fp1 channel for responder (R) and non-responder (NR) patients across SSRI and rTMS therapies. The images illustrate the spectral-temporal structure of EEG activity used for predicting treatment outcomes.
4.1.2. Variational Mode Decomposition
Variational Mode Decomposition (VMD) is an adaptive, non-recursive signal decomposition method that overcomes the mode-mixing and boundary-effect limitations of Empirical Mode Decomposition [28,41]. VMD simultaneously decomposes a real-valued signal into K band-limited intrinsic mode functions (IMFs), denoted , each centred around a distinct centre frequency . The decomposition is formulated as a constrained variational optimisation problem that minimises the total bandwidth of all modes subject to their sum reconstructing the original signal [42].
For each mode , its analytic signal is computed via the Hilbert transform, and the spectrum is frequency-shifted to baseband by modulating with [43]. The constrained optimisation problem is defined as:
where is the Dirac delta, ∗ denotes convolution, and is the partial derivative with respect to time. The -norm of the time-derivative of the analytic baseband signal serves as a proxy for the instantaneous bandwidth of each mode.
To solve Equation (4), the constrained problem is converted into an unconstrained saddle-point problem using an augmented Lagrangian:
where is a quadratic penalty parameter controlling mode bandwidth, and is the Lagrange multiplier enforcing the reconstruction constraint. The modes, centre frequencies, and multipliers are iteratively updated in the Fourier domain until convergence. The spectral update rule for each mode is:
and the centre frequency is updated as:
where and denote the Fourier transforms of and , respectively. Equation (7) computes the centre of gravity of the power spectrum of the k-th mode, which drives each mode towards its dominant frequency bin. The Lagrange multiplier is updated as:
where is the dual ascent step size. Iterations continue until the convergence criterion is satisfied.
Unlike CWT, which directly produces a time-frequency map, VMD yields a set of narrowband temporal waveforms. To obtain a structured two-dimensional time-frequency representation compatible with CNN input, the Short-Time Fourier Transform (STFT) spectrogram was computed for each IMF and the resulting complex spectrograms were accumulated across all modes [44]. For a given IMF , the STFT is defined as:
where is a sliding analysis window centred at time t. The aggregate complex spectrogram across all K modes is then:
and the magnitude was rendered as a two-dimensional image. This summation strategy integrates the spectral contributions of each narrow-band mode, thereby recovering a full time-frequency picture of the original EEG epoch while preserving the sharp frequency localisation achieved by VMD decomposition.
In the implementation, each denoised 15-s EEG was processed independently for each of the 19 channels. MATLAB’s vmd function was applied with IMFs, decomposing each single-channel epoch into 20 band-limited components (i.e., IMFs). The STFT spectrogram of each IMF was subsequently computed using MATLAB’s spectrogram function with a Hamming window of length 256 samples, an overlap of 250 samples (step size of 6 samples), and 7680 FFT points at the resampled rate of Hz. The high overlap ensured dense temporal sampling of the spectrogram to preserve fine-grained temporal dynamics within each epoch. The 20 complex spectrograms were accumulated as in Equation (10), and the magnitude of the summed spectrogram was visualised. The frequency axis was restricted to Hz to focus on the clinically relevant delta, theta, alpha, beta and gamma EEG rhythms, consistent with the spectral range of interest for depression biomarker analysis. The jet colormap was applied to map magnitude values to colour, axes and tick labels were suppressed, and each image was saved in PNG format. The resulting VMD-based images constitute structured time-frequency representations suitable for direct input to pretrained CNNs. Figure 4 shows representative VMD-based spectrogram images generated from the Fp1 channel for R and NR subjects from each dataset.
Figure 4.
VMD-based time-frequency images generated from the Fp1 channel for responder (R) and non-responder (NR) patients across SSRI and rTMS therapies. The images highlight intrinsic mode variations that differentiate treatment-responsive from treatment-resistant neural activity.
4.1.3. Fusion of CWT and VMD Representations
CWT and VMD capture complementary aspects of EEG spectral-temporal dynamics. CWT provides a continuous, multi-resolution time-frequency representation that tracks non-stationary oscillatory activity across a broad frequency range, whereas VMD yields a set of narrowband, adaptively decomposed modes whose accumulated spectrogram reflects stable rhythmic components with fine frequency localisation. Since the two representations encode different but mutually reinforcing aspects of the same underlying neural signal, combining them into a single fused image is expected to improve the feature space available to the downstream CNN classifier [29,30].
Let and denote the CWT scalogram image and the VMD-based spectrogram image, respectively, both corresponding to the same EEG signal, channel, and subject. Prior to fusion, each image is independently normalised to the unit interval using min-max normalisation:
where and are computed over all pixel values of the individual image. This per-image normalisation ensures that neither representation dominates the fused output due to differences in absolute amplitude scale arising from the distinct computational pathways of CWT and VMD.
The fused image is obtained by pixel-wise arithmetic averaging of the two normalised representations:
where indexes a spatial pixel location. The averaging operation in Equation (12) preserves the dynamic range within , avoids saturation artefacts, and assigns equal weight to each representation. At each pixel, the fused value reflects the joint contribution of the instantaneous wavelet energy at that time-frequency coordinate from the CWT and the accumulated narrowband spectral energy from the VMD-derived spectrogram. Regions of the time-frequency plane that are consistently active in both representations are reinforced in the fused image, while modality-specific transient features are attenuated, yielding a more robust composite representation.
From an information-theoretic perspective, the pixel-wise average can be understood as a linear combination of two complementary feature maps. Because CWT and VMD decompose the same signal through non-adaptive and adaptive mechanisms respectively, their representations are not identical, and the fusion can capture structured patterns that neither representation alone encodes completely. This complementarity has been shown to improve classification performance in multimodal signal analysis tasks [29,30].
In the implementation, each CWT image and its corresponding VMD image, matched by subject, segment index, and channel identifier embedded in the filename, were loaded. Both images were converted to double-precision floating-point arrays using im2double and subsequently normalised to using mat2gray, which implements the min-max operation defined in Equation (11). The normalised images were then averaged pixel-wise as in Equation (12), and the resulting fused image was saved in PNG format using imwrite. The same procedure was applied independently to all 19 EEG channels across all subjects and segments in both the SSRI and rTMS datasets. The resulting fused images share the same spatial dimensions and filename convention as the individual CWT and VMD images, and serve as a third distinct input modality for the pretrained CNN classifiers evaluated in this study. Figure 5 shows representative fused images for R and NR subjects from each dataset. Table 3 summarises the complete set of parameters used for CWT, VMD, and Fusion image generation.
Figure 5.
Fusion-based EEG representations of responder (R) and non-responder (NR) across SSRI and rTMS therapies, generated from the Fp1 channel. The fused time-frequency patterns combine the information of CWT scalograms and VMD-derived spectrograms via pixel-wise averaging to emphasise discriminative structures for treatment-outcome prediction.
Table 3.
Summary of image generation parameters for CWT, VMD, and Fusion representations.
4.2. Pretrained CNN Models
Four pretrained convolutional neural network architectures are evaluated as classifiers in the proposed CAD system. All models are initialised with ImageNet pretrained weights and fine-tuned end-to-end on the EEG time-frequency images. The final classification layer of each model is replaced with a new linear layer with two output neurons corresponding to the R and NR classes.
ResNet-18 is an 18-layer residual network in which skip connections enable gradient flow across deeper layers, mitigating the vanishing gradient problem [31]. The original fully connected layer with 1000 outputs is replaced by a linear layer mapping the 512-dimensional feature vector to the two-class output space.
MobileNet-V3 Large is a lightweight architecture that combines inverted residual blocks with hard-swish activations and a Squeeze-and-Excitation attention mechanism, designed for efficient inference [32,33]. The last linear layer of its three-layer classifier head is replaced with a binary output layer while the preceding adaptive average pooling and the first two classifier layers are retained.
EfficientNet-B0 is the baseline model of the EfficientNet family, which applies compound scaling to jointly increase network depth, width, and input resolution [34,35]. Its final classification layer is replaced with a two-output linear layer while the preceding dropout and the feature extraction backbone remain intact.
TinyViT-Hybrid is a hybrid architecture proposed in this work that couples a convolutional feature extractor with a compact Transformer encoder [36,37,38]. The ResNet-18 backbone is retained up to and including layer4, producing spatial feature maps of shape for inputs. These feature maps are reshaped into a sequence of 49 non-overlapping tokens of dimension 512, each corresponding to one spatial position in the grid. Learnable 1-D positional embeddings are added to each token to inject spatial order information. The token sequence is then passed through a Transformer encoder consisting of two Multi-Head Self-Attention layers with attention heads, a feed-forward inner dimension of (MLP ratio of 4), and a dropout rate of 0.1. The self-attention mechanism allows each spatial token to attend to all other tokens, capturing global spatial dependencies across the feature map. The attended token sequence is normalised with Layer Normalisation, and the 512-dimensional mean-pooled representation is passed to a linear classification head. The Transformer block is implemented using timm’s Vision Transformer Block.
Detailed architectural diagrams of the four pretrained convolutional neural network models evaluated in this study, namely ResNet-18 (Figure S1), MobileNet-V3 (Figure S2), EfficientNet-B0 (Figure S3), and TinyViT-Hybrid (Figure S4), accompanied by descriptive text explaining the layer structure, spatial resolution evolution, and task-specific modifications applied to each model (Sections S1–S4).
All experiments were implemented in Python 3.14.3 using PyTorch 2.10.0. Each model was initialised with ImageNet pretrained weights provided by torchvision and fine-tuned end-to-end on the EEG time-frequency images. Training was performed using the AdamW optimiser with a learning rate of and a weight decay of . A cosine annealing learning rate scheduler was applied over epochs with a minimum learning rate of . The maximum number of training epochs per fold was set to 8, with early stopping applied on the validation loss with a patience of 3 epochs to prevent overfitting. The batch size was set to 32 for all models and experiments, and the cross-entropy loss was used as the training objective for all four architectures. All input images were resized to pixels and normalised using ImageNet statistics. No augmentation was applied during training and testing. A fixed random seed of 42 was used to ensure reproducibility across all runs, with the seed incremented by fold index at the start of each fold to diversify weight initialisation across folds. A summary of the architecture-specific and shared training hyperparameters is provided in Table 4.
Table 4.
Architecture configurations and shared training hyperparameters for the four pretrained CNN models.
4.3. Dataset Statistics and Validation Strategy
The CAD system is evaluated on the SSRI and rTMS datasets. In each dataset, EEG recordings are available for both R and NR patients. Regardless of class, all EEG signals are segmented into non-overlapping 15-s segments, and a time-frequency image is generated for each segment (each segment has 19 channels). Table 5 provides the total number of R and NR EEG images generated from the SSRI and rTMS datasets. The table shows that the number of images is large enough to support training of the pretrained CNN models. However, in the SSRI dataset the number of NR images exceeds that of R images, which is a direct consequence of the larger number of NR patients in the dataset.
Table 5.
Number of R and NR EEG images in the SSRI and rTMS datasets.
The test data can be categorised into four situations based on the predictions of the CAD system. A true positive (TP) occurs when a R image is correctly predicted as R by the model. A true negative (TN) occurs when a NR image is correctly predicted as NR. A false positive (FP) occurs when a NR image is incorrectly predicted as R. A false negative (FN) occurs when a R image is incorrectly predicted as NR. The performance of each CAD system was quantified using five standard evaluation metrics: accuracy, precision, recall, specificity, and F1-score. Accuracy reflects the overall correctness of the classifier and is defined as:
Precision reflects the model’s confidence in its R predictions by measuring the proportion of predicted R samples that are truly responders:
Recall quantifies how accurately the model detects R samples:
Specificity measures how effectively the model identifies non-R samples:
The F1-score combines precision and recall into a single balanced measure:
To minimise result bias, each model is evaluated under two complementary K-fold cross-validation (CV) strategies. The first is an image-independent 6-fold CV, in which all images from both classes are pooled and partitioned into six folds using stratified sampling to preserve the class ratio in each fold. Within each fold, 90% of the non-test images form the training set and the remaining 10% serve as the validation set. This strategy evaluates how well the model distinguishes R from NR images when training and test images may originate from the same subject and, therefore, provides an upper-bound estimate of image-level discrimination performance. Performance is assessed at the image level by comparing the predicted class label of every test image against its true label to derive a fold-level confusion matrix.
The second strategy is a subject-independent 6-fold CV, in which the data are partitioned at the subject level so that all images from a given subject appear exclusively in either the training or the test set within each fold, but never in both. For the SSRI dataset, which contains 12 responders and 18 non-responders, each test fold consists of 2 R and 3 NR, ensuring that every subject appears in the test set exactly once across the six folds. For the rTMS dataset, which contains 23 R and 23 NR, the first five test folds each contain 4 R and 4 NR, while the sixth fold contains 3 R and 3 NR to account for the remainder. In both cases, the remaining training subjects are further split into a 90% training subset and a 10% stratified validation subset. Subject-independent CV provides a more clinically realistic estimate of generalisation, as it simulates prediction for a previously unseen patient. In the subject-level CV strategy, a subject is considered correctly classified when the majority of the subject’s images are correctly classified. Table 6 summarises the configuration and purpose of the two CV strategies employed in this study.
Table 6.
Summary of the two CV strategies employed in this study.
In image-level strategies, accuracy, precision, recall, specificity, and F1-score are computed from the fold-level confusion matrix and reported as the mean and standard deviation across the six folds. For the subject-independent strategy, subject-level accuracy along with a 95% confidence interval (CI) is reported. Additionally, a per-channel accuracy analysis is performed in the subject-level CV strategy.
5. Results and Discussion
This section reports and discusses the experimental results. Two independent datasets are used to evaluate the proposed CAD system for predicting the outcomes of SSRI and rTMS therapies, as described in Section 2. In the first step of the proposed pipeline, the EEG signals are denoised following the steps described in Section 3 and segmented into non-overlapping 15-s windows. The time-frequency methods are then applied to the EEG signals to convert them into images. The EEG images are generated using CWT, VMD, and their fusion as described in Section 4.1. The resulting images are subsequently fed into four pretrained CNN architectures with fine-tuned classification layers, whose structures are described in Section 4.2.
5.1. Performance of CAD System for SSRI Therapy
Table 7 reports the image-level classification performance of the four pretrained CNN models under image-independent 6-fold CV for the SSRI dataset across all three time-frequency representations. Table 8 reports the corresponding subject-level accuracy under subject-independent CV. The average confusion matrices for all twelve model–representation combinations are visualised in Figure 6. The training and validation dynamics are presented in Figure 7, Figure 8 and Figure 9 for CWT, VMD, and Fusion representations respectively. The per-channel subject-level accuracy under subject-independent CV is shown in Figure 10.
Table 7.
Image-level classification performance (%) for the SSRI dataset using three time–frequency representations (CWT, VMD, and Fusion) and four deep learning models under image-level CV. Values are reported as mean ± standard deviation across 6 folds.
Table 8.
Subject-level classification accuracy (%) for the SSRI dataset under subject-independent CV.
Figure 6.
Average 6-fold confusion matrices for SSRI treatment response prediction under image-independent CV. Results are shown for three time-frequency representations—CWT (panels a–d), VMD (panels e–h), and Fusion (panels i–l)—and four CNN architectures (ResNet-18, EfficientNet-B0, MobileNet-V3, TinyViT-Hybrid). Each cell reports the mean count averaged across the six folds; non-integer values arise from this averaging. Rows indicate the true class label and columns the predicted label, where Resp = responder and Non_Resp = non-responder to SSRI therapy. The colour scale reflects the normalised proportion per cell. The approximate number of test images per fold is 2159 (∼709R and ∼1090NR).
Figure 7.
Training and validation loss and accuracy curves for CWT-based representations on the SSRI dataset under image-independent CV. Results are averaged across six folds with standard deviation shading.
Figure 8.
Training and validation loss and accuracy curves for VMD-based representations on the SSRI dataset under image-independent CV. Results are averaged across six folds with standard deviation shading.
Figure 9.
Training and validation loss and accuracy curves for the Fusion (CWT+VMD) representation on the SSRI dataset under image-independent CV. Results are averaged across six folds with standard deviation shading.
Figure 10.
Per-channel subject-level accuracy (%) for the SSRI dataset under subject-independent CV across CWT, VMD, and Fusion representations and four CNN models.
Across all representations, ResNet-18 consistently achieves the highest image-level performance. With CWT-based images, ResNet-18 attains the best overall accuracy of 99.43% ± 0.14%, a precision of 99.37% ± 0.43%, a recall of 99.20% ± 0.37%, a specificity of 99.59% ± 0.28%, and an F1-score of 99.28% ± 0.18%. The corresponding confusion matrix confirms this result, showing mean per-fold misclassification counts of only 4.50 false positives and 5.67 false negatives out of approximately 1798 test images per fold. The TinyViT-Hybrid hybrid achieves comparable performance with CWT images, reaching an accuracy of 99.28% ± 0.16% and an F1-score of 99.08% ± 0.21%, indicating that the integration of local convolutional features with global Transformer-based attention does not substantially compromise accuracy relative to the plain residual architecture on this dataset.
MobileNet-V3 and EfficientNet-B0 achieve lower but still competitive performance under CWT. MobileNet-V3 reaches 98.23% ± 0.47% accuracy and EfficientNet-B0 reaches 97.92% ± 0.43%, both exhibiting slightly larger standard deviations across folds compared to ResNet-18 and TinyViT-Hybrid, suggesting reduced stability. The confusion matrices for these two models show noticeably higher off-diagonal counts, with MobileNet-V3 producing mean false positive and false negative counts of 15.33 and 16.50 respectively, and EfficientNet-B0 producing 19.50 and 17.83 respectively, confirming their lower discriminative reliability relative to the residual-based models.
Comparing representations, CWT-based images yield the highest image-level accuracy across all four models for the SSRI dataset. VMD-based representations produce a modest but consistent reduction in performance, with ResNet-18 dropping to 98.84% ± 0.18% and TinyViT-Hybrid to 98.72% ± 0.26%. The confusion matrices for VMD confirm this trend: the ResNet-18 VMD matrix shows mean off-diagonal counts of 10.67 false positives and 10.17 false negatives, compared to only 4.50 and 5.67 for CWT. MobileNet-V3 and EfficientNet-B0 are more severely affected by the change in representation, falling to 97.19% and 96.75% accuracy respectively under VMD, with EfficientNet-B0 exhibiting the highest false positive count of 33.33 among all VMD configurations. These results suggest that, for the SSRI dataset, the multi-resolution continuous wavelet representation captures more discriminative time-frequency structure than the adaptive narrowband decomposition provided by VMD.
The fusion of CWT and VMD representations does not yield consistent improvements for SSRI therapy prediction. Under the fusion representation, ResNet-18 achieves 99.02% ± 0.24% accuracy, which is lower than the 99.43% obtained with CWT alone. A similar degradation is observed for all other models: MobileNet-V3 and EfficientNet-B0 achieve 97.21% and 97.08% respectively under fusion, compared to 98.23% and 97.92% with CWT. The confusion matrices for fusion-based models show higher false positive counts relative to CWT, particularly for MobileNet-V3 (28.33) and EfficientNet-B0 (34.50), indicating increased misclassification of non-R images as responders when the two representations are combined. This performance reduction suggests that the pixel-level averaging fusion introduces competing spectral information that partially obscures the discriminative patterns captured individually by CWT for SSRI therapy.
The learning curves in Figure 7, Figure 8 and Figure 9 provide further insight into the training dynamics of each model across all three representations. For CWT-based images, ResNet-18 and TinyViT-Hybrid exhibit the most stable convergence behaviour, with validation loss closely tracking training loss from epoch 1 and both quantities reaching near-zero values by epoch 8. The narrow standard deviation bands across folds confirm that these two models converge consistently regardless of the fold assignment. The validation accuracy for ResNet-18 and TinyViT-Hybrid under CWT starts at approximately 90% at epoch 1 and saturates near 99.5% by epoch 8, with very low inter-fold variance throughout training. MobileNet-V3 exhibits a noticeably wider standard deviation band in validation loss around epoch 5 under both CWT and fusion representations, indicating sensitivity to the particular fold composition. EfficientNet-B0 shows a persistent gap between training and validation loss across all three representations, with training loss remaining consistently higher than validation loss, suggesting that the model benefits from the generalisation properties of ImageNet pretraining but converges more slowly to the target distribution.
Under VMD-based images, all four models exhibit a steeper initial loss decline compared to CWT, with validation loss beginning at higher values in epoch 1 but dropping sharply by epoch 3. ResNet-18 and TinyViT-Hybrid again demonstrate the tightest inter-fold consistency, with standard deviation bands narrowing rapidly after epoch 2. MobileNet-V3 under VMD begins training with a notably high initial loss near 0.58 and wide confidence bands in early epochs, yet converges to a stable final accuracy of approximately 97%, consistent with its image-level result. For the fusion representation, convergence patterns closely resemble those of CWT for ResNet-18 and TinyViT-Hybrid, while MobileNet-V3 again displays elevated variance in validation loss around epoch 5, reflecting the broader instability of this architecture on the SSRI dataset. Across all representations, no model shows signs of overfitting within the 8-epoch training budget, as validation loss consistently decreases or stabilises without diverging from training loss.
The per-channel subject-level accuracy analysis presented in Figure 10 reveals clear spatial patterns in the discriminative information available for SSRI therapy prediction. Under CWT-based representations, the occipital channels O1 and O2 yield the highest subject-level accuracy, reaching 90% for both ResNet-18 and TinyViT-Hybrid, and 83.3% for EfficientNet-B0 and MobileNet-V3. This finding indicates that posterior cortical oscillatory activity, which is closely associated with alpha-band rhythms at rest, carries particularly informative pre-treatment signatures for identifying SSRI responders. The frontal channels F3, F4, F7, and Fp1 produce the lowest channel-level accuracy under CWT for most models, with ResNet-18 dropping to 61.1% at F3 and MobileNet-V3 to 60.8% at F3, reflecting the higher noise and artefact susceptibility of anterior electrodes in resting-state recordings. Under VMD, the channel accuracy profile shifts towards the temporal and central regions, with T3 and T4 yielding the highest accuracy for ResNet-18 (89.2%) and TinyViT-Hybrid (89.2%), while occipital channels remain informative but show less pronounced dominance compared to CWT. MobileNet-V3 achieves its peak channel accuracy of 85% at F4 under VMD, suggesting that the narrowband decomposition of VMD exposes different frequency-specific spatial patterns compared to the continuous multi-resolution representation of CWT. Under the fusion representation, the channel accuracy profiles become more uniform across the scalp, with most channels falling in the range of 70–85% and no single electrode showing a strong advantage, consistent with the slight overall performance degradation observed under fusion at the image level. These spatial differences between representations further support the observation that CWT and VMD encode complementary but non-redundant neural features, and that their combination through pixel-level averaging attenuates rather than reinforces the most discriminative spatial signals for SSRI response prediction.
At the subject level, Table 8 shows that the highest subject-level accuracy of 82.50% is achieved by multiple model–representation combinations, including CWT with ResNet-18, VMD with ResNet-18, VMD with TinyViT-Hybrid, Fusion with MobileNet-V3, and Fusion with EfficientNet-B0. However, all subject-level results are associated with relatively large standard deviations in the range of 11–17%, which reflects the limited number of test subjects per fold (five per fold: two responders and three non-responders) inherent to the subject-independent split of the 30-subject SSRI cohort. The wide 95% confidence intervals reported in Table 8 further underscore this variability. Notably, CWT with MobileNet-V3 yields the lowest subject-level accuracy of 71.67%, consistent with its lower image-level performance. Overall, the subject-independent results indicate that ResNet-18 and TinyViT-Hybrid generalise most reliably to unseen SSRI patients across all three time-frequency representations.
5.2. Performance of CAD System for rTMS Therapy
Table 9 reports the image-level classification performance of the four pretrained CNN models under image-independent 6-fold CV for the rTMS dataset across all three time-frequency representations. Table 10 reports the corresponding subject-level accuracy under subject-independent CV. The average confusion matrices for all twelve model–representation combinations are visualised in Figure 11. The training and validation dynamics are presented in Figure 12, Figure 13 and Figure 14 for CWT, VMD, and Fusion representations respectively. The per-channel subject-level accuracy under subject-independent CV is shown in Figure 15.
Table 9.
Image-level classification performance (%) for the rTMS dataset using three time–frequency representations (CWT, VMD, and Fusion) and four deep learning models under image-level CV. Values are reported as mean ± standard deviation across six folds.
Table 10.
Subject-level classification accuracy (%) for the rTMS dataset under subject-independent CV.
Figure 11.
Average 6-fold confusion matrices for rTMS treatment response prediction under image-independent CV. Results are shown for three time-frequency representations—CWT (panels a–d), VMD (panels e–h), and Fusion (panels i–l)—and four CNN architectures (ResNet-18, EfficientNet-B0, MobileNet-V3, TinyViT-Hybrid). Each cell reports the mean count averaged across the six folds; non-integer values arise from this averaging. Rows indicate the true class label and columns the predicted label, where Resp = responder and Non_Resp = non-responder to rTMS therapy. The colour scale reflects the normalised proportion per cell. The approximate number of test images per fold is 2711 (∼1384R and ∼1327NR), reflecting the balanced class distribution of the rTMS dataset.
Figure 12.
Training and validation loss and accuracy curves for CWT-based representations on the rTMS dataset under image-independent CV. Results are averaged across six folds with standard deviation shading.
Figure 13.
Training and validation loss and accuracy curves for VMD-based representations on the rTMS dataset under image-independent CV. Curves show mean performance across six folds with standard deviation shading.
Figure 14.
Training and validation loss and accuracy curves for the Fusion (CWT+VMD) representation on the rTMS dataset under image-independent CV. Shaded regions represent one standard deviation across six folds.
Figure 15.
Per-channel subject-level accuracy (%) for the rTMS dataset under subject-independent CV across CWT, VMD, and Fusion representations and four CNN models.
The rTMS dataset presents a notably different pattern of results compared to SSRI therapy. Whereas CWT was the best-performing representation for SSRI, VMD-based images achieve the highest overall image-level accuracy for rTMS. ResNet-18 with VMD attains the best overall performance of 98.77% ± 0.14% accuracy, a precision of 98.75% ± 0.32%, a recall of 98.84% ± 0.30%, a specificity of 98.69% ± 0.34%, and an F1-score of 98.80% ± 0.13%. The corresponding confusion matrix confirms this, showing mean off-diagonal counts of only 18.67 false positives and 22.33 false negatives out of approximately 2710 test images per fold. TinyViT-Hybrid under VMD achieves closely comparable performance at 98.59% ± 0.28% accuracy with 21.00 false positives and 18.67 false negatives, demonstrating that the hybrid attention mechanism provides a modest complementary benefit over the plain residual architecture on the VMD representation for rTMS.
The superiority of VMD over CWT for rTMS therapy is consistent across all four models, though the margin is small for the stronger architectures. ResNet-18 improves from 98.65% under CWT to 98.77% under VMD, and TinyViT-Hybrid improves from 98.40% to 98.59%. For the lighter architectures the advantage of VMD is also evident: MobileNet-V3 improves from 96.90% to 96.98% and EfficientNet-B0 improves from 96.48% to 96.88%. The confusion matrices confirm this trend for EfficientNet-B0, where the VMD matrix shows 38.50 false positives and 46.17 false negatives compared to 51.17 and 44.17 respectively under CWT, indicating that VMD reduces non-R misclassification for this model in the rTMS setting. This result suggests that the adaptive narrowband decomposition of VMD captures spectral features of rTMS-relevant neural oscillations more effectively than the continuous multi-resolution representation of CWT, reflecting a therapy-specific difference in the discriminative information encoded in the EEG signal.
In contrast to the SSRI results, the fusion of CWT and VMD representations yields a more nuanced outcome for rTMS. ResNet-18 under fusion achieves 98.49% accuracy, which is slightly lower than the 98.77% obtained with VMD alone but comparable to the 98.65% obtained with CWT alone. TinyViT-Hybrid under fusion achieves 98.54% accuracy, marginally lower than its VMD counterpart. Importantly, MobileNet-V3 achieves its highest image-level accuracy of 97.07% ± 0.47% under the fusion representation, surpassing both its CWT (96.90%) and VMD (96.98%) results, which is the only instance across all experiments where fusion produces a clear improvement. EfficientNet-B0 under fusion achieves 96.43% accuracy, marginally lower than its VMD result, with 47.83 false positives and 49.00 false negatives, the highest false negative count among all rTMS configurations.
Across all representations and models, the rTMS results exhibit lower inter-fold standard deviations compared to SSRI, reflecting the larger size of the rTMS dataset and its balanced class distribution. The standard deviation of accuracy across the six folds is at most 0.47% for any model–representation combination in the rTMS dataset. ResNet-18 consistently achieves the smallest standard deviation across all three representations (0.34% for CWT, 0.14% for VMD, and 0.12% for fusion), confirming its superior stability relative to the other architectures for rTMS therapy prediction.
The learning curves in Figure 12, Figure 13 and Figure 14 reveal important differences in convergence behaviour between models and representations on the rTMS dataset. Under CWT, ResNet-18 and TinyViT-Hybrid demonstrate the most stable and rapid convergence, with validation loss closely tracking training loss from epoch 1, narrow standard deviation bands throughout training, and validation accuracy saturating near 98.5–99% by epoch 8. TinyViT-Hybrid exhibits a notably lower initial training loss of approximately 0.35 under CWT compared to ResNet-18 at 0.47, and its validation accuracy starts at 88% at epoch 1, the highest among all models under this representation, reflecting the benefit of its hybrid attention mechanism for capturing global spatial dependencies in CWT scalograms from the very first epoch. EfficientNet-B0 under CWT exhibits the largest persistent gap between training and validation loss across all eight epochs, with training loss remaining above validation loss throughout, and a wider standard deviation band in validation accuracy during epochs 2–5, indicating higher variability across folds. MobileNet-V3 under CWT shows broadly similar convergence to EfficientNet-B0 but with a somewhat tighter validation band after epoch 4.
Under VMD-based images, the convergence dynamics for ResNet-18 and TinyViT-Hybrid are qualitatively similar to CWT but with slightly higher initial loss values. ResNet-18 begins training at a loss of approximately 0.51 under VMD and descends smoothly to below 0.05 by epoch 8, with very narrow inter-fold standard deviation bands confirming consistent convergence. TinyViT-Hybrid under VMD shows the tightest validation loss band of any model–representation combination across the entire rTMS experiment, reaching near-zero validation loss by epoch 7 with negligible spread across folds, consistent with its strong and stable image-level accuracy. EfficientNet-B0 under VMD shows a persistent training-to-validation loss gap that remains visible at epoch 8, suggesting that this model continues to generalise better than it trains on VMD features, which is atypical and may reflect the mismatch between the VMD spectrogram’s spectral structure and the compound-scaling architecture of EfficientNet-B0. MobileNet-V3 under VMD exhibits a notably wide validation loss standard deviation band at epoch 4, caused by inter-fold variability in how quickly this model adapts to the VMD representation, though the band narrows substantially by epoch 6 and the model ultimately converges to a stable accuracy near 97%.
Under the fusion representation, the learning curve profiles of ResNet-18 and TinyViT-Hybrid remain stable and closely mimic their VMD counterparts, with validation accuracy reaching approximately 98.5% and 99% respectively by epoch 8. MobileNet-V3 under fusion displays a pronounced spike in the validation loss standard deviation band around epoch 5, a pattern also observed under CWT for the same model on the SSRI dataset, suggesting that MobileNet-V3 is systematically sensitive to fold composition when trained on mixed-domain representations. Despite this transient instability, the model recovers to a stable accuracy by epoch 7, consistent with its best image-level performance under fusion among the lighter architectures. EfficientNet-B0 under fusion shows the most gradual convergence of all rTMS configurations, with training accuracy still climbing noticeably between epochs 6 and 8 and the validation accuracy standard deviation band remaining wider than for ResNet-18 and TinyViT-Hybrid throughout training.
The per-channel subject-level accuracy analysis presented in Figure 15 reveals clear spatial and representation-dependent differences in the discriminative power of individual EEG channels for rTMS therapy prediction. Under VMD-based representations, the temporal and frontal regions yield the highest per-channel accuracy. ResNet-18 achieves its peak channel accuracy of 85.6% at F8 under VMD, and TinyViT-Hybrid reaches 87.7% at Fp2, also under VMD. MobileNet-V3 achieves peak channel accuracies of 83.2% at C3 and C4, and EfficientNet-B0 peaks at 83.5% at P3 and T3 under VMD. These results confirm that frontotemporal channels carry particularly strong discriminative information for rTMS response prediction, consistent with the known involvement of the prefrontal cortex and its connectivity with the default mode network in rTMS treatment effects. Under CWT-based representations, the channel accuracy profiles are more uniform and generally lower, with most models achieving peak accuracies between 78% and 82% at parietal and temporal channels including Pz, T3, and T5, and dropping noticeably at Fp1 and Fp2 for EfficientNet-B0 and MobileNet-V3, which reach as low as 62.6% and 62.2% respectively at those channels. This frontal weakness under CWT, also observed for SSRI but to a lesser extent, reflects the greater susceptibility of CWT scalograms to frontal artefact contamination in the resting-state EEG. Under the fusion representation, the channel accuracy profiles are intermediate between CWT and VMD for most channels, with TinyViT-Hybrid achieving its best fusion channel accuracy of 83.2% at Fz, and MobileNet-V3 peaking at 80.9% at T3. The overall flattening of the channel accuracy profile under fusion, combined with the reduction in peak accuracy relative to VMD, further supports the conclusion that pixel-level averaging of CWT and VMD representations partially attenuates the spatially localised discriminative features most relevant for rTMS outcome prediction.
At the subject level, Table 10 shows that the highest subject-level accuracy of 83.53% is jointly achieved by VMD with ResNet-18 and VMD with EfficientNet-B0, both with a standard deviation of 12.30% and a 95% CI of [72.81, 94.25]%. These results confirm that VMD-based representations are particularly effective for rTMS outcome prediction at the subject level, consistent with the image-level findings. CWT with MobileNet-V3 achieves the next highest subject-level accuracy of 81.15% ± 15.41%, while CWT with EfficientNet-B0 yields the lowest subject-level accuracy of 71.63% ± 21.63%, the widest CI among all rTMS configurations. Under the fusion representation, all four models achieve subject-level accuracy in the range of 71.63–76.69%, consistently lower than the best VMD results. Overall, the subject-independent results confirm that VMD with ResNet-18 provides the most reliable framework for predicting rTMS therapy outcomes in unseen patients.
Complete per-fold image-level classification results for all model–representation–dataset combinations (Tables S1–S24), reporting Accuracy, Precision, Recall, Specificity, and F1-Score across the six image-level CV folds together with the mean and standard deviation, enabling a transparent assessment of fold-to-fold variability and performance stability. Supplementary Materials S3: Per-fold training and validation loss and accuracy curves (Figures S5–S28) for all combinations of four models (ResNet-18, MobileNet-V3, EfficientNet-B0, TinyViT-Hybrid), three time-frequency representations (CWT, VMD, Fusion), and two datasets (SSRI, rTMS), each presented in a grid layout for direct visual inspection of convergence behaviour and consistency across folds.
5.3. Comparative Performance and Architectural Analysis of the Evaluated Networks
To quantify whether the observed performance differences between architectures are statistically significant, pairwise Wilcoxon signed-rank tests were conducted on the per-fold accuracy scores across all model pairs, representations, and both therapy datasets. The resulting p-values are reported in Table 11, with a significance threshold of . It should be noted that all statistical comparisons are based exclusively on image-level CV results, where the larger number of test samples per fold provides sufficient statistical power for pairwise testing. Subject-level results, which involve only a small number of test subjects per fold, are not included in this analysis as the limited sample size would preclude reliable non-parametric significance testing.
Table 11.
Pairwise statistical comparison (Wilcoxon signed-rank test, p-values) of per-fold image-level accuracy between the four CNN architectures across three time-frequency representations for the SSRI and rTMS datasets.
The statistical analysis reveals a clear and consistent two-tier architectural hierarchy across both datasets and all representations. ResNet-18 and TinyViT-Hybrid form a statistically indistinguishable pair in every experimental condition, with all pairwise p-values exceeding 0.08 for SSRI and 0.22 for rTMS. This confirms that the addition of the Transformer encoder to the ResNet-18 backbone does not yield a statistically meaningful accuracy gain over the plain residual architecture within the training budget used in this study.
Both residual-family models are significantly superior to MobileNet-V3 and EfficientNet-B0 across nearly all conditions, with p-values consistently below 0.001 for both datasets. The performance gap between the two tiers becomes most pronounced under the fusion representation, where the lighter architectures appear less capable of simultaneously integrating the mixed-domain features arising from the pixel-level combination of CWT and VMD images. The comparison between EfficientNet-B0 and MobileNet-V3 is inconsistent across conditions, reaching significance only under CWT for SSRI () and under fusion for rTMS (), while remaining non-significant in all other cases. This indicates that these two architectures occupy a statistically equivalent performance tier, and the choice between them should be guided by deployment constraints such as model size and inference latency rather than classification accuracy.
5.4. Influence of Time-Frequency Image Type on Classification Efficacy
To assess whether the choice of time-frequency representation significantly affects classification performance, pairwise Wilcoxon signed-rank tests were conducted on the per-fold image-level accuracy scores across the three representations for each therapy dataset. All accuracy scores were pooled across the four CNN architectures within each therapy to maximise statistical power. The resulting p-values are summarised in Table 12, with a significance threshold of . As with the architectural comparisons reported in Section 5.3, all statistical tests are based exclusively on image-level CV results, as the limited number of test subjects per fold in the subject-independent setting does not provide sufficient statistical power for reliable non-parametric significance testing.
Table 12.
Pairwise statistical comparison (Wilcoxon signed-rank test, p-values) of per-fold image-level accuracy between the three time-frequency representations for the SSRI and rTMS datasets.
The results reveal a clear therapy-dependent pattern in how much the choice of representation influences classification performance. For the SSRI dataset, CWT is significantly superior to both VMD and fusion, with p-values below 0.0001 for both comparisons, confirming that the multi-resolution continuous wavelet representation captures substantially more discriminative information for SSRI therapy prediction than either the adaptive narrowband VMD decomposition or their pixel-level combination. The comparison between VMD and fusion for SSRI does not reach significance (), indicating that fusing CWT and VMD through pixel-level averaging does not produce a representation that is statistically distinguishable from VMD alone in this therapy context. Taken together, these findings confirm that CWT is the representation of choice for SSRI therapy, and that fusion provides no statistically meaningful benefit over the individual representations.
For the rTMS dataset, the representation differences are smaller in magnitude and the statistical picture is more nuanced. CWT and VMD differ significantly (), with VMD yielding higher accuracy as reported in Section 5.2, confirming that the adaptive decomposition of VMD is statistically superior to CWT for rTMS outcome prediction. In contrast, CWT and fusion are not significantly different (), and VMD versus fusion only approaches significance (), suggesting that for rTMS the three representations produce broadly comparable performance at the image level even though VMD holds a marginal numerical advantage. These results indicate that the rTMS classification task is less sensitive to the choice of time-frequency representation than SSRI, likely reflecting the more balanced class distribution and larger dataset size of the rTMS cohort, which allows all three representations to extract sufficient discriminative information. Overall, the statistical analysis of representation effects reinforces the conclusion that the optimal time-frequency image type is therapy-specific: CWT is preferred for SSRI and VMD is preferred for rTMS, while fusion does not consistently improve upon the best single-representation result in either therapy context.
5.5. Comparison with State-of-the-Art Studies
Table 13 summarises the most relevant prior studies on EEG-based prediction of depression therapy outcomes alongside the best image-level results achieved by the proposed framework. Direct numerical comparison across studies should be interpreted with caution, as differences in dataset size, CV strategy, feature extraction pipeline, and class balance can substantially influence reported accuracy figures.
Table 13.
Comparison with state-of-the-art studies on EEG-based prediction of depression therapy outcome on image level.
Early studies on EEG-based SSRI response prediction relied on hand-crafted spectral and connectivity features combined with shallow classifiers. Mumtaz et al. [14] combined wavelet transform, STFT, and empirical mode decomposition features with logistic regression under 10-fold CV, achieving 91.6% accuracy on the SSRI dataset. Khodayari et al. [15] extracted power spectral density, coherence, and multifractal analysis features and reported 87.4% accuracy with an 80/20 train-test split. While these approaches established the feasibility of EEG-based SSRI outcome prediction, their dependence on manually engineered features limits their capacity to capture the complex nonlinear and non-stationary dynamics inherent in pre-treatment EEG recordings.
More recent deep learning approaches have substantially improved performance by learning representations directly from structured EEG inputs. Shahabi et al. [5] applied an ensemble transfer learning framework to CWT images for SSRI prediction, achieving 96.55% accuracy under 5-fold CV. A subsequent study by the same group [21] incorporated a recurrent attention mechanism applied to raw EEG signals and reported 98.84% accuracy, which represents the strongest prior result on the SSRI dataset. For rTMS therapy, Shahabi et al. [7] combined CWT images with a CNN-LSTM-Attention architecture and achieved 97.1% accuracy, while a bidirectional LSTM ensemble model applied to raw EEG signals reached 98.51% [24].
The proposed framework achieves 99.43% image-level accuracy for SSRI using CWT images with ResNet-18, surpassing all prior SSRI results including the recurrent attention model of [21]. For rTMS therapy, the proposed VMD-based ResNet-18 framework attains 98.77% accuracy, exceeding the CWT-CNN-LSTM-Attention result of [7] and closely matching the bidirectional LSTM ensemble of [24]. Notably, these competitive results are achieved without any recurrent or attention layers in the classification network, relying instead on the inductive bias of pretrained residual convolutional features applied to carefully constructed time-frequency images. This demonstrates that the choice of time-frequency representation is at least as important as architectural complexity for EEG-based therapy prediction, and that simple fine-tuning of a pretrained ResNet-18 on well-designed spectral images can match or outperform substantially more complex sequence modelling architectures. Furthermore, unlike most prior studies that evaluate a single therapy modality, the proposed framework provides a unified comparative analysis across both SSRI and rTMS therapies, revealing the therapy-specific nature of the optimal time-frequency representation, a finding that has direct implications for the design of future EEG-based depression therapy prediction systems.
The proposed CAD system offers several notable advantages over existing approaches for EEG-based depression therapy outcome prediction.
- Therapy-specific representation analysis: The framework systematically compares three time-frequency representations across two therapy datasets, revealing that the optimal representation is therapy-specific: CWT for SSRI and VMD for rTMS. This finding has direct implications for the design of future EEG-based treatment selection systems.
- No recurrent or attention layers required: State-of-the-art image-level accuracy is achieved by simply fine-tuning a pretrained ResNet-18 on well-constructed time-frequency images, without any recurrent or attention mechanisms, reducing training complexity and computational cost relative to LSTM- or Transformer-based classifiers.
- Computationally efficient: All four architectures are lightweight pretrained CNNs that converge within 8 epochs on a standard GPU, making the framework practical for clinical research settings where computational resources and labelled data are limited.
- Rigorous dual-strategy evaluation: The system is assessed under both image-independent and subject-independent 6-fold CV, providing both an upper-bound discrimination estimate and a clinically realistic generalisation estimate, complemented by per-channel accuracy analysis across all 19 EEG electrodes.
- Adaptive decomposition via VMD: VMD adaptively decomposes EEG into narrowband intrinsic mode functions with fine frequency localisation, capturing oscillatory components inaccessible to fixed-scale wavelet analysis. Its statistically confirmed superiority for rTMS demonstrates the value of adaptive decomposition in EEG-based deep learning pipelines.
- Statistically validated conclusions: All findings on architectural hierarchy and representation efficacy are supported by pairwise Wilcoxon signed-rank tests, confirming that observed differences are genuine and reproducible rather than artefacts of fold assignment.
- Spatial interpretability: Per-channel accuracy analysis identifies occipital dominance for SSRI under CWT and frontotemporal dominance for rTMS under VMD, consistent with known neurophysiological mechanisms and offering guidance for electrode selection in resource-constrained deployments.
Despite the promising results, several limitations of the present study should be acknowledged. First, the patient groups are small (30 SSRI and 46 rTMS patients), and although the subject-independent CV provides a realistic estimate of generalisation, the subject-level accuracy values are associated with wide CIs that reflect this limited sample size. Replication on larger, multi-centre datasets is therefore necessary before any clinical conclusions can be drawn. Second, the absence of an external independent validation cohort is a key limitation. The subject-independent CV results provide a realistic estimate of generalisation to unseen patients within each cohort, but multi-centre external validation is necessary before clinical deployment. Third, the fusion strategy evaluated here-pixel-wise arithmetic averaging of per-image normalised representations-represents one of many possible fusion paradigms. Feature-level, channel-wise, or attention-guided late-fusion strategies may yield different results and are identified as a direction for future work. Fourth, the deep learning models employed in this study offer limited interpretability. Future work should integrate explainable AI techniques such as gradient-weighted class activation mapping (Grad-CAM) to identify the EEG frequency bands, temporal segments, and scalp regions that contribute most to treatment-response predictions, thereby supporting transparent and clinically interpretable decision-making by psychiatrists. Finally, no data augmentation was applied in this study; exploring augmentation strategies tailored to EEG time-frequency images may further improve generalisation in low-sample settings.
6. Conclusions
This study presented a comprehensive deep learning framework for predicting depression therapy outcomes from pre-treatment EEG signals using three time-frequency representations and four pretrained CNN architectures evaluated on two independent therapy datasets. The key findings can be summarised as follows.
The choice of time-frequency representation is therapy-specific and has a statistically significant impact on classification performance. CWT-based scalograms yield superior discrimination for SSRI therapy prediction, with ResNet-18 achieving 99.43% image-level accuracy, while VMD-based spectrograms are statistically superior for rTMS outcome prediction, with ResNet-18 reaching 98.77%. Pixel-level fusion of CWT and VMD does not consistently improve performance over the best single representation in either therapy context. These findings demonstrate that the neurophysiological signatures predictive of SSRI and rTMS response are encoded in different spectral-temporal structures, and that selecting the representation to match the therapy is a critical design decision for EEG-based CAD systems.
A clear two-tier architectural hierarchy was confirmed by pairwise Wilcoxon signed-rank tests. ResNet-18 and TinyViT-Hybrid are statistically indistinguishable and both significantly outperform MobileNet-V3 and EfficientNet-B0 across all conditions and both datasets. Notably, competitive results are achieved without any recurrent or attention mechanisms, demonstrating that careful signal representation combined with a simple pretrained residual network can match or surpass substantially more complex architectures reported in prior work.
Per-channel analysis revealed that occipital channels carry the most discriminative information for SSRI response under CWT, while frontotemporal channels are most predictive for rTMS response under VMD, findings that are consistent with known neurophysiological mechanisms and offer practical guidance for electrode selection in resource-constrained clinical settings. At the subject level, the framework achieves 82.50% and 83.53% accuracy for SSRIs and rTMS respectively under strict subject-independent CV, highlighting the remaining challenge of generalisation to unseen patients in small clinical cohorts.
Future work should focus on validating the proposed framework on larger, multi-centre datasets to improve subject-level generalisation, and on integrating explainable AI methods to identify the EEG biomarkers and frequency bands that are most informative for each therapy, thereby supporting transparent and clinically interpretable treatment selection.
Supplementary Materials
The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/brainsci16030301/s1, Supplementary Materials S1: Detailed architectural diagrams of the four pretrained convolutional neural network models evaluated in this study, namely ResNet-18 (Figure S1), MobileNet-V3 (Figure S2), EfficientNet-B0 (Figure S3), and TinyViT-Hybrid (Figure S4), accompanied by descriptive text explaining the layer structure, spatial resolution evolution, and task-specific modifications applied to each model (Sections S1–S4). Supplementary Materials S2: Complete per-fold image-level classification results for all model–representation–dataset combinations (Tables S1–S24), reporting Accuracy, Precision, Recall, Specificity, and F1-Score across the six image-level CV folds together with the mean and standard deviation, enabling a transparent assessment of fold-to-fold variability and performance stability. Supplementary Materials S3: Per-fold training and validation loss and accuracy curves (Figures S5–S28) for all combinations of four models (ResNet-18, MobileNet-V3, EfficientNet-B0, TinyViT-Hybrid), three time-frequency representations (CWT, VMD, Fusion), and two datasets (SSRI, rTMS), each presented in a grid layout for direct visual inspection of convergence behaviour and consistency across folds.
Author Contributions
Conceptualization, H.A., S.B. and J.F.S.; Methodology, H.A., S.B. and J.F.S.; Software, H.A. and J.F.S.; Validation, H.A., S.B. and J.F.S.; Formal analysis, H.A., S.B. and J.F.S.; Investigation, H.A., S.B. and J.F.S.; Data curation, H.A., R.R. and R.K.; Writing—original draft preparation, H.A. and S.B.; Writing—review and editing, H.A., R.N., S.M., H.C. and M.M.; Visualization, H.A.; Supervision, R.N., S.M., H.C. and M.M.; Project administration, R.N., S.M., H.C. and M.M. All authors have read and agreed to the published version of the manuscript.
Funding
This research received no external funding.
Institutional Review Board Statement
This study made use of two EEG signal datasets: the Mumtaz database and the Atieh Hospital database. The protocols for SSRI therapy in the Mumtaz database were approved by the Ethics Committee of Universiti Teknologi PETRONAS. The rTMS therapy protocols in the Atieh Hospital database were approved by the Ethics Committee of Shahid Beheshti University of Medical Sciences. All procedures were conducted in accordance with the Declaration of Helsinki.
Informed Consent Statement
Informed consent was obtained by the original investigators for both datasets. No additional data were collected for this study.
Data Availability Statement
The SSRI dataset used in this study are publicly available on Figshare at https://figshare.com/articles/dataset/EEG_Data_New/4244171 (accessed on 1 January 2026). The rTMS dataset was collected at Atieh Hospital, Tehran, Iran, under the ethical approval of Shahid Beheshti University of Medical Sciences, and is available from the corresponding author upon reasonable request.
Acknowledgments
The authors would like to thank the contributors of the Mumtaz EEG dataset on Figshare. We also acknowledge the valuable discussions and support received from colleagues and our respective institutions. We acknowledge the great support received from Bizhan Alipour Pijani.
Conflicts of Interest
The authors declare no conflict of interest.
Abbreviations
| ACC | Accuracy |
| AdamW | Adaptive Moment Estimation with Weight Decay |
| ASR | Artifact Subspace Reconstruction |
| BDI | Beck Depression Inventory |
| CAD | Computer-Aided Decision |
| CAR | Common Average Reference |
| CI | Confidence Interval |
| CNN | Convolutional Neural Network |
| CWT | Continuous Wavelet Transform |
| EEG | Electroencephalography |
| EMD | Empirical Mode Decomposition |
| F1 | F1-score |
| FFT | Fast Fourier Transform |
| FN | False Negative |
| FP | False Positive |
| GPU | Graphics Processing Unit |
| Grad-CAM | Gradient-weighted Class Activation Mapping |
| ICA | Independent Component Analysis |
| IMF | Intrinsic Mode Function |
| LR | Logistic Regression |
| LSTM | Long Short-Term Memory |
| MFA | Multifractal Analysis |
| MLP | Multi-Layer Perceptron |
| MSPCA | Multiscale Principal Component Analysis |
| NR | Non-Responder |
| PCA | Principal Component Analysis |
| PNG | Portable Network Graphics |
| PRE | Precision |
| PSD | Power Spectral Density |
| R | Responder |
| REC | Recall |
| ROC | Receiver Operating Characteristic |
| rTMS | Repetitive Transcranial Magnetic Stimulation |
| SPE | Specificity |
| SSRI | Selective Serotonin Reuptake Inhibitor |
| STD | Standard Deviation |
| STFT | Short-Time Fourier Transform |
| SVM | Support Vector Machine |
| TL | Transfer Learning |
| TN | True Negative |
| TP | True Positive |
| VMD | Variational Mode Decomposition |
| ViT | Vision Transformer |
References
- Modonesi, C.; Serafini, S.; Giuliani, A. Urban Systems Between the Environment, Human Health and Society: An Overview. Systems 2025, 13, 487. [Google Scholar] [CrossRef]
- Chen, M.; Li, M.; Wan, D.; Zha, J. Global, regional, and national burden of depressive disorders in older adults from 1990 to 2021, with projections of prevalence to 2050: A trend analysis for the Global Burden of Disease Study 2021. J. Affect. Disord. 2026, 405, 121531. [Google Scholar] [CrossRef]
- Akbari, H.; Korani, W.; Kumar, P.M.; Rostami, R.; Kazemi, R. Dynamical pattern of successive bits to predict the outcomes of the SSRI and rTMS depression therapies using EEG signals. Comput. Biol. Med. 2025, 196, 110500. [Google Scholar] [CrossRef] [PubMed]
- Chi, S.; Song, M.; Lee, J.h.; Ko, M.; Suh, S.i.; Lee, M.S. Prospective study on resting state functional connectivity in adolescents with major depressive disorder after antidepressant treatment. J. Psychiatr. Res. 2021, 142, 369–375. [Google Scholar] [CrossRef] [PubMed]
- Shahabi, M.S.; Shalbaf, A.; Maghsoudi, A. Prediction of drug response in major depressive disorder using ensemble of transfer learning with convolutional neural network based on EEG. Biocybern. Biomed. Eng. 2021, 41, 946–959. [Google Scholar] [CrossRef]
- Akbari, H.; Korani, W.; Rostami, R.; Kazemi, R. Predicting the outcome of depression therapies using a pattern of sixteen consecutive bits of EEG signals. Biomed. Signal Process. Control 2026, 113, 108808. [Google Scholar] [CrossRef]
- Shahabi, M.S.; Shalbaf, A.; Rostami, R.; Kazemi, R. A convolutional recurrent neural network with attention for response prediction to repetitive transcranial magnetic stimulation in major depressive disorder. Sci. Rep. 2023, 13, 10147. [Google Scholar] [CrossRef]
- Button, K.S.; Kounali, D.; Thomas, L.; Wiles, N.J.; Peters, T.; Welton, N.J.; Ades, A.; Lewis, G. Minimal clinically important difference on the Beck Depression Inventory-II according to the patient’s perspective. Psychol. Med. 2015, 45, 3269–3279. [Google Scholar] [CrossRef]
- Reeves, G.M.; Rohan, K.J.; Langenberg, P.; Snitker, S.; Postolache, T.T. Validation of BDI-II response and remission cut-points for assessment of seasonal affective disorder patients. J. Affect. Disord. 2012, 138, 123. [Google Scholar] [CrossRef] [PubMed]
- Wang, Y.P.; Gorenstein, C. Assessment of depression in medical patients: A systematic review of the utility of the Beck Depression Inventory-II. Clinics 2013, 68, 1274–1287. [Google Scholar] [CrossRef]
- Akbari, H.; Sadiq, M.T.; Rehman, A.U.; Ghazvini, M.; Naqvi, R.A.; Payan, M.; Bagheri, H.; Bagheri, H. Depression recognition based on the reconstruction of phase space of EEG signals and geometrical features. Appl. Acoust. 2021, 179, 108078. [Google Scholar] [CrossRef]
- Khare, S.K.; Blanes-Vidal, V.; Nadimi, E.S.; Acharya, U.R. Emotion recognition and artificial intelligence: A systematic review (2014–2023) and research recommendations. Inf. Fusion 2024, 102, 102019. [Google Scholar] [CrossRef]
- Sadiq, M.T.; Yousaf, A.; Siuly, S.; Almogren, A. Fast Fractional Fourier Transform-Aided Novel Graphical Approach for EEG Alcoholism Detection. Bioengineering 2024, 11, 464. [Google Scholar] [CrossRef] [PubMed]
- Mumtaz, W.; Xia, L.; Mohd Yasin, M.A.; Azhar Ali, S.S.; Malik, A.S. A wavelet-based technique to predict treatment outcome for major depressive disorder. PLoS ONE 2017, 12, e0171409. [Google Scholar] [CrossRef] [PubMed]
- Khodayari-Rostamabad, A.; Reilly, J.P.; Hasey, G.M.; De Bruin, H.; MacCrimmon, D.J. A machine learning approach using EEG data to predict response to SSRI treatment for major depressive disorder. Clin. Neurophysiol. 2013, 124, 1975–1985. [Google Scholar] [CrossRef]
- Zhdanov, A.; Atluri, S.; Wong, W.; Vaghei, Y.; Daskalakis, Z.J.; Blumberger, D.M.; Frey, B.N.; Giacobbe, P.; Lam, R.W.; Milev, R.; et al. Use of machine learning for predicting escitalopram treatment outcome from electroencephalography recordings in adult patients with depression. JAMA Netw. Open 2020, 3, e1918377. [Google Scholar] [CrossRef]
- Jaworska, N.; De la Salle, S.; Ibrahim, M.H.; Blier, P.; Knott, V. Leveraging machine learning approaches for predicting antidepressant treatment response using electroencephalography (EEG) and clinical data. Front. Psychiatry 2019, 9, 768. [Google Scholar] [CrossRef]
- Rajpurkar, P.; Yang, J.; Dass, N.; Vale, V.; Keller, A.S.; Irvin, J.; Taylor, Z.; Basu, S.; Ng, A.; Williams, L.M. Evaluation of a machine learning model based on pretreatment symptoms and electroencephalographic features to predict outcomes of antidepressant treatment in adults with depression: A prespecified secondary analysis of a randomized clinical trial. JAMA Netw. Open 2020, 3, e206653. [Google Scholar] [CrossRef]
- Wu, W.; Zhang, Y.; Jiang, J.; Lucas, M.V.; Fonzo, G.A.; Rolle, C.E.; Cooper, C.; Chin-Fatt, C.; Krepel, N.; Cornelssen, C.A.; et al. An electroencephalographic signature predicts antidepressant response in major depression. Nat. Biotechnol. 2020, 38, 439–447. [Google Scholar] [CrossRef]
- Minami, S.; Kato, M.; Ikeda, S.; Yoshimura, M.; Ueda, S.; Koshikawa, Y.; Takekita, Y.; Kinoshita, T.; Nishida, K. Association between the rostral anterior cingulate cortex and anterior insula in the salience network on response to antidepressants in major depressive disorder as revealed by isolated effective coherence. Neuropsychobiology 2022, 81, 475–483. [Google Scholar] [CrossRef]
- Shahabi, M.S.; Shalbaf, A. Prediction of Treatment Outcome in Major Depressive Disorder using Ensemble of Hybrid Transfer Learning and Long Short Term Memory based on EEG Signal. IEEE Trans. Cogn. Dev. Syst. 2022, 15, 1279–1288. [Google Scholar] [CrossRef]
- Mirjebreili, S.M.; Shalbaf, R.; Shalbaf, A. Prediction of treatment response in major depressive disorder using a hybrid of convolutional recurrent deep neural networks and effective connectivity based on EEG signal. Phys. Eng. Sci. Med. 2024, 47, 633–642. [Google Scholar] [CrossRef]
- Shahabi, M.S.; Shalbaf, A.; Nobakhsh, B.; Rostami, R.; Kazemi, R. Attention-based convolutional recurrent deep neural networks for the prediction of response to repetitive transcranial magnetic stimulation for major depressive disorder. Int. J. Neural Syst. 2023, 33, 2350007. [Google Scholar] [CrossRef]
- Shahabi, M.S.; Shalbaf, A.; Rostami, R. Prediction of response to repetitive transcranial magnetic stimulation for major depressive disorder using hybrid Convolutional recurrent neural networks and raw Electroencephalogram Signal. Cogn. Neurodyn. 2023, 17, 909–920. [Google Scholar] [CrossRef]
- Shahabi, M.S.; Nobakhsh, B.; Shalbaf, A.; Rostami, R.; Kazemi, R. Prediction of treatment outcome for repetitive transcranial magnetic stimulation in major depressive disorder using connectivity measures and ensemble of pre-trained deep learning models. Biomed. Signal Process. Control 2023, 85, 104822. [Google Scholar] [CrossRef]
- Arts, L.P.; Van den Broek, E.L. The fast continuous wavelet transformation (fCWT) for real-time, high-quality, noise-resistant time–frequency analysis. Nat. Comput. Sci. 2022, 2, 47–58. [Google Scholar] [CrossRef] [PubMed]
- Hazrati, H.; Daliri, M.R. Decoding covert visual attention of electroencephalography signals using continuous wavelet transform and deep learning approach. Sci. Rep. 2025, 15, 37503. [Google Scholar] [CrossRef] [PubMed]
- Dragomiretskiy, K.; Zosso, D. Variational mode decomposition. IEEE Trans. Signal Process. 2013, 62, 531–544. [Google Scholar] [CrossRef]
- Peng, Z.; Zhang, W.; Lang, Z.; Meng, G.; Chu, F. Time–frequency data fusion technique with application to vibration signal analysis. Mech. Syst. Signal Process. 2012, 29, 164–173. [Google Scholar] [CrossRef]
- Lampropoulos, G.A.; Thayaparan, T.; Xie, N. Fusion of time-frequency distributions and applications to radar signals. J. Electron. Imaging 2006, 15, 023007. [Google Scholar] [CrossRef]
- Bagherzadeh, S.; Norouzi, M.R.; Hampa, S.B.; Ghasri, A.; Kouroshi, P.T.; Hosseininasab, S.; Zadeh, M.A.G.; Nasrabadi, A.M. A subject-independent portable emotion recognition system using synchrosqueezing wavelet transform maps of EEG signals and ResNet-18. Biomed. Signal Process. Control 2024, 90, 105875. [Google Scholar] [CrossRef]
- Pechetti, S.; Rao, B.S. Optimized MobileNetV3: A deep learning-based Parkinson’s disease classification using fused images. PeerJ Comput. Sci. 2023, 9, e1702. [Google Scholar] [CrossRef]
- Yao, D.; Li, G.; Liu, H.; Yang, J. An intelligent method of roller bearing fault diagnosis and fault characteristic frequency visualization based on improved MobileNet V3. Meas. Sci. Technol. 2021, 32, 124009. [Google Scholar] [CrossRef]
- Bagherzadeh, S.; Shahabi, M.S.; Shalbaf, A. Detection of schizophrenia using hybrid of deep learning and brain effective connectivity image from electroencephalogram signal. Comput. Biol. Med. 2022, 146, 105570. [Google Scholar] [CrossRef]
- Awan, T.; Aftab, M.; Naveed, A.; Shahid, M.H.; Sajid, A.; Khan, K.B. Advanced EEG-based brain monitoring for early detection of harmful neural activities using EfficientNet architecture. Biomed. Signal Process. Control 2026, 112, 108360. [Google Scholar] [CrossRef]
- Wu, K.; Zhang, J.; Peng, H.; Liu, M.; Xiao, B.; Fu, J.; Yuan, L. Tinyvit: Fast pretraining distillation for small vision transformers. In Proceedings of the European Conference on Computer Vision; Springer: Berlin/Heidelberg, Germany, 2022; pp. 68–85. [Google Scholar]
- Kim, J.W.; Khan, A.U.; Banerjee, I. Systematic review of hybrid vision transformer architectures for radiological image analysis. J. Imaging Inform. Med. 2025, 38, 3248–3262. [Google Scholar] [CrossRef] [PubMed]
- Mehta, S.; Rastegari, M. Mobilevit: Light-weight, general-purpose, and mobile-friendly vision transformer. arXiv 2021, arXiv:2110.02178. [Google Scholar]
- Delorme, A.; Makeig, S. EEGLAB: An open source toolbox for analysis of single-trial EEG dynamics including independent component analysis. J. Neurosci. Methods 2004, 134, 9–21. [Google Scholar] [CrossRef]
- Shirodkar, V.; Edla, D.R.; Kumari, A. Generative Adversarial Networks for Motor Imagery Classification using Wavelet Packet Decomposition and Complex Morlet Transform. Multimed. Tools Appl. 2025, 84, 34377–34400. [Google Scholar] [CrossRef]
- Nazari, M.; Sakhaei, S.M. Variational mode extraction: A new efficient method to derive respiratory signals from ECG. IEEE J. Biomed. Health Inform. 2017, 22, 1059–1067. [Google Scholar] [CrossRef]
- Nazari, M.; Sakhaei, S.M. Successive variational mode decomposition. Signal Process. 2020, 174, 107610. [Google Scholar] [CrossRef]
- Nazari, M.; Korshøj, A.R.; ur Rehman, N. Multiscale dynamic graph signal analysis. Signal Process. 2024, 222, 109519. [Google Scholar] [CrossRef]
- Shahbakhti, M.; Beiramvand, M.; Nazari, M.; Broniec-Wójcik, A.; Augustyniak, P.; Rodrigues, A.S.; Wierzchon, M.; Marozas, V. VME-DWT: An efficient algorithm for detection and elimination of eye blink from short segments of single EEG channel. IEEE Trans. Neural Syst. Rehabil. Eng. 2021, 29, 408–417. [Google Scholar] [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2026 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license.














