Previous Article in Journal
Transformer-Based HER2 Scoring in Breast Cancer: Comparative Performance of a Foundation and a Lightweight Model
Previous Article in Special Issue
An Explainable Approach to Parkinson’s Diagnosis Using the Contrastive Explanation Method—CEM
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Enhancing Diagnostic Accuracy of Neurological Disorders Through Feature-Driven Multi-Class Classification with Machine Learning

by
Çiğdem Gülüzar Altıntop
Department of Biomedical Engineering, Faculty of Engineering, Erciyes University, Kayseri 38280, Türkiye
Diagnostics 2025, 15(17), 2132; https://doi.org/10.3390/diagnostics15172132 (registering DOI)
Submission received: 10 July 2025 / Revised: 17 August 2025 / Accepted: 21 August 2025 / Published: 23 August 2025
(This article belongs to the Special Issue Artificial Intelligence in Brain Diseases)

Abstract

Background/Objectives: Neurological disorders (ND) are a global health challenge, affecting millions and greatly reducing quality of life. Disorders such as Alzheimer’s disease, mild cognitive impairment (MCI), schizophrenia, and depression often share overlapping symptoms, complicating diagnosis and treatment. Early detection is crucial for timely intervention; however, traditional diagnostic methods rely on subjective assessments and costly imaging, which are not universally accessible. Addressing these challenges, this study investigates the classification of multiple ND using electroencephalography (EEG) signals. Methods: Various feature extraction methods were employed, and the Least Absolute Shrinkage and Selection Operator (Lasso) algorithm was utilized for effective feature selection. Two-class (disease–disease and healthy control–disease), three-class (healthy control and two ND, as well as three ND), and four-class (healthy control and three ND) classifications were conducted using different machine learning algorithms with the selected features. An EEG dataset comprising 40 Alzheimer’s patients, 43 healthy controls, 42 schizophrenia patients, 28 MCI patients, and 28 depression patients served as the experimental benchmark. Results: The Linear Discriminant Analysis (LDA) classifier achieved the highest accuracy, distinguishing between healthy controls and Alzheimer’s with 100% accuracy and demonstrating strong performance in other comparisons. Multi-class classification reached 84.67% accuracy for distinguishing depression, MCI, and schizophrenia, while four-class classification achieved 57.89%, highlighting the complexity of differentiating among multiple ND. The frequent selection of frontal lobe channels across ND indicates their critical role in classification. Conclusions: This study contributes to the literature by emphasizing disease-to-disease classification over the traditional control-versus-patient framework, highlighting the potential for more effective diagnostic tools in clinical settings.

1. Introduction

Neurological disorders (ND) encompass both central and peripheral nervous system diseases, including neurodevelopmental, neurodegenerative, and psychiatric conditions [1]. NDs are a primary cause of mortality and impaired quality of life worldwide. Early diagnosis can help reduce the course of many illnesses, if not totally eliminate them [2]. Common ND include Parkinson’s disease (PD), epilepsy, mild cognitive impairment (MCI), schizophrenia, and Alzheimer’s disease (AD), as well as cerebrovascular diseases like stroke, brain tumors, and developmental disorders like autism and attention deficit hyperactivity disorder (ADHD) [1]. In addition, mood disorders such as depression are regarded as major ND due to their substantial influence on cognitive and emotional processes. Depression, which is frequently associated with disorders such as Alzheimer’s and schizophrenia, is increasingly being explored for its neurological roots and consequences on brain function and structure [3].
These disorders can cause major complications, including memory loss and neurological malfunction, and have a significant impact on patients’ and families’ daily lives. As a result, early diagnosis is essential for timely treatment. However, early detection of diseases might be challenging. Current diagnostic methods, including clinical examination, neuropsychological assessment, and neuroimaging (MRI, fMRI, PET), are effective but often costly, time-consuming, and dependent on specialized expertise [4,5]. In regions with limited access to specialists, delays in diagnosis are common. Additionally, some imaging procedures involve invasive or radioactive materials, posing further risks. As a result, electroencephalography (EEG) has emerged as an attractive alternative due to its non-invasive nature, high temporal resolution, portability, and low cost [5,6]. EEG records brain electrical activity and can capture functional abnormalities across neurological and psychiatric disorders [7]. However, manual EEG interpretation is labor-intensive, prone to subjectivity, and challenged by the signal’s low signal-to-noise ratio and complexity [8].
To overcome these challenges, computer-aided diagnosis (CAD) systems using machine learning (ML) have been developed to automate EEG analysis. ML algorithms can identify patterns that may not be apparent to human observers, improving diagnostic accuracy and reducing interpretation time. Few research have explored developing a unified framework for multi-class classification of different ND. Most recent EEG-based studies have focused on binary classification of one ND (epilepsy [9], schizophrenia [10], AD [11], MCI [12], depression [13]) versus healthy controls. Deep learning methods, such as convolutional neural networks (CNNs) and recurrent neural networks (RNNs), have also been applied to EEG-based ND classification [14,15,16]. While these approaches can capture complex hierarchical patterns, they require large datasets for optimal performance, demand high computational resources, and often act as “black boxes” with limited interpretability [17]. These limitations make traditional ML approaches with engineered features more practical in many clinical settings, especially when datasets are small or medium-sized.
Table 1 summarizes representative EEG-based classification studies, highlighting the diversity in disorders, features, and classifiers, as well as the performance achieved.
Although many binary classification studies have achieved high performance, multi-class classification—which allows the simultaneous differentiation of multiple disorders—remains less explored. This limitation necessitates multiple CAD systems to cover different diseases, increasing cost and complexity. Most multi-class EEG studies are limited to three-class classification, such as AD vs. MCI vs. controls [24,25,26,27,28] or focus on specific neuropsychiatric disorders like depression and schizophrenia [29,30,31,32,33]. Disease-to-disease classification (e.g., AD vs. schizophrenia or AD vs. depression) is particularly rare. Nevertheless, to the best of my knowledge, there have not been any EEG-based studies that compare or classify AD and depression within a complete machine learning framework. Addressing these gaps might provide more understanding of the unique characteristics and similarities across these diseases, perhaps leading to more effective diagnostic and treatment methods. A unified framework for binary and multi-class classification of multiple NDs could improve diagnostic specificity, reduce cost, and enhance clinical applicability.
This work addresses this gap by proposing an ML framework for binary, three-class, and four-class classification of AD, depression, schizophrenia, MCI, and healthy controls. Features from time, frequency, entropy, and complexity measures were extracted, feature selection was applied to identify the most informative EEG channels, and performance was evaluated across multiple classifiers. Unlike studies in the literature, a multi-class classification of multiple disorders was proposed within the same framework. Beyond classification, it is investigated whether certain EEG channels could serve as potential biomarkers for these disorders. A cost-effective, interpretable, and scalable CAD system suitable for clinical environments with limited resources is aimed at being provided by this approach.

2. Methodology

In this study, EEG signals from four different ND and healthy controls were preprocessed, followed by feature extraction. After feature selection, two-class (control vs. disease, disease vs. disease), three-class (control vs. two diseases, three diseases), and four-class (control vs. three diseases, four diseases) classifications were conducted using various machine learning algorithms. Figure 1 illustrates the overall research framework.

2.1. Dataset

In this study, a publicly available dataset [34] containing EEG signals from AD, schizophrenia, MCI, depression patients, and control groups was used. EEG recordings were conducted in a standard clinical environment by a trained technician. Each participant underwent EEG monitoring between 8 AM and 1 PM, using a Nihon Kohden device with 19 electrodes arranged according to the international 10–20 system, with a sampling rate of 500 Hz. During the procedure, participants alternated between periods of rest with their eyes open and closed. Individuals who had undergone sleep EEGs were excluded from the study group. In the dataset, there were 230 subjects, including 28 with major depression, 42 with schizophrenia, 68 with cognitive impairment (40 AD, 28 MCI), and 95 controls. To ensure a balanced classification of the dataset, EEG signals of 43 randomly selected healthy controls were used. Table 2 presents demographic information for each neurological disease and for healthy controls.
This study was conducted without applying data augmentation techniques or generating synthetic samples. All analyses were performed exclusively on authentic EEG recordings to ensure that the results reflect genuine neurophysiological patterns.

2.2. EEG Signals and Preprocessing

The amplitude of EEG recordings might vary depending on factors such as electrode location, skin impedance, and equipment variability. Normalization ensures that these differences do not affect the analysis, allowing for more consistent comparisons between subjects. Hence, the EEG signals were first normalized to a range between 0 and 1. To eliminate mains interference, a 50 Hz notch filter was applied. Afterward, a Butterworth bandpass filter [1–30 Hz bandpass] was used to filter the signals. The 10–20 international system is illustrated in Figure 2, while the specific electrode placements for each channel are detailed in Table 3. In Figure 3, the filtered EEG signals of the F7 channel are shown for a period of 1 min.

2.3. Feature Extraction

In the study, various features were extracted from EEG signals in the time axis and frequency axis to examine their distinctiveness in ND. While the features extracted in the time axis consist of statistical calculations, entropy values of different algorithms, and complexity measurements, calculations in the frequency axis are based on the results obtained from the power spectral density (PSD) analysis of frequency values and power values. PSD is a method for examining the frequency characteristics of a signal by evaluating how its power is distributed across different frequency bands [35]. This approach is particularly useful in EEG analysis, where it helps to identify dominant brainwave frequencies that are associated with various neurological states. By using the Fourier transform, PSD allows for a detailed analysis of the signal’s frequency content. Choosing an appropriate window length is essential to balance the trade-off between frequency resolution and variance in the estimation process [35]. The window length was heuristically selected at 512 after testing different values to achieve an optimal balance between frequency resolution and spectral estimation stability. Table 4 is provided to group the extracted features and briefly explain them with their abbreviations.

2.3.1. Entropy Measures

In this study, instead of calculating a single entropy value, entropy values with different methodologies were calculated. Since entropy calculations are sensitive to length, calculating entropy by dividing the signal into certain windows and averaging the entropy matrix is a more accurate approach. Short windows may not accurately reflect slower processes, but lengthy windows may ignore faster alterations, affecting entropy estimates. For example, in EEG analysis, selecting a window that is too small might underestimate signal complexity, whereas an overly long window may smooth out crucial characteristics. Therefore, selecting a suitable window size is crucial for reliable entropy measurement [36,37]. In this study, after examining many studies, entropy calculation was performed for every two seconds of data (1000 samples) [38,39,40]. The entropy measurements are presented below with their formulations.
Approximation Entropy (ApEn) [41] is a measure of signal complexity, particularly resistant to low-frequency noise. Higher ApEn values suggest greater complexity. Given a time series XN with N samples, subsets pm(i) are formed using m samples starting at the i-th position. Two subsets pm(i) and pm(j), are considered similar if the Euclidean distance between them is less than a threshold r. The set pm represents all patterns of length m within XN. The similarity calculation is formalized in Equations (1) and (2).
X i + k X j + k < r ,   0 k m
A p E n m , r , N = I n C m ( r ) C m + 1 ( r )
C i , m r = n i , m ( r ) N m + 1
ni,m(r) represents the number of patterns in the set pm that are similar to pm(i), according to the similarity coefficient r. This calculation is performed for each pattern in pm. Cm(r) is the average of all Ci,m(r) values (see Equation (3)). For ApEn calculations, the parameters are usually chosen as m = 2 or m = 3, and r = 0.15 × σX (σX is the standard deviation of the original time series), or r = 0.25 × σX or r = 0.2 × σX. In this study, the embedding dimension is set to m = 2 and the similarity coefficient to r = 0.25 × σX.
Logenergy entropy (logenergy) [42] is predicted based on the energy present in the signal, which can result in higher entropy values compared to other algorithms. For a data series X with N samples, the energy of the signal and the logenergy value are calculated as follows:
l o g e n e r g y ( x i ) = log ( x i 2 ) l o g e n e r g y ( x ) = i = 1 n log ( x i 2 )
Unlike other entropy methods, Permutation Entropy (PermEn) [43] relies solely on the order of the time series amplitudes. This makes PermEn more computationally efficient [44]. The PermEn calculation is described by Equation (5).
P e r m E n = i = 1 N ( m 1 ) τ p i π log 2 p i π
In Equation (5), p(π) represents the relative frequency of each permutation π, with the embedding dimension m and lag τ. For the PermEn calculation, the embedding dimension was empirically set to six, and the lag was set to one. The PermEn value is 1 when all permutations are equally probable and decreases if the time series exhibits regularity.
Renyi entropy (renyi) is an extension of Shannon Entropy [45]. The definition of renyi with a specific order is given in Equation (6). In this context, pi denotes the probability of the time series, and α ≠ 1 indicates the order. When α = 1, Renyi entropy simplifies to Shannon entropy. Renyi entropy remains consistent across various density functions.
r e n y i x = 1 log 2 p i
Tsallis entropy [46] for an EEG signal is derived from a generalized entropy measure. In Equation (7), k represents the length of the sub-dataset, pi is the probability function, and q is the Tsallis parameter (with k = 1000 and q = 2).
t s a l l i s q x = i = 1 k ( P i P i q ) q 1
Spectral entropy (spentropy) quantifies the unpredictability or complexity of power distribution across distinct frequency bands in a signal’s spectrum, providing information on the signal’s spectral properties. Spectral Entropy involves substituting the probability density function in Equation (5) with the power spectral density of the signal. It is calculated as follows
s p e n t r o p y F i = i = f l f h F i   log ( F i   )
where Fi represents the normalized power in the i-th frequency band, and fl and fh represent the frequency bands.

2.3.2. Hjorth Parameters

Hjorth parameters, introduced by Hjorth [47], include activity, mobility, and complexity. The activity parameter reflects the signal’s overall power: higher activity values indicate greater amounts of high-frequency components. Mobility is related to the standard deviation of the power spectrum, measuring how power distribution varies. Complexity assesses how much the signal deviates from a pure sine wave, with a maximum value of 1 indicating perfect similarity. The formulas for calculating Hjorth parameters are detailed in [47]. In this study, Hjorth parameters were calculated separately for the entire signal and for one-minute windows.

2.3.3. Complexity Measures

Kolmogorov complexity [48], often known as algorithmic complexity, is the amount of information needed to describe a string or dataset in the most compact manner. It is defined as the length of the shortest program (in a certain programming language) that produces the supplied text when executed. Essentially, it measures the complexity of a string based on the shortest feasible explanation, indicating how intricate or simple the data is.
Lempel–Ziv complexity [49] measures the complexity of a sequence by measuring the number of unique substrings or patterns that may be seen while the sequence is processed gradually. It measures the sequence’s richness or regularity, indicating how much new information or variety is introduced as the sequence progresses. To calculate the Lempel value, the signal is first converted to a binary data array as in Equation (9). The threshold value T is determined by calculating the median or mean value of the time series (x(n)). In this study, the median value is used. Because the median is commonly selected as the threshold (T) due to its resistance to the effects of outliers. Its ability to remain stable despite extreme values makes it a reliable choice for setting thresholds [50].
s n = 1 ,     i f   x n > T , 0 ,     i f   x n T ,
All features, except for frequency axis measures (freqmax, powmax), were calculated for signals divided into 2-s windows. The mean and variance of the resulting matrices were evaluated for each calculation (yielding the final feature value). Additionally, Hjorth parameters, kurtosis, and skewness values were obtained for 60-s windows across all signals. As a result, a total of 646 features were obtained (19 EEG channels × 34 features).

2.4. Feature Selection

In recent years, the volume of high-dimensional data accessible online has surged significantly. As a result, machine learning techniques face challenges in managing the vast array of input features, presenting an intriguing problem for researchers. To apply machine learning algorithms efficiently, data preprocessing is crucial. Among the various preprocessing techniques, feature selection stands out as one of the most vital and widely used methods, playing a key role in the success of machine learning algorithms [51]. Feature selection methods are grouped into three groups based on evaluation criteria and relation to the learning algorithm as the filtering technique, the Wrapper approach, and the embedded method [51]. Although each category has advantages and disadvantages, feature selection methods from each category have been used in biomedical image analysis and biomedical signal processing [52]. In recent research, mutual information-based feature selection [53], Recursive Feature Elimination (RFE) method [54], correlation-based feature selection (CBFS) [55], Kruskal–Wallis (KW) [56], and ReliefF [57] method are some of the methods used in the selection of features extracted from ND EEG signals. Since researchers continue to explore which feature selection method can be used and apply feature selection techniques tailored to the unique characteristics of each dataset, it suggests that the choice of method is largely based on experimentation. In this research, the Least Absolute Shrinkage and Selection Operator (Lasso) algorithm was chosen for feature selection through empirical testing. Lasso regularization [56] is an embedding approach based on the ℓ1-norm of a linear classifier’s coefficient in Equation (10).
β ^ = m i n β c ( β , X ) + λ   i = 1 M β i
where i = 1 M β i is regularization term, M is the number of features, λ shrinkage (regularization) parameter, and c(.) is the classification objective function.

2.5. Statistical Analysis

Selecting an appropriate statistical approach is crucial for analyzing biological data. Selecting the improper statistical approach might lead to interpretation issues and negatively impact study conclusions. For features that do not follow a normal distribution, non-parametric tests should be used, while parametric statistical tests are appropriate for features that follow a normal distribution [58]. Student’s t-test is applied to compare the means of two groups, whereas the ANOVA test, an extension of the t-test, is employed to compare the means across three or more groups. Non-parametric alternatives to these parametric methods exist as well. For instance, the Mann–Whitney U test serves as a substitute for the t-test, and the Kruskal–Wallis H test is the non-parametric counterpart to ANOVA [58]. Dunn’s test with Bonferroni correction was applied for non-parametric pairwise comparisons between independent groups to identify significant differences among ND. The Bonferroni adjustment is frequently used to control the family-wise error rate. For parametric pairwise comparisons, Tukey’s test was utilized to detect significant differences in ND. Tukey’s test is widely used in biomedical research and is especially appropriate when group sizes are balanced [59]. In this study, the R statistics software (R 4.4.2 version) was used for statistical analysis processes.

2.6. Machine Learning Algorithms

In this research, features derived from EEG signals were classified using several ML models, including SVM, LDA, ANN, and RF algorithms. The classification process was conducted separately for two-class (healthy control vs. disease, disease vs. disease), three-class, and four-class classification to assess their ability to distinguish ND. The classifiers were trained using features selected through the Lasso method. The model performance was evaluated using 10-fold cross-validation. In the 10-fold cross-validation, 90% of the data was used for training, while the remaining 10% served as the test set, with the procedure repeated for each fold. Care was taken to maintain class balance during k-fold cross-validation. Preserving class proportions in each fold helps reduce the risk of overfitting to certain classes, especially in small datasets. Performance metrics were calculated at each iteration.
SVM is a widely used non-probabilistic classifier that operates by finding a decision boundary, or hyperplane, that maximizes the margin between support vectors belonging to different classes. For linearly separable data, SVM uses hyperplanes in an N-dimensional space to distinguish between different feature groups. Kernel functions are employed when the data is not linearly separable, mapping it to a higher-dimensional space to facilitate classification [60].
RF is an ensemble learning technique that uses numerous decision trees to increase accuracy while minimizing overfitting. It works by training several trees on random samples of data and averaging their predictions, making it reliable and effective for classification and regression applications [61]. LDA is a statistical technique for dimensionality reduction and classification. It selects a linear combination of characteristics that best distinguishes several classes by maximizing the distance between the means of distinct classes while minimizing variation within each [62].
ANN consists of three key layers: the input layer, one or more hidden layers, and the output layer. The number of neurons in the input layer is equal to the number of features representing the classified objects, while the output layer’s neurons correspond to the total number of classes [63]. In a fully connected backpropagation network, each neuron in the hidden and output layers is connected to every neuron in the preceding layer through a set of numerical weights. The training process allows the network to learn the relationship between inputs and outputs, using backpropagation as a common supervised learning technique. Training data includes pairs of inputs and their associated outputs, and the network adjusts its weights and biases to minimize prediction errors on the training set [63].
In this study, the MATLAB (2024a version) program was used for feature extraction, feature selection, and the classification process.

2.7. Evaluation Metrics

This study’s evaluation metrics include mostly used metrics: accuracy, sensitivity, specificity, precision, F-score, and Area Under the Curve (AUC) for the Receiver Operating Characteristic (ROC) curve [64].

3. Experimental Results

In this section, the results of feature selection, statistical analysis, and classification studies are given.

3.1. Feature Selection Results

The Lasso algorithm was used for feature selection to reduce the data burden and improve classification performance. However, determining the optimal value of the lambda (λ) parameter in Lasso poses a challenge. Therefore, the lambda value was iteratively changed starting from 0.001 up to 0.05, and the lambda value that yielded the highest classification accuracy was selected, as seen in Figure 4. So, the shrinkage parameter (λ) was examined with 10-fold cross-validation, and the value with the lowest cross-validation loss (λ = 0.001) was chosen. Information about the features selected for binary classification was given in Figure 5. The number of features selected for two-class classification (binary classification) was given in Table 5.
In Figure 6, the numbers of features selected for binary classification according to channels were given. In the examination of EEG channel selections for different ND, specific channels emerged for each disease. For AD, channel F7 had the highest number of selected features (9), followed by O2 (8 features), indicating its importance in distinguishing this disorder. In depression, Fp2 and Fz showed significant relevance, with Fp2 having 9 features and Fz having 10 features. For MCI, P4 was the most significant channel with 8 selected features, followed by Fp1 and Fz, both showing 7 features. Finally, in schizophrenia, F7 again had the highest feature count with 10 selected features, followed by Fp2 and F4 (9 and 7 features). These channels highlight critical brain regions for classifying these disorders.

3.2. Statistical Analysis Results

The discrimination of features among the five groups (healthy controls, AD, depression, MCI, and schizophrenia) was assessed using both ANOVA and Kruskal–Wallis tests. Following these tests, post hoc analyses were conducted to determine which groups differed significantly from each other. The detailed results, including specific group comparisons and statistical significance, are provided in Supplementary Files for further reference. As seen in Supplementary File S1, 239 of the features yielded results that created a significant difference between the groups (p < 0.05). The post hoc analysis revealed that AD exhibited significant differences 184 times compared to other groups. The control group showed differences 151 times, depression in 119, schizophrenia in 52, and MCI in 28. These findings highlight the varying degrees of distinction between the conditions across different features, which are detailed further in the Supplementary File S1. Box-plot graphs for complexity and spentropyDyd, which are some of the features that reveal the most differences between the groups, were given in Figure 7. As seen in Figure 7, the complexity feature of the 8th EEG channel created a significant difference between the AD–depression, depression–control, depression–MCI, and depression–schizophrenia groups. The post hoc analysis results for the 4 features in Figure 7 according to the Tukey test were given in Supplementary File S2.

3.3. Parameter Tuning and Model Performance Timings

In the parameter tuning process for machine learning algorithms, a Bayesian optimization approach has been employed to efficiently explore the hyperparameter space. To ensure robust evaluation, a 10-fold cross-validation method was used to partition the data into training and test sets. Training and testing times of the algorithms were recorded. This approach not only provided a comprehensive assessment of model performance but also enabled a detailed comparison of algorithms in terms of their training and testing durations. By analyzing these timings, insights were gained into the computational efficiency of each algorithm, highlighting the trade-offs between parameter optimization and execution speed. As shown in Table 6, the running times of the algorithms for two-class classification were evaluated to compare their computational efficiency.
Upon analyzing the training times, it is evident that RF has the shortest training time across all diseases, taking less than a second, which is significantly quicker than the other algorithms like SVM, LDA, and ANN. On the other hand, LDA also shows relatively fast training times, though slightly longer than RF. ANN, while exhibiting the slowest training times, particularly for AD, outperforms others in testing time, showing consistently faster results in comparison to SVM, LDA, and RF, especially for quick evaluations. These results suggest that LDA is efficient for both training and testing, while ANN takes longer to train but is very fast in testing.

3.4. Classification Results

In this study, the performance of various classification algorithms (SVM, LDA, ANN, RF) for the classification of ND using features extracted from EEG signals was evaluated with 10-fold cross-validation. Classification results are presented in tables as mean ± standard deviation of each fold of 10-fold cross-validation. In Table 7, the performance measures of two-class classification (binary classification) according to control–disease classification were given. Table 8 shows the results of binary classification for disease–disease classification. The classification results for control–disease and disease–disease comparisons reveal varying levels of effectiveness across machine learning models like SVM, LDA, ANN, and RF. In the control–AD classification, LDA performed exceptionally well, achieving perfect scores for accuracy (1.0000 ± 0.0000), sensitivity, specificity, precision, F1 score, and AUC. This indicates that the model is perfectly classified between the control and AD groups across all metrics. In contrast, RF performed significantly worse, with lower accuracy (0.7208 ± 0.1498), and only moderate AUC (0.7225 ± 0.1474). This pattern highlights LDA’s robustness in distinguishing AD compared to other models. For the control–depression comparison, SVM delivered a high accuracy of 0.9589 ± 0.0945 and an impressive AUC of 0.9917 ± 0.0264, making it highly reliable in this classification. LDA also showed excellent results, with an accuracy of 0.9857 ± 0.0452 and similar performance in sensitivity and specificity. RF again underperformed with an accuracy of 0.7339 ± 0.1186, showing reduced sensitivity but moderate specificity (0.8550 ± 0.2088). The results emphasize how LDA and SVM succeed in capturing both the true positives (sensitivity) and true negatives (specificity) in the depression classification. In the control–schizophrenia comparison, both LDA and SVM provided consistently strong performances. LDA achieved an accuracy of 0.9778 ± 0.0468, and SVM had 0.9403 ± 0.1010, with high AUC values for both models (0.9800 ± 0.0422 and 0.9738 ± 0.0641, respectively). Meanwhile, RF continued to lag behind, offering an accuracy of 0.7278 ± 0.1386, reflecting its challenges in accurately distinguishing schizophrenia from the control group. In the control–MCI classification, LDA maintained strong performance with an accuracy of 0.9714 ± 0.0602, indicating its high reliability in distinguishing control subjects from MCI patients. SVM and ANN followed closely with the same accuracy of 0.9571, and SVM had a high AUC value (1.0000 ± 0.0000). In comparison, RF lagged behind, showing lower accuracy (0.7304 ± 0.1589) and reduced sensitivity, again confirming its relative inefficiency in handling neurological disorder classifications. Overall, LDA excelled in this control–MCI distinction. These findings underline LDA’s superior generalizability across various ND.
In the disease–disease classification, particularly for the AD–MCI comparison, LDA remained highly effective with 0.9690 ± 0.0655 accuracy, further supported by perfect specificity and precision (1.0000 ± 0.0000). Similarly, SVM and ANN yielded high accuracies (0.9262 ± 0.0781) but still underperformed compared to LDA. RF, once again, demonstrated lower efficiency, showing variability across metrics, particularly in specificity and AUC. For the AD–schizophrenia classification, LDA performed well with 0.9278 ± 0.0624 accuracy and balanced sensitivity/specificity. Although SVM showed an accuracy of 0.8667 ± 0.1220, the model’s AUC was still relatively strong (0.9238 ± 0.1256). ANN outperformed SVM with an accuracy of 0.9028 ± 0.1143. RF, however, had the lowest accuracy (0.7181 ± 0.1058), confirming the trend of reduced effectiveness compared to other methods. For the depression–AD classification, LDA achieved an accuracy of 0.9429 ± 0.0738, demonstrating solid performance across most metrics. SVM followed with an accuracy of 0.8810 ± 0.0635, and ANN closely matched with 0.8238 ± 0.1118, showing comparable AUCs. RF, however, had lower accuracy at 0.7333 ± 0.1830, confirming its relative inefficiency. In the depression–MCI classification, SVM achieved the highest performance with an accuracy of 0.9667 ± 0.0703, displaying excellent consistency across other metrics like specificity and precision (0.9667 ± 0.1054 for both). LDA followed closely with an accuracy of 0.9633 ± 0.0777, showing perfect precision and specificity (1.0000 ± 0.0000). ANN performed comparably with an accuracy of 0.9467 ± 0.0864, while RF struggled with an accuracy of 0.6700 ± 0.2457, highlighting its lower effectiveness in this classification task. In the schizophrenia–MCI classification, LDA outperformed other models with an accuracy of 0.9571 ± 0.0690, along with strong specificity and precision (0.9750 ± 0.0791 for both). SVM showed a slightly lower accuracy of 0.8714 ± 0.1251 but still delivered reliable precision (0.9000 ± 0.1748). ANN underperformed compared to the others, with an accuracy of 0.7857 ± 0.2156, reflecting its variability. RF also showed moderate performance with an accuracy of 0.7429 ± 0.1622, lagging behind in both sensitivity and specificity compared to LDA and SVM. Similarly, in the depression–schizophrenia comparison, LDA excelled, showing an accuracy of 0.9571 ± 0.0690 and nearly perfect AUC, indicating a strong ability to differentiate between these disorders. RF’s performance remained significantly lower, with an accuracy of 0.6857 ± 0.1475 and sensitivity of 0.4333 ± 0.3063, confirming its struggle with disease-to-disease distinctions.
Overall, LDA consistently delivered the highest classification performance in both control–disease and disease–disease comparisons, across multiple ND. SVM also performed well, especially in distinguishing depression and AD from the control group. ANN, similarly to SVM, demonstrated effective performance in most cases, offering a viable alternative for classification. RF, on the other hand, displayed the weakest performance in most cases, with lower accuracy, sensitivity, and AUC scores, suggesting that it may not be as suitable for these medical classifications. These results indicate that LDA and SVM are generally better suited for binary classifications involving ND, while RF may require optimization to improve its performance in these contexts.
In the disease–disease classification, AD-MCI was the best-performing classification, with LDA achieving the highest accuracy of 0.9690 ± 0.0655. This result was reinforced by perfect specificity and precision, making LDA the most reliable model in this context. Similarly, for the Depression-MCI comparison, SVM stood out with an accuracy of 0.9667 ± 0.0703, demonstrating its consistency across various metrics like sensitivity, specificity, and AUC, indicating its strong capability in differentiating between these two conditions effectively.
Table 9 and Table 10 present the results of the three-class classification using the LDA algorithm. Since the LDA algorithm is more successful than the others, only the results of the classification with LDA were given for the 3-class and 4-class classification. In the three-class classification results for the LDA algorithm, control–MCI–AD achieved the highest accuracy at 0.7932 ± 0.1041, along with strong sensitivity and specificity values. Among the disorder-only groups, the depression–MCI–schizophrenia classification performed best, with 0.8467 ± 0.1088 accuracy and a high AUC of 0.9128 ± 0.1090, indicating effective separation between these three disorders. In contrast, the depression–schizophrenia–AD group showed the lowest performance, with accuracy dropping to 0.6455 ± 0.1088 and lower precision and F-score values. In Table 9 and Table 10, the values highlighted in red correspond to the best-performing results among the evaluated algorithms.
Table 11 presents the results of the four-class classification using the LDA algorithm. The classification accuracy varied across different combinations of classes. The highest accuracy was observed in the control–depression–schizophrenia–cognitive decline classification, achieving an accuracy of 0.5789 (±0.1309). This was followed closely by the control–depression–schizophrenia–MCI classification, with an accuracy of 0.5605 (±0.1410). The control–depression–schizophrenia–AD classification exhibited a lower accuracy of 0.5146 (±0.1148), indicating challenges in distinguishing between these conditions. The lowest accuracy was recorded in the depression–schizophrenia–AD–MCI classification, at 0.4857 (±0.1171). Across all classifications for four-class, sensitivity, specificity, precision, and F-score values also reflected the varying degrees of model performance, highlighting the complexities in multi-class classification tasks.
The confusion matrices in Figure 8 illustrate the classification performance of the LDA model across four experimental scenarios. In the control–AD–MCI task (a), the control class achieved the highest correct classification rate, whereas partial misclassification occurred between AD and MCI. In the depression–MCI–schizophrenia task (b), schizophrenia was identified with high accuracy, while MCI and depression exhibited greater overlap. The four-class control–depression–schizophrenia–MCI scenario (c) proved more challenging, with notable misclassification between depression and schizophrenia. Similarly, in the control–depression–schizophrenia–cognitive decline scenario (d), confusion was observed between Control and cognitive decline, as well as between depression and schizophrenia. Overall, multi-class configurations with more than three categories showed increased misclassification rates, reflecting the higher complexity and feature overlap among certain neurological disorders.

4. Discussion

Clinical EEG diagnosis presents several challenges: firstly, the accuracy of diagnoses is largely contingent upon the expertise of highly trained EEG specialists. Secondly, acquiring the skills necessary to interpret EEG recordings requires extensive pathological education over several years. Lastly, the process of analyzing EEG data is labor-intensive and can be both time-consuming and mentally taxing [65]. Therefore, there is a growing need for a computer-based system capable of automatically diagnosing multiple ND, which would streamline the process, reduce the burden on specialists, and improve diagnostic accuracy.
Very limited research on ND uses nonlinear EEG analytic techniques. In most studies, linear approaches such as power spectral density and frequency analysis have been used more often. Nonlinear approaches, on the other hand, are gaining popularity due to their ability to find more complicated dynamics in brain signals, providing crucial insights that linear methods may overlook, particularly when discriminating between various neurological diseases. The EEG signals may also be analyzed in a nonlinear way, which makes it possible to obtain an understanding that is not given by the linear measures. For example, Higuchi’s fractal dimension (HFD) has been shown to reflect higher complexity inside depressed patients as compared to healthy control individuals in all brain areas. Such studies have reported such significant success in classifying depression that an enhanced probabilistic neural network model has achieved 91.3% accuracy based on seven frontal EEG channels, indicating its possible applicability in the clinical field for mental health diagnostics [66]. Also, The Lempel–Ziv complexity (LZC) of multi-channel resting EEG has been shown to be effective in evaluating a variety of neurological and mental diseases, including serious depression [67,68].
Numerous studies show a significant correlation between cognitive decline and the reduction in EEG irregularity [40,69]. For this reason, in this study, both frequency axis and time axis (entropy, statistical measures) features are extracted from EEG signals. Although previous research suggests a framework for multi-class EEG classification, to the best of my knowledge, no studies have used the machine learning model to classify more than two diseases in healthy people and disease–disease classification. The aim of this study is to develop a system capable of automatically detecting four ND—Alzheimer, schizophrenia, depression, and MCI—using EEG signals, improving diagnostic speed and accuracy. Following this context, some studies in the literature are shown in Table 12.
Comparing the present results with selected studies in Table 12 the proposed LDA model achieved notably higher accuracy in several cases. For instance, in the classification of schizophrenia and depression, this study reached 95.71% accuracy with 95% sensitivity and specificity, outperforming the results of Jang et al. [71]. In the work of Wang et al. [33], a 10-fold cross-validation approach with LDA yielded 74.32% accuracy, which is lower than the performance observed in the present study for similar classification tasks. Likewise, when compared with the results of Hassanzadeh et al. [73], the proposed method achieved 92.78% accuracy in the same two-class classification and 76.03% accuracy in the equivalent three-class classification, both exceeding the accuracies reported in their work.
Also, in this study, the LDA classifier produced accuracies of 100% for control–AD, 97.14% for control–MCI, 96.90% for MCI-AD, and 79.32% for 3-class classification (AD–MCI–control), respectively. This performance is consistent with previous EEG-based ND classification studies, where accuracies typically ranged between 70% and 98% depending on the dataset size, feature set, and classification task. For example, Huang et al. [70] reported 88.2% accuracy in distinguishing depression from healthy controls using KNN after Lasso-based feature selection. Wang et al. [33] achieved 79.27% accuracy in a three-class classification of schizophrenia, depression, and healthy controls using a convolutional neural network (MUCHf-Net). Cheng et al. [72] applied dynamic functional connectivity features with a random forest model to classify four groups (nonpsychotic major depression, psychotic major depression, schizophrenia, and healthy controls), obtaining 73.1% accuracy. Compared to these works, the present study’s LDA approach demonstrated competitive or superior performance in both binary and multi-class settings while using a broader classification framework encompassing multiple ND combinations.
According to Table 7 and Table 8, SVM also performed strongly, particularly in control–depression classification, with an accuracy of 0.9589 ± 0.0945 and an AUC of 0.9917 ± 0.0264. In the control–schizophrenia classification, the ANN method outperforms SVM. Specifically, ANN achieves an accuracy of 0.9639 ± 0.0583, while SVM shows an accuracy of 0.9403 ± 0.1010. However, RF exhibited weaker results, with its accuracy ranging from 0.6700 to 0.7524, depending on the comparison, often underperforming in sensitivity and AUC scores.
In the disease–disease comparisons, similar trends were observed. LDA consistently achieved high accuracy, sensitivity, and specificity values, especially in the Alzheimer–MCI comparison (accuracy of 0.9690 ± 0.0655, AUC of 0.9750 ± 0.0527). SVM also showed solid performance, with accuracy values exceeding 0.90 in most comparisons. Although the ANN method generally performs well, it falls short of LDA in some classifications, particularly in the depression–MCI comparison, where ANN shows an impressive accuracy of 0.9467 ± 0.0864. However, RF again struggled, with accuracy values ranging from 0.6857 to 0.7524 and often demonstrating lower sensitivity and AUC values compared to LDA, ANN, and SVM.
The results from Table 9 and Table 10 demonstrate the performance of the LDA algorithm in three-class classification tasks involving various ND. In the control–disease–disease classification (Table 9), the highest accuracy was observed in the control–MCI–Alzheimer classification, with an accuracy of 0.7932 ± 0.1041, highlighting the model’s strength in distinguishing between these conditions. Conversely, the lowest performance was seen in the control–depression–MCI classification, with an accuracy of 0.7114 ± 0.1513. In the three disorders comparison (Table 10), the depression–MCI–schizophrenia classification yielded the highest accuracy of 0.8467 ± 0.1088, indicating effective differentiation among these disorders.
The results of the four-class classification indicate variability in the LDA algorithm’s performance across different combinations. The highest accuracy was achieved in the control–depression–schizophrenia–cognitive decline classification (57.89%), while the lowest accuracy was found in the depression–schizophrenia–Alzheimer–MCI classification (48.57%). These results highlight the challenges in distinguishing between certain conditions and demonstrate significant differences in model performance. Overall, the sensitivity, specificity, and accuracy values reflect the complexities involved in multi-class classification tasks.
Overall, LDA proved to be the most effective classification method across the board, while RF demonstrated significant variability and generally lower performance across different metrics. The RF classifier consistently showed lower performance across classification tasks. This may reflect the combination of a relatively small dataset, high feature dimensionality, and a reduced number of samples per class. Although hyperparameters such as the number of trees were optimized, these factors may have limited the classifier’s generalizability. Future studies should consider alternative balancing techniques, feature reduction strategies, or ensemble approaches to enhance RF performance.
According to the statistical analysis results, as seen in Figure 6, the channel Fp1 is selected for all four ND (Alzheimer’s, depression, MCI, and schizophrenia). While the number of features selected for this channel varies, its consistent appearance across all conditions suggests its potential importance in distinguishing these disorders. Additionally, other channels like Fp2, F7, and Fz are frequently selected, though not universally across all disorders. These shared channels might provide valuable insights for comparative analysis in multi-class classification tasks.
The frequent selection of frontal lobe channels (Fp1, Fp2, F3, F7, Fz) across the different ND in the Figure 6 suggests that frontal regions of the brain may play a critical role in differentiating these disorders. The frontal lobe is responsible for higher cognitive functions such as decision-making, attention, and emotional regulation, which are often impaired in these disorders. This focus on frontal channels could indicate that EEG signals from this region are particularly informative for identifying changes in brain activity associated with these conditions. Many neurological and psychiatric disorders studied (e.g., depression, Alzheimer’s, schizophrenia, MCI) involve functional alterations in frontal lobe networks, which are reflected in EEG measures of complexity, entropy, and spectral power [75,78,79,80]. Prior studies have also shown that frontal EEG abnormalities are robust biomarkers in these conditions, supporting this study’s findings [75,78,79,80]. However, further work with larger and more diverse datasets is needed to confirm the generalizability of these results.

5. Limitations and Future Work

This study is limited by a class imbalance problem. Class imbalance is a well-known challenge in EEG-based classification tasks, as it can bias model training and reduce generalizability. While oversampling methods such as SMOTE and ADASYN are commonly employed in general machine learning research, recent studies have highlighted potential pitfalls in biomedical contexts. Bunterngchit et al. [81] emphasize the importance of carefully imbalance-handling techniques to preserve the neurophysiological validity of EEG signals. Similarly, Rasool et al. [82] review deep learning-based strategies for addressing class imbalance, noting that although synthetic data generation can improve performance in some cases, it may also introduce artifacts that reduce clinical reliability. In this study, to maintain the authenticity of EEG recordings and avoid the potential biases of synthetic samples, only real, experimentally acquired data were used. Another limitation of this study is the absence of detailed clinical characterization of the participant groups. Key information, such as disease stage, medication status, comorbid conditions, and other relevant clinical variables, was not available, which may have influenced EEG patterns. Future studies should incorporate these factors to enhance the interpretability and generalizability of the findings. Also, in future studies, the analysis will be expanded to include other ND, such as epilepsy, to further investigate classification capabilities.

6. Conclusions

In this study, various features were extracted from EEG signals to achieve high-accuracy classification of multiple ND and to investigate the differentiation among these disorders. Feature selection was conducted using the Lasso algorithm, followed by classification using SVM, LDA, ANN, and RF algorithms. The results revealed that in binary classification, AD was accurately classified with 100% accuracy from healthy controls, while the Alzheimer–MCI classification achieved an accuracy of 96.90%. In the three-class classification involving depression, MCI, and schizophrenia, an accuracy of 84.67% was obtained. However, in the four-class classification, the maximum accuracy reached 57.89%. These findings highlight the effectiveness of EEG signal analysis in diagnosing and differentiating multiple ND. Furthermore, this research contributes to the field by utilizing a unique dataset that has not been previously examined in relation to these disorders in this manner. In conclusion, this study contributes to the literature by focusing on disease-to-disease classification and multi-class differentiation among ND, rather than control vs. patient classification.

Supplementary Materials

The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/diagnostics15172132/s1, Supplementary File S1: groups with significant differences as a result of post hoc analysis; Supplementary File S2: Significant features and different groups.

Funding

This research received no external funding.

Institutional Review Board Statement

Ethical review and approval were waived for this study, as the data were obtained from an open-access repository provided by Benninger, Felix et al. (https://doi.org/10.5061/dryad.8gtht76pw).

Informed Consent Statement

Informed consent was not required, since the dataset is openly available in a public repository.

Data Availability Statement

The data that support the findings of this study are openly available in [Dryad] at [https://doi.org/10.5061/dryad.8gtht76pw], reference number [34].

Conflicts of Interest

The author declares no conflicts of interest.

References

  1. World Health Organization. Neurological Disorders: Public Health Challenges; World Health Organization: Geneva, Switzerland, 2006; ISBN 9241563362.
  2. Raghavaiah, P.; Varadarajan, S. A CAD System Design for Alzheimer’s Disease Diagnosis Using Temporally Consistent Clustering and Hybrid Deep Learning Models. Biomed. Signal Process. Control 2022, 75, 103571. [Google Scholar] [CrossRef]
  3. Rakofsky, J.; Rapaport, M. Mood Disorders. Contin. Lifelong Learn. Neurol. 2018, 24, 804–827. [Google Scholar] [CrossRef]
  4. Stone, J. Functional Neurological Disorders: The Neurological Assessment as Treatment. Pract. Neurol. 2016, 16, 7–17. [Google Scholar] [CrossRef]
  5. Yan, Y.; Liu, G.; Cai, H.; Wu, E.Q.; Cai, J.; Cheok, A.D.; Liu, N.; Li, T.; Fan, Z. A Review of Graph Theory-Based Diagnosis of Neurological Disorders Based on EEG and MRI. Neurocomputing 2024, 599, 128098. [Google Scholar] [CrossRef]
  6. Vandana, J.; Nirali, N. A Review of EEG Signal Analysis for Diagnosis of Neurological Disorders Using Machine Learning. J. Biomed. Photonics Eng. 2021, 7, 40201. [Google Scholar]
  7. Jaiswal, A.; Kruiper, R.; Rasool, A.; Nandkeolyar, A.; Wall, D.P.; Washington, P. Digitally Diagnosing Multiple Developmental Delays Using Crowdsourcing Fused with Machine Learning: Protocol for a Human-in-the-Loop Machine Learning Study. JMIR Res. Protoc. 2024, 13, e52205. [Google Scholar] [CrossRef] [PubMed]
  8. Subha, D.P.; Joseph, P.K.; Acharya, U.R.; Lim, C.M. EEG Signal Analysis: A Survey. J. Med. Syst. 2010, 34, 195–212. [Google Scholar] [CrossRef]
  9. Asadi-Pooya, A.A.; Fattahi, D.; Abolpour, N.; Boostani, R.; Farazdaghi, M.; Sharifi, M. Epilepsy Classification Using Artificial Intelligence: A Web-based Application. Epilepsia Open 2023, 8, 1362–1368. [Google Scholar] [CrossRef]
  10. De Miras, J.R.; Ibáñez-Molina, A.J.; Soriano, M.F.; Iglesias-Parro, S. Schizophrenia Classification Using Machine Learning on Resting State EEG Signal. Biomed. Signal Process. Control 2023, 79, 104233. [Google Scholar] [CrossRef]
  11. Raza, N.; Naseer, A.; Tamoor, M.; Zafar, K. Alzheimer Disease Classification through Transfer Learning Approach. Diagnostics 2023, 13, 801. [Google Scholar] [CrossRef]
  12. Sibilano, E.; Brunetti, A.; Buongiorno, D.; Lassi, M.; Grippo, A.; Bessi, V.; Micera, S.; Mazzoni, A.; Bevilacqua, V. An Attention-Based Deep Learning Approach for the Classification of Subjective Cognitive Decline and Mild Cognitive Impairment Using Resting-State EEG. J. Neural Eng. 2023, 20, 16048. [Google Scholar] [CrossRef]
  13. Hosseinzadeh Kasani, P.; Lee, J.E.; Park, C.; Yun, C.-H.; Jang, J.-W.; Lee, S.-A. Evaluation of Nutritional Status and Clinical Depression Classification Using an Explainable Machine Learning Method. Front. Nutr. 2023, 10, 1165854. [Google Scholar] [CrossRef]
  14. Gautam, R.; Sharma, M. Prevalence and Diagnosis of Neurological Disorders Using Different Deep Learning Techniques: A Meta-Analysis. J. Med. Syst. 2020, 44, 49. [Google Scholar] [CrossRef]
  15. Ahmad Mir, W.; Izharuddin; Nissar, I. Contribution of Application of Deep Learning Approaches on Biomedical Data in the Diagnosis of Neurological Disorders: A Review on Recent Findings. In Advances in Computational Intelligence, Proceedings of the Security and Internet of Things: Second International Conference, ICCISIoT 2019, Agartala, India, 13–14 December 2019; Proceedings 2; Springer: Berlin/Heidelberg, Germany, 2020; pp. 87–97. [Google Scholar]
  16. Noor, M.B.T.; Zenia, N.Z.; Kaiser, M.S.; Mamun, S.A.; Mahmud, M. Application of Deep Learning in Detecting Neurological Disorders from Magnetic Resonance Images: A Survey on the Detection of Alzheimer’s Disease, Parkinson’s Disease and Schizophrenia. Brain Inform. 2020, 7, 11. [Google Scholar] [CrossRef]
  17. Marcus, G. Deep Learning: A Critical Appraisal. arXiv 2018, arXiv:1801.00631. [Google Scholar] [CrossRef]
  18. Bevilacqua, V.; Salatino, A.A.; Di Leo, C.; Tattoli, G.; Buongiorno, D.; Signorile, D.; Babiloni, C.; Del Percio, C.; Triggiani, A.I.; Gesualdo, L. Advanced Classification of Alzheimer’s Disease and Healthy Subjects Based on EEG Markers. In Proceedings of the 2015 International Joint Conference on Neural Networks (IJCNN), Killarney, Ireland, 12–17 July 2015; pp. 1–5. [Google Scholar]
  19. Bachmann, M.; Päeske, L.; Kalev, K.; Aarma, K.; Lehtmets, A.; Ööpik, P.; Lass, J.; Hinrikus, H. Methods for Classifying Depression in Single Channel EEG Using Linear and Nonlinear Signal Analysis. Comput. Methods Programs Biomed. 2018, 155, 11–17. [Google Scholar] [CrossRef]
  20. Hosseinifard, B.; Moradi, M.H.; Rostami, R. Classifying Depression Patients and Normal Subjects Using Machine Learning Techniques and Nonlinear Features from EEG Signal. Comput. Methods Programs Biomed. 2013, 109, 339–345. [Google Scholar] [CrossRef] [PubMed]
  21. Aljalal, M.; Molinas, M.; Aldosari, S.A.; AlSharabi, K.; Abdurraqeeb, A.M.; Alturki, F.A. Mild Cognitive Impairment Detection with Optimally Selected EEG Channels Based on Variational Mode Decomposition and Supervised Machine Learning. Biomed. Signal Process. Control 2024, 87, 105462. [Google Scholar] [CrossRef]
  22. Deng, J.; Sun, B.; Kavcic, V.; Liu, M.; Giordani, B.; Li, T. Novel Methodology for Detection and Prediction of Mild Cognitive Impairment Using Resting-state EEG. Alzheimer’s Dement. 2024, 20, 145–158. [Google Scholar] [CrossRef]
  23. Gosala, B.; Kapgate, P.D.; Jain, P.; Chaurasia, R.N.; Gupta, M. Wavelet Transforms for Feature Engineering in EEG Data Processing: An Application on Schizophrenia. Biomed. Signal Process. Control 2023, 85, 104811. [Google Scholar] [CrossRef]
  24. Alahmadi, T.J.; Rahman, A.U.; Alhababi, Z.A.; Ali, S.; Alkahtani, H.K. Prediction of Mild Cognitive Impairment Using EEG Signal and BiLSTM Network. Mach. Learn. Sci. Technol. 2024, 5, 25028. [Google Scholar] [CrossRef]
  25. Oltu, B.; Akşahin, M.F.; Kibaroğlu, S. A Novel Electroencephalography Based Approach for Alzheimer’s Disease and Mild Cognitive Impairment Detection. Biomed. Signal Process. Control 2021, 63, 102223. [Google Scholar] [CrossRef]
  26. Pirrone, D.; Weitschek, E.; Di Paolo, P.; De Salvo, S.; De Cola, M.C. EEG Signal Processing and Supervised Machine Learning to Early Diagnose Alzheimer’s Disease. Appl. Sci. 2022, 12, 5413. [Google Scholar] [CrossRef]
  27. Kim, M.; Youn, Y.C.; Paik, J. Deep Learning-Based EEG Analysis to Classify Normal, Mild Cognitive Impairment, and Dementia: Algorithms and Dataset. Neuroimage 2023, 272, 120054. [Google Scholar] [CrossRef]
  28. Sharma, N.; Kolekar, M.H.; Jha, K.; Kumar, Y. EEG and Cognitive Biomarkers Based Mild Cognitive Impairment Diagnosis. Innov. Res. Biomed. Eng. 2019, 40, 113–121. [Google Scholar] [CrossRef]
  29. Parsa, M.; Rad, H.Y.; Vaezi, H.; Hossein-Zadeh, G.-A.; Setarehdan, S.K.; Rostami, R.; Rostami, H.; Vahabie, A.-H. EEG-Based Classification of Individuals with Neuropsychiatric Disorders Using Deep Neural Networks: A Systematic Review of Current Status and Future Directions. Comput. Methods Programs Biomed. 2023, 240, 107683. [Google Scholar] [CrossRef]
  30. Shor, O.; Yaniv-Rosenfeld, A.; Valevski, A.; Weizman, A.; Khrennikov, A.; Benninger, F. EEG-Based Spatio-Temporal Relation Signatures for the Diagnosis of Depression and Schizophrenia. Sci. Rep. 2023, 13, 776. [Google Scholar] [CrossRef]
  31. Sarisik, E.; Popovic, D.; Keeser, D.; Khuntia, A.; Schiltz, K.; Falkai, P.; Pogarell, O.; Koutsouleris, N. EEG-Based Signatures of Schizophrenia, Depression, and Aberrant Aging: A Supervised Machine Learning Investigation. Schizophr. Bull. 2024, 51, sbae150. [Google Scholar] [CrossRef]
  32. Emre, I.E.; Erol, Ç.; Taş, C.; Tarhan, N. Multi-Class Classification Model for Psychiatric Disorder Discrimination. Int. J. Med. Inform. 2023, 170, 104926. [Google Scholar] [CrossRef] [PubMed]
  33. Wang, Z.; Feng, J.; Jiang, R.; Shi, Y.; Li, X.; Xue, R.; Du, X.; Ji, M.; Zhong, F.; Meng, Y. Automated Rest Eeg-Based Diagnosis of Depression and Schizophrenia Using a Deep Convolutional Neural Network. IEEE Access 2022, 10, 104472–104485. [Google Scholar] [CrossRef]
  34. Shor, O.; Glik, A.; Yaniv-Rosenfeld, A.; Valevski, A.; Weizman, A.; Khrennikov, A.; Benninger, F. EEG P-Adic Quantum Potential Accurately Identifies Depression, Schizophrenia and Cognitive Decline [Dataset]. Dryad. PLoS ONE 2021, 16, e0255529. [Google Scholar] [CrossRef]
  35. Stoica, P.; Moses, R.L. Spectral Analysis of Signals; Pearson Prentice Hall: Upper Saddle River, NJ, USA, 2005. [Google Scholar]
  36. Cohen, M.X. Analyzing Neural Time Series Data: Theory and Practice; MIT Press: Cambridge, MA, USA, 2014. [Google Scholar]
  37. Richman, J.S.; Moorman, J.R. Physiological Time-Series Analysis Using Approximate Entropy and Sample Entropy Maturity in Premature Infants Physiological Time-Series Analysis Using Approximate Entropy and Sample Entropy. Am. J. Physiol. Heart Circ. Physiol. 2000, 278, H2039–H2049. [Google Scholar] [CrossRef]
  38. Molina-Picó, A.; Cuesta-Frau, D.; Aboy, M.; Crespo, C.; Miró-Martínez, P.; Oltra-Crespo, S. Comparative Study of Approximate Entropy and Sample Entropy Robustness to Spikes. Artif. Intell. Med. 2011, 53, 97–106. [Google Scholar] [CrossRef]
  39. Cuesta-Frau, D.; Murillo-Escobar, J.P.; Orrego, D.A.; Delgado-Trejos, E. Embedded Dimension and Time Series Length. Practical Influence on Permutation Entropy and Its Applications. Entropy 2019, 21, 385. [Google Scholar] [CrossRef]
  40. Altıntop, Ç.G.; Latifoğlu, F.; Akın, A.K.; Bayram, A.; Çiftçi, M. Classification of Depth of Coma Using Complexity Measures and Nonlinear Features of Electroencephalogram Signals. Int. J. Neural Syst. 2022, 32, 2250018. [Google Scholar] [CrossRef]
  41. Pincus, S.M.S.M. Approximate Entropy as a Measure of System Complexity. Proc. Natl. Acad. Sci. USA 1991, 88, 2297–2301. [Google Scholar] [CrossRef]
  42. Coifman, R.R.; Meyer, Y.; Quake, S.; Wickerhauser, M.V. Signal Processing and Compression with Wavelet Packets. In Wavelets and Their Applications; Springer: Berlin/Heidelberg, Germany, 1994; pp. 363–379. [Google Scholar]
  43. Bandt, C.; Pompe, B. Permutation Entropy: A Natural Complexity Measure for Time Series. Phys. Rev. Lett. 2002, 88, 4. [Google Scholar] [CrossRef] [PubMed]
  44. Li, Y.; Wang, X.; Liu, Z.; Liang, X.; Si, S. The Entropy Algorithm and Its Variants in the Fault Diagnosis of Rotating Machinery: A Review. IEEE Access 2018, 6, 66723–66741. [Google Scholar] [CrossRef]
  45. Rényi, A. On Measures of Entropy and Information. In Proceedings of the Fourth Berkeley Symposium on Mathematical Statistics and Probability, Volume 1: Contributions to the Theory of Statistics; University of California Press: Oakland, CA, USA, 1961; pp. 547–561. [Google Scholar]
  46. Gell-Mann, M.; Tsallis, C. Nonextensive Entropy: Interdisciplinary Applications; Oxford University Press: Oxford, UK, 2004; ISBN 0198036213. [Google Scholar]
  47. Hjorth, B. EEG Analysis Based on Time Domain Properties. Electroencephalogr. Clin. Neurophysiol. 1970, 29, 306–310. [Google Scholar] [CrossRef] [PubMed]
  48. Kolmogorov, A.N. Three Approaches to the Definition of the Concept “Quantity of Information”. Probl. Peredachi Informatsii 1965, 1, 3–11. [Google Scholar]
  49. Lempel, A.; Ziv, J. On the Complexity of Finite Sequences. IEEE Trans. Inf. Theory 1976, 22, 75–81. [Google Scholar] [CrossRef]
  50. Nagarajan, R. Quantifying Physiological Data with Lempel-Ziv Complexity-Certain Issues. IEEE Trans. Biomed. Eng. 2002, 49, 1371–1373. [Google Scholar] [CrossRef]
  51. Kalousis, A.; Prados, J.; Hilario, M. Stability of Feature Selection Algorithms: A Study on High-Dimensional Spaces. Knowl. Inf. Syst. 2007, 12, 95–116. [Google Scholar] [CrossRef]
  52. Remeseiro, B.; Bolon-Canedo, V. A Review of Feature Selection Methods in Medical Applications. Comput. Biol. Med. 2019, 112, 103375. [Google Scholar] [CrossRef]
  53. Sakar, C.O.; Kursun, O. Telediagnosis of Parkinson’s Disease Using Measurements of Dysphonia. J. Med. Syst. 2010, 34, 591–599. [Google Scholar] [CrossRef]
  54. Richhariya, B.; Tanveer, M.; Rashid, A.H.; Initiative, A.D.N. Diagnosis of Alzheimer’s Disease Using Universum Support Vector Machine Based Recursive Feature Elimination (USVM-RFE). Biomed. Signal Process. Control 2020, 59, 101903. [Google Scholar] [CrossRef]
  55. Tejeswinee, K.; Shomona, G.J.; Athilakshmi, R. Feature Selection Techniques for Prediction of Neuro-Degenerative Disorders: A Case-Study with Alzheimer’s and Parkinson’s Disease. Procedia Comput. Sci. 2017, 115, 188–194. [Google Scholar] [CrossRef]
  56. Priya, S.J.; Rani, A.J.; Subathra, M.S.P.; Mohammed, M.A.; Damaševičius, R.; Ubendran, N. Local Pattern Transformation Based Feature Extraction for Recognition of Parkinson’s Disease Based on Gait Signals. Diagnostics 2021, 11, 1395. [Google Scholar] [CrossRef]
  57. Baygin, M. An Accurate Automated Schizophrenia Detection Using TQWT and Statistical Moment Based Feature Extraction. Biomed. Signal Process. Control 2021, 68, 102777. [Google Scholar] [CrossRef]
  58. Mishra, P.; Pandey, C.M.; Singh, U.; Keshri, A.; Sabaretnam, M. Selection of Appropriate Statistical Methods for Data Analysis. Ann. Card. Anaesth. 2019, 22, 297–301. [Google Scholar] [CrossRef] [PubMed]
  59. Lee, S.; Lee, D.K. What Is the Proper Way to Apply the Multiple Comparison Test? Korean J. Anesthesiol. 2018, 71, 353. [Google Scholar] [CrossRef] [PubMed]
  60. Pisner, D.A.; Schnyer, D.M. Support Vector Machine. In Machine Learning; Elsevier: Amsterdam, The Netherlands, 2020; pp. 101–121. [Google Scholar]
  61. Altıntop, Ç.G.; Latifoğlu, F.; Akın, A.K. Can Patients in Deep Coma Hear Us? Examination of Coma Depth Using Physiological Signals. Biomed. Signal Process. Control 2022, 77, 103756. [Google Scholar] [CrossRef]
  62. Zhao, H.; Lai, Z.; Leung, H.; Zhang, X.; Zhao, H.; Lai, Z.; Leung, H.; Zhang, X. Linear Discriminant Analysis. Featur. Learn. Underst. Algorithms Appl. 2020, 4, 71–85. [Google Scholar]
  63. Féraud, R.; Clérot, F. A Methodology to Explain Neural Network Classification. Neural Netw. 2002, 15, 237–246. [Google Scholar] [CrossRef]
  64. Sokolova, M.; Lapalme, G. A Systematic Analysis of Performance Measures for Classification Tasks. Inf. Process. Manag. 2009, 45, 427–437. [Google Scholar] [CrossRef]
  65. Gemein, L.A.W.; Schirrmeister, R.T.; Chrabąszcz, P.; Wilson, D.; Boedecker, J.; Schulze-Bonhage, A.; Hutter, F.; Ball, T. Machine-Learning-Based Diagnostics of EEG Pathology. Neuroimage 2020, 220, 117021. [Google Scholar] [CrossRef]
  66. Ahmadlou, M.; Adeli, H.; Adeli, A. Fractality Analysis of Frontal Brain in Major Depressive Disorder. Int. J. Psychophysiol. 2012, 85, 206–211. [Google Scholar] [CrossRef]
  67. Kang, X.; Liu, X.; Chen, S.; Zhang, W.; Liu, S.; Ming, D. Major Depressive Disorder Recognition by Quantifying EEG Signal Complexity Using Proposed APLZC and AWPLZC. J. Affect. Disord. 2024, 356, 105–114. [Google Scholar] [CrossRef]
  68. Li, Y.; Tong, S.; Liu, D.; Gai, Y.; Wang, X.; Wang, J.; Qiu, Y.; Zhu, Y. Abnormal EEG Complexity in Patients with Schizophrenia and Depression. Clin. Neurophysiol. 2008, 119, 1232–1241. [Google Scholar] [CrossRef] [PubMed]
  69. Abásolo1, D.; Hornero, R.; Espino, P. Approximate Entropy of EEG Background Activity in Alzheimer’s Disease Patients. Intell. Autom. Soft Comput. 2009, 15, 591–603. [Google Scholar]
  70. Huang, Y.; Yi, Y.; Chen, Q.; Li, H.; Feng, S.; Zhou, S.; Zhang, Z.; Liu, C.; Li, J.; Lu, Q. Analysis of EEG Features and Study of Automatic Classification in First-Episode and Drug-Naïve Patients with Major Depressive Disorder. BMC Psychiatry 2023, 23, 832. [Google Scholar] [CrossRef]
  71. Jang, K.-I.; Kim, S.; Kim, S.Y.; Lee, C.; Chae, J.-H. Machine Learning-Based Electroencephalographic Phenotypes of Schizophrenia and Major Depressive Disorder. Front. Psychiatry 2021, 12, 745458. [Google Scholar] [CrossRef]
  72. Chen, H.; Lei, Y.; Li, R.; Xia, X.; Cui, N.; Chen, X.; Liu, J.; Tang, H.; Zhou, J.; Huang, Y. Resting-State EEG Dynamic Functional Connectivity Distinguishes Non-Psychotic Major Depression, Psychotic Major Depression and Schizophrenia. Mol. Psychiatry 2024, 29, 1088–1098. [Google Scholar] [CrossRef]
  73. Hassanzadeh, R.; Abrol, A.; Calhoun, V. Classification of Schizophrenia and Alzheimer’s Disease Using Resting-State Functional Network Connectivity. In Proceedings of the 2022 IEEE-EMBS International Conference on Biomedical and Health Informatics (BHI), Ioannina, Greece, 27–30 September 2022; pp. 1–4. [Google Scholar]
  74. Akrofi, K.; Baker, M.C.; O’Boyle, M.W.; Schiffer, R.B. Clustering and Modeling of EEG Coherence Features of Alzheimer’s and Mild Cognitive Impairment Patients. In Proceedings of the 2008 30th Annual International Conference of the IEEE Engineering in Medicine and Biology Society, Vancouver, BC, Canada, 20–25 August 2008; pp. 1092–1095. [Google Scholar]
  75. McBride, J.C.; Zhao, X.; Munro, N.B.; Smith, C.D.; Jicha, G.A.; Hively, L.; Broster, L.S.; Schmitt, F.A.; Kryscio, R.J.; Jiang, Y. Spectral and Complexity Analysis of Scalp EEG Characteristics for Mild Cognitive Impairment and Early Alzheimer’s Disease. Comput. Methods Programs Biomed. 2014, 114, 153–163. [Google Scholar] [CrossRef]
  76. Fiscon, G.; Weitschek, E.; Cialini, A.; Felici, G.; Bertolazzi, P.; De Salvo, S.; Bramanti, A.; Bramanti, P.; De Cola, M.C. Combining EEG Signal Processing with Supervised Methods for Alzheimer’s Patients Classification. BMC Med. Inform. Decis. Mak. 2018, 18, 35. [Google Scholar] [CrossRef]
  77. Ieracitano, C.; Mammone, N.; Hussain, A.; Morabito, F.C. A Novel Multi-Modal Machine Learning Based Approach for Automatic Classification of EEG Recordings in Dementia. Neural Netw. 2020, 123, 176–190. [Google Scholar] [CrossRef] [PubMed]
  78. Mizuno, T.; Takahashi, T.; Cho, R.Y.; Kikuchi, M.; Murata, T.; Takahashi, K.; Wada, Y. Assessment of EEG Dynamical Complexity in Alzheimer’s Disease Using Multiscale Entropy. Clin. Neurophysiol. Off. J. Int. Fed. Clin. Neurophysiol. 2010, 121, 1438–1446. [Google Scholar] [CrossRef] [PubMed]
  79. Sun, J.; Wang, B.; Niu, Y.; Tan, Y.; Fan, C.; Zhang, N.; Xue, J.; Wei, J.; Xiang, J. Complexity Analysis of EEG, MEG, and FMRI in Mild Cognitive Impairment and Alzheimer’s Disease: A Review. Entropy 2020, 22, 239. [Google Scholar] [CrossRef]
  80. Čukić, M.; Stokić, M.; Radenković, S.; Ljubisavljević, M.; Simić, S.; Savić, D. Nonlinear Analysis of EEG Complexity in Episode and Remission Phase of Recurrent Depression. Int. J. Methods Psychiatr. Res. 2020, 29, e1816. [Google Scholar] [CrossRef]
  81. Bunterngchit, C.; Wang, J.; Su, J.; Wang, Y.; Liu, S.; Hou, Z.-G. Temporal Attention Fusion Network with Custom Loss Function for EEG–FNIRS Classification. J. Neural Eng. 2024, 21, 66016. [Google Scholar] [CrossRef] [PubMed]
  82. Rasool, A.; Aslam, S.; Xu, Y.; Wang, Y.; Pan, Y.; Chen, W. Deep Neurocomputational Fusion for ASD Diagnosis Using Multi-Domain EEG Analysis. Neurocomputing 2025, 641, 130353. [Google Scholar] [CrossRef]
Figure 1. Workflow of EEG signal processing and classification.
Figure 1. Workflow of EEG signal processing and classification.
Diagnostics 15 02132 g001
Figure 2. The standard 10–20 EEG electrode placements with 19 channels.
Figure 2. The standard 10–20 EEG electrode placements with 19 channels.
Diagnostics 15 02132 g002
Figure 3. Filtered one-minute EEG signals from one person with each neurological disorder and healthy controls.
Figure 3. Filtered one-minute EEG signals from one person with each neurological disorder and healthy controls.
Diagnostics 15 02132 g003
Figure 4. Lambda value determination for classification healthy control and AD using LDA algorithm.
Figure 4. Lambda value determination for classification healthy control and AD using LDA algorithm.
Diagnostics 15 02132 g004
Figure 5. The number of selected features for the binary classification and the total count of selected features.
Figure 5. The number of selected features for the binary classification and the total count of selected features.
Diagnostics 15 02132 g005
Figure 6. Selected feature counts for binary classification according to EEG channels: (a) control–Alzheimer, (b) control–depression (c) control–MCI, (d) control–schizophrenia.
Figure 6. Selected feature counts for binary classification according to EEG channels: (a) control–Alzheimer, (b) control–depression (c) control–MCI, (d) control–schizophrenia.
Diagnostics 15 02132 g006
Figure 7. Box-plot graphics for ND and healthy control. Statistical significance between groups is indicated by asterisks (* p < 0.05; ** p < 0.01; *** p < 0.001; **** p < 0.0001).
Figure 7. Box-plot graphics for ND and healthy control. Statistical significance between groups is indicated by asterisks (* p < 0.05; ** p < 0.01; *** p < 0.001; **** p < 0.0001).
Diagnostics 15 02132 g007
Figure 8. Overall confusion matrices obtained using the LDA classifier for (a) control–AD–MCI, (b) depression–MCI–schizophrenia, (c) control–depression–schizophrenia–MCI, and (d) control–depression–schizophrenia–cognitive decline classifications.
Figure 8. Overall confusion matrices obtained using the LDA classifier for (a) control–AD–MCI, (b) depression–MCI–schizophrenia, (c) control–depression–schizophrenia–MCI, and (d) control–depression–schizophrenia–cognitive decline classifications.
Diagnostics 15 02132 g008
Table 1. Some binary classification studies.
Table 1. Some binary classification studies.
Ref.Disorder(s)Feature Extraction and SelectionClassifier(s)Accuracy
[18]AD vs. ControlFrequency axis features, power, coherence,
power ratio, frequency bandwidth;
SVM-RFE, PCA (feature selection)
ANN, SVM86% (ANN)
[19]Depression vs. ControlPSD, Higuchi fractal dimensionLogistic Regression (LR)92%
[20]Depression vs. ControlDetrended Fluctuation Analysis (DFA),
correlation dimension, higuchi fractal, and lyapunov exponent
LDA, K-nearest neighbor (KNN), LR90% (LR)
[21]MCI vs. ControlVariational Mode Decomposition (VMD)KNN, non-dominated sorting genetic algorithm (NSGA)99.81%
[22]MCI vs. ControlSoft-decision score featuresAdaBoost91.97%
[10]Schizophrenia vs. ControlLinear and nonlinear features, PCA (feature reduction)SVM, LR, KNN, Random Forest (RF), Decision Tree (DT)89% (SVM)
[23]Schizophrenia vs. ControlStatistical + Wavelet featuresLR, SVM, DT, RF, AdaBoost, and Gradient Boost97.98% (DT)
Table 2. Demographic information of participants.
Table 2. Demographic information of participants.
GroupAge (Mean ± Standard Deviation)FemaleMaleTotal Volunteer
Control40.3 ± 13.8271643
Depression69.8 ± 14.928-28
Alzheimer70.7 ± 8.5211940
Schizophrenia41.4 ± 16.8152742
MCI73.6 ± 5.5141428
Table 3. Electrode locations for EEG recordings.
Table 3. Electrode locations for EEG recordings.
Channel Number12345678910111213141516171819
Fp1Fp2F3F4C3C4P3P4O1O2F7F8T3T4T5T6FzCzPz
Table 4. Extracted features.
Table 4. Extracted features.
IDGroupFeature NameAbbreviationBrief Explanation of Feature
1Frequency axis measuresFrequency valuefreqmaxThe frequency value that gives the peak power value as a result of the power spectral density (PSD) analysis.
2Power contentpowmaxMaximum power value of PSD, indicating the dominant power in the signal
3Spectral edge frequencyspedgeSpectral edge frequency indicates the frequency limit below which a certain percentage (in this case 50%) of the total spectral energy of a signal lies.
4Entropy measuresApproximation EntropyApEnA measure of the regularity and complexity of time-series data, with lower values indicating more regular patterns.
5LogEnergy EntropylogenergyQuantifies the entropy or unpredictability based on logarithmic energy values of the signal
6Permutation EntropyPermEnA measure of complexity that reflects the order and patterns in time-series data.
7Tsallis EntropytsallisA generalization of Shannon entropy that provides information about the complexity and structure in the data.
8Renyi EntropyrenyiA measure of entropy that generalizes the concept by focusing on different aspects of data distribution.
9Spectral EntropyspentropyEntropy computed from the power spectral density, indicating the signal’s frequency-domain complexity.
10Spectral entropy for dyadic levelsspentropDydSpectral entropy calculated over dyadic levels, offering insights into multi-scale signal complexity.
11Hjorth parametersActivityactivityThe variation in the time-series signal is related to its power and total energy.
12MobilitymobilityMeasures the frequency variation in the signal, related to the ratio of the standard deviation of the first derivative.
13ComplexitycomplexityQuantifies the deviation from pure sine wave activity, assessing the signal’s frequency modulations.
14Statistical measuresKurtosiskurtDescribes the “tailedness” of the probability distribution, indicating outlier presence in the signal.
15SkewnessskewMeasures the asymmetry of the signal’s amplitude distribution, indicating if values are skewed towards higher or lower values.
16Complexity measuresLempel–Ziv complexitylempelA measure of complexity based on the compressibility of the signal; higher values indicate more irregularity.
17Kolmogorov complexitykolmogorovA measure of the complexity of a signal, representing the length of the shortest possible description (algorithm) that can generate it.
Table 5. Selected feature numbers.
Table 5. Selected feature numbers.
Classified ClassesSelected Feature NumbersTotal Feature Numbers Before Selection
Control–depression8619 channel × 34 features = 646 features
Control–AD99
Control–schizophrenia100
Control–MCI90
Depression–AD79
Depression–schizophrenia91
Depression–MCI74
AD–schizophrenia108
AD–MCI88
Schizophrenia–MCI87
Table 6. Algorithms’ running times for two-class classification (control–disease).
Table 6. Algorithms’ running times for two-class classification (control–disease).
Training Time
DiseaseSVMLDAANNRF
Alzheimer18.35665.353661.51400.5953
Depression15.31655.446735.50720.3766
Schizophrenia18.52795.437442.07070.6761
MCI16.79285.364929.70700.1619
Testing Time
DiseaseSVMLDAANNRF
Alzheimer0.00340.00210.00140.3098
Depression0.00370.00210.00130.2496
Schizophrenia0.00360.00260.00130.3622
MCI0.00360.00200.00130.0896
Table 7. Binary classification results (control–disease).
Table 7. Binary classification results (control–disease).
Control–Alzheimer
MethodAccuracySensitivitySpecificityPrecisionFscoreAUC
SVM0.9250 ± 0.10540.9500 ± 0.10540.9000 ± 0.17480.9217 ± 0.13010.9296 ± 0.09760.9688 ± 0.0988
LDA1.0000 ± 0.00001.0000 ± 0.00001.0000 ± 0.00001.0000 ± 0.00001.0000 ± 0.00001.0000 ± 0.0000
ANN0.8750 ± 0.18630.8750 ± 0.31730.8750 ± 0.21250.8963 ± 0.16370.9284 ± 0.12110.8750 ± 0.1863
RF0.7208 ± 0.14980.7250 ± 0.21890.7200 ± 0.16360.7067 ± 0.14930.7060 ± 0.18180.7225 ± 0.1474
Control–Depression
MethodAccuracySensitivitySpecificityPrecisionFscoreAUC
SVM0.9589 ± 0.09450.9333 ± 0.14050.9750 ± 0.07910.9667 ± 0.10540.9467 ± 0.11670.9917 ± 0.0264
LDA0.9857 ± 0.04521.0000 ± 0.00000.9800 ± 0.06320.9667 ± 0.10540.9800 ± 0.06320.9900 ± 0.0316
ANN0.9446 ± 0.07170.9667 ± 0.10540.9300 ± 0.11350.9167 ± 0.13610.9314 ± 0.09060.9483 ± 0.0686
RF0.7339 ± 0.11860.5333 ± 0.21940.8550 ± 0.20880.8100 ± 0.24700.6055 ± 0.15340.6942 ± 0.1061
Control–Schizophrenia
MethodAccuracySensitivitySpecificityPrecisionFscoreAUC
SVM0.9403 ± 0.10100.9500 ± 0.10540.9300 ± 0.11350.9300 ± 0.11350.9389 ± 0.10540.9738 ± 0.0641
LDA0.9778 ± 0.04680.9600 ± 0.08431.0000 ± 0.00001.0000 ± 0.00000.9778 ± 0.04680.9800 ± 0.0422
ANN0.9639 ± 0.05830.9750 ± 0.07910.9500 ± 0.10540.9600 ± 0.08430.9635 ± 0.05940.9625 ± 0.0604
RF0.7278 ± 0.13860.6550 ± 0.21400.7800 ± 0.24860.7842 ± 0.20770.6962 ± 0.17740.7175 ± 0.1419
Control–Mild cognitive impairement
MethodAccuracySensitivitySpecificityPrecisionFscoreAUC
SVM0.9571 ± 0.06900.9333 ± 0.14050.9750 ± 0.07910.9750 ± 0.07910.9457 ± 0.08881.0000 ± 0.0000
LDA0.9714 ± 0.06020.9667 ± 0.10540.9750 ± 0.07910.9750 ± 0.07910.9657 ± 0.07350.9708 ± 0.0623
ANN0.9571 ± 0.09640.9333 ± 0.14050.9750 ± 0.07910.9667 ± 0.10540.9467 ± 0.11670.9542 ± 0.1010
RF0.7304 ± 0.15890.4833 ± 0.27720.9000 ± 0.17480.8333 ± 0.25000.6265 ± 0.20580.6917 ± 0.1610
Table 8. Binary classification results (disease–disease).
Table 8. Binary classification results (disease–disease).
Alzheimer–MCI
MethodAccuracySensitivitySpecificityPrecisionFscoreAUC
SVM0.9262 ± 0.07810.9500 ± 0.10540.9000 ± 0.16100.9400 ± 0.09660.9381 ± 0.06630.9833 ± 0.0351
LDA0.9690 ± 0.06550.9500 ± 0.10541.0000 ± 0.00001.0000 ± 0.00000.9714 ± 0.06020.9750 ± 0.0527
ANN0.9262 ± 0.07810.9500 ± 0.10540.9000 ± 0.16100.9400 ± 0.09660.9381 ± 0.06630.9250 ± 0.0805
RF0.7524 ± 0.16770.9000 ± 0.12910.5500 ± 0.31480.7538 ± 0.15410.8138 ± 0.12460.7250 ± 0.1845
Alzheimer–Schizophrenia
MethodAccuracySensitivitySpecificityPrecisionFscoreAUC
SVM0.8667 ± 0.12200.8500 ± 0.17480.8850 ± 0.17000.8950 ± 0.14620.8575 ± 0.12790.9238 ± 0.1256
LDA0.9278 ± 0.06240.9250 ± 0.12080.9300 ± 0.11350.9400 ± 0.09660.9238 ± 0.06680.9275 ± 0.0629
ANN0.9028 ± 0.11430.8500 ± 0.24150.9550 ± 0.09560.9600 ± 0.08430.8749 ± 0.17930.9025 ± 0.1145
RF0.7181 ± 0.10580.7750 ± 0.14190.6550 ± 0.24430.7271 ± 0.16470.7307 ± 0.08320.7150 ± 0.1042
Depression–Alzheimer
MethodAccuracySensitivitySpecificityPrecisionFscoreAUC
SVM0.8810 ± 0.06350.9000 ± 0.16100.8750 ± 0.13180.8583 ± 0.15240.8571 ± 0.07970.9208 ± 0.1010
LDA0.9429 ± 0.07380.9667 ± 0.10540.9250 ± 0.12080.9250 ± 0.12080.9371 ± 0.08280.9458 ± 0.0710
ANN0.8238 ± 0.11180.7667 ± 0.31620.8750 ± 0.13180.8333 ± 0.16140.8201 ± 0.08710.8208 ± 0.1333
RF0.7333 ± 0.18300.5500 ± 0.35180.8500 ± 0.17480.6917 ± 0.40450.7327 ± 0.16250.7000 ± 0.2068
Depression–MCI
MethodAccuracySensitivitySpecificityPrecisionFscoreAUC
SVM0.9667 ± 0.07030.9667 ± 0.10540.9667 ± 0.10540.9750 ± 0.07910.9657 ± 0.07350.9889 ± 0.0351
LDA0.9633 ± 0.07770.9333 ± 0.14051.0000 ± 0.00001.0000 ± 0.00000.9600 ± 0.08430.9667 ± 0.0703
ANN0.9467 ± 0.08640.9333 ± 0.14050.9667 ± 0.10540.9750 ± 0.07910.9457 ± 0.08880.9500 ± 0.0805
RF0.6700 ± 0.24570.7000 ± 0.35830.6667 ± 0.31430.6778 ± 0.24380.7061 ± 0.23470.6833 ± 0.2250
Depression–Schizophrenia
MethodAccuracySensitivitySpecificityPrecisionFscoreAUC
SVM0.9143 ± 0.09990.8833 ± 0.19330.9300 ± 0.11350.9083 ± 0.14930.8790 ± 0.14000.9467 ± 0.1068
LDA0.9571 ± 0.06900.9500 ± 0.15810.9500 ± 0.10540.9500 ± 0.10540.9381 ± 0.11240.9500 ± 0.0874
ANN0.9143 ± 0.09990.9167 ± 0.18000.9050 ± 0.12350.8833 ± 0.15320.8848 ± 0.13750.9108 ± 0.1102
RF0.6857 ± 0.14750.4333 ± 0.30630.8600 ± 0.16960.6979 ± 0.27440.5863 ± 0.20200.6467 ± 0.1596
Schizophrenia–MCI
MethodAccuracySensitivitySpecificityPrecisionFscoreAUC
SVM0.8714 ± 0.12510.8167 ± 0.25400.9000 ± 0.17480.9017 ± 0.16220.8240 ± 0.17810.9833 ± 0.0351
LDA0.9571 ± 0.06900.9333 ± 0.14050.9750 ± 0.07910.9750 ± 0.07910.9457 ± 0.08880.9542 ± 0.0747
ANN0.7857 ± 0.21560.8000 ± 0.28110.7850 ± 0.26570.7767 ± 0.26820.7526 ± 0.23410.7925 ± 0.2143
RF0.7429 ± 0.16220.6500 ± 0.27720.8100 ± 0.21710.7667 ± 0.25090.6543 ± 0.20150.7300 ± 0.1626
Table 9. Three-class classification results for the LDA algorithm (control–disease–disease).
Table 9. Three-class classification results for the LDA algorithm (control–disease–disease).
ClassessAccuracySensitivitySpecificityPrecisionFscoreAUC
Cont-Depr-Alz0.7583 ± 0.14950.7350 ± 0.15730.8808 ± 0.07350.7488 ± 0.16950.7576 ± 0.15400.7840 ± 0.1939
Cont-Depr-MCI0.7114 ± 0.15130.7294 ± 0.14490.8565 ± 0.07800.7250 ± 0.15130.7127 ± 0.14420.5702 ± 0.1262
Cont-Depr-Sch0.7432 ± 0.14220.7333 ± 0.14680.8779 ± 0.06710.7672 ± 0.12800.7280 ± 0.14490.5325 ± 0.1116
Cont-Sch-Alz0.7603 ± 0.14320.7583 ± 0.14450.8787 ± 0.07230.7938 ± 0.14940.7490 ± 0.14520.4484 ± 0.0788
Cont-Sch-MCI0.7424 ± 0.10150.7489 ± 0.10800.8677 ± 0.05330.7853 ± 0.09390.7379 ± 0.12030.5421 ± 0.0977
Cont-MCI-Alz0.7932 ± 0.10410.7950 ± 0.09430.8950 ± 0.05750.8128 ± 0.11730.7863 ± 0.10150.5975 ± 0.1133
Cont = ‘Control’, Depr = ‘Depression’, Alz = ‘Alzheimer’, MCI = ‘Mild cognitive impairment’, Sch = ‘Schizophrenia’.
Table 10. Three-class classification results for the LDA algorithm (three disorders).
Table 10. Three-class classification results for the LDA algorithm (three disorders).
ClassessAccuracySensitivitySpecificityPrecisionFscoreAUC
Depr-Alz-MCI0.8100 ± 0.07320.8028 ± 0.10510.8984 ± 0.04560.8698 ± 0.05370.8203 ± 0.06350.8925 ± 0.1131
Depr-MCI-Sch0.8467 ± 0.10880.8278 ± 0.11840.9208 ± 0.05370.8650 ± 0.12780.8296 ± 0.12160.9128 ± 0.1090
Depr-Sch-Alz0.6455 ± 0.10880.6428 ± 0.09940.8214 ± 0.05420.6678 ± 0.11610.6268 ± 0.10470.7298 ± 0.1536
Alz-MCI-Sch0.7636 ± 0.13690.7450 ± 0.16360.8843 ± 0.06930.7600 ± 0.15190.7771 ± 0.11860.7875 ± 0.1602
Cont = ‘dontrol’, Depr = ‘depression’, Alz = ‘Alzheimer’, MCI = ‘mild cognitive impairment’, Sch = ‘schizophrenia’.
Table 11. Four-class classification results for the LDA algorithm.
Table 11. Four-class classification results for the LDA algorithm.
ClassessAccuracySensitivitySpecificityPrecisionFscore
Cont-Depr-Sch-Alz0.5146 ± 0.11480.5029 ± 0.10980.8388 ± 0.03910.5338 ± 0.11750.5148 ± 0.1044
Cont-Depr-Sch-MCI0.5605 ± 0.14100.5575 ± 0.15030.8533 ± 0.04750.5971 ± 0.18020.6122 ± 0.0647
Cont-Depr-Sch-CD0.5789 ± 0.13090.5389 ± 0.14270.8589 ± 0.04350.5475 ± 0.15590.5966 ± 0.1141
Cont-Sch-Alz-MCI0.5617 ± 0.08880.5533 ± 0.09020.8534 ± 0.03050.6150 ± 0.07420.5701 ± 0.0856
Cont-Depr-Alz-MCI0.5555 ± 0.14710.5296 ± 0.14360.8540 ± 0.04960.5429 ± 0.16700.6064 ± 0.1548
Depr-Sch-Alz-MCI0.4857 ± 0.11710.4667 ± 0.12350.8272 ± 0.03830.4904 ± 0.14850.5410 ± 0.1033
Cont = ‘control’, Depr = ‘depression’, Alz = ‘Alzheimer’, MCI = ‘mild cognitive impairement’, Sch = ‘schizophrenia’, CD = ‘cognitive decline’. CD designates Alzheimer and MCI as the same class.
Table 12. Summary of related studies.
Table 12. Summary of related studies.
StudyDisorders ComparedFeatures/MethodsClassifierClassesAccuracy/Metrics
Huang et al. [70]Depression vs. HCSVM-RFE, Lasso, PCAKNN288.2% Acc
Shor et al. [30] Schizophrenia, Depression, HCp-adic spatio-temporal EEG analysis2AUC 99.89% (HC–Depr), 86.08% (Depr–Sch)
Jang et al. [71]Schizophrenia, Depression, HCN1, P3 amplitudes (P300 task)LDA, SVM2Acc 71.31% (Sch–HC), 74.55% (Depr–HC), 59.71% (Sch–Depr)
Emre et al. [32]Multiple disorders & HCFrequency contentC5.0, RF, SVM, ANNMulti0.762–0.841 Acc
Sarisik et al. [31]Schizophrenia, Depression, HCCentral alpha power, EphysAGESVM2Acc 72.7% (Sch–HC), 67% (Depr–HC), 63.2% (Sch–Depr)
Wang et al. [33] Schizophrenia, Depression, HCMUCHf-Net (CNN)CNN379.27% Acc
Cheng et al. [72]NPMD, PMD, Schizophrenia, HCDynamic Functional ConnectivityRF473.1% Acc
Hassanzadeh et al. [73]AD, Schizophrenia, HCICA, Pearson correlationKNN, RF, LR, SVM2/3Acc 89% (AD-Sch), 68% (AD-Sch-HC)
Akrofi et al. [74]AD, MCI, HCEEG featuresKNN, MDA384% Acc
McBride et al. [75]AD, MCI, HCEntropy, spectral power, complexitySVM385.4% Acc
Fiscon et al. [76]AD, MCI, HCFrequency featuresDT2/373.4% (MCI-AD-HC)–91.7% (HC-MCI) Acc
Pirrone et al. [26]AD, MCI, HCPower band differencesKNN, SVM, DT2/375% (HC-MCI-AD)–97% (HC-MCI) Acc
Ieracitano et al. [77]AD, MCI, HCCWT + bispectrumMLP389.22% Acc
Oltu et al. [25] AD, MCI, HCDWT, PSD, coherenceBagged Tree396.5% Acc
HC = ’healthy control’, Depr = ‘depression’, AD = ‘Alzheimer’s disease’, MCI = ‘mild cognitive impairment’, Sch = ‘schizophrenia’, NPMD = ‘nonpsychotic major depression’, PMD = ‘psychotic major depression’.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Altıntop, Ç.G. Enhancing Diagnostic Accuracy of Neurological Disorders Through Feature-Driven Multi-Class Classification with Machine Learning. Diagnostics 2025, 15, 2132. https://doi.org/10.3390/diagnostics15172132

AMA Style

Altıntop ÇG. Enhancing Diagnostic Accuracy of Neurological Disorders Through Feature-Driven Multi-Class Classification with Machine Learning. Diagnostics. 2025; 15(17):2132. https://doi.org/10.3390/diagnostics15172132

Chicago/Turabian Style

Altıntop, Çiğdem Gülüzar. 2025. "Enhancing Diagnostic Accuracy of Neurological Disorders Through Feature-Driven Multi-Class Classification with Machine Learning" Diagnostics 15, no. 17: 2132. https://doi.org/10.3390/diagnostics15172132

APA Style

Altıntop, Ç. G. (2025). Enhancing Diagnostic Accuracy of Neurological Disorders Through Feature-Driven Multi-Class Classification with Machine Learning. Diagnostics, 15(17), 2132. https://doi.org/10.3390/diagnostics15172132

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop