Next Article in Journal
Evaluation of CK8/18 and CK19 Expression as Adjunct Immunohistochemical Markers in Non-Keratinizing Nasopharyngeal Carcinoma
Previous Article in Journal
A YOLO Ensemble Framework for Detection of Barrett’s Esophagus Lesions in Endoscopic Images
Previous Article in Special Issue
Quantitative DSA Analysis of MCA Aneurysms Using SymDIRECT Pixel Clustering: A Novel Framework for Objective Post-Treatment Evaluation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Recognition of EEG Features in Autism Disorder Using SWT and Fisher Linear Discriminant Analysis

by
Fahmi Fahmi
1,*,
Melinda Melinda
2,
Prima Dewi Purnamasari
3,
Elizar Elizar
2 and
Aufa Rafiki
2
1
Department of Electrical Engineering, Universitas Sumatera Utara, Medan 20155, Indonesia
2
Department of Electrical and Computer Engineering, Universitas Syiah Kuala, Banda Aceh 23111, Indonesia
3
Department of Electrical Engineering, Universitas Indonesia, Depok 16424, Indonesia
*
Author to whom correspondence should be addressed.
Diagnostics 2025, 15(18), 2291; https://doi.org/10.3390/diagnostics15182291
Submission received: 3 July 2025 / Revised: 28 August 2025 / Accepted: 3 September 2025 / Published: 10 September 2025
(This article belongs to the Special Issue Advances in the Diagnosis of Nervous System Diseases—3rd Edition)

Abstract

Background/Objectives: An ASD diagnosis from EEG is challenging due to non-stationary, low-SNR signals and small cohorts. We propose a compact, interpretable pipeline that pairs a shift-invariant Stationary Wavelet Transform (SWT) with Fisher’s Linear Discriminant (FLDA) as a supervised projection method, delivering band-level insight and subject-wise evaluation suitable for resource-constrained clinics. Methods: EEG from the KAU dataset (eight ASD, eight controls; 256 Hz) was decomposed with SWT (db4). We retained levels 3, 4, and 6 (γ/β/θ) as features. FLDA learned a low-dimensional discriminant subspace, followed by a linear decision rule. Evaluation was conducted using a subject-wise 70/30 split (no subject overlap) with accuracy, precision, recall, F1, and confusion matrices. Results: The β band (Level 4) achieved the best performance (accuracy/precision/recall/F1 = 0.95), followed by γ (0.92) and θ (0.85). Despite partial overlap in FLDA scores, the projection maximized between-class separation relative to within-class variance, yielding robust linear decisions. Conclusions: Unlike earlier FLDA-only pipelines and wavelet–entropy–ANN approaches, our study (1) employs SWT (undecimated, shift-invariant) rather than DWT to stabilize sub-band features on short resting segments, (2) uses FLDA as a supervised projection to mitigate small-sample covariance pathologies before classification, (3) provides band-specific discriminative insight (β > γ/θ) under a subject-wise protocol, and (4) targets low-compute deployment. These choices yield a reproducible baseline with competitive accuracy and clear clinical interpretability. Future work will benchmark kernel/regularized discriminants and lightweight deep models as cohort size and compute permit.

1. Introduction

Neurodevelopmental Autism Spectrum Disorder (ASD) is a complex cranial nerve disorder that significantly impacts the growth and behavioral development of affected children [1]. This syndrome is characterized by repetitive behavioral patterns [2]. Based on research, many children with autism remain undiagnosed or do not receive appropriate treatment, largely due to environmental factors and the inadequacy of diagnostic equipment [3]. With advances in technology, especially in the use of an electroencephalogram (EEG) [4,5], detecting autism in children is potentially becoming easier. This can be achieved by recording brain signals via a Brain–Computer Interface (BCI) with a collection of electrodes placed on the scalp [4]. However, with complex EEG signals, this research requires a Machine Learning (ML) scheme to analyze signal patterns that are difficult to analyze. The ML technique effectively utilizes real-time processing methods for EEG signals [5]. Automation of the detection method reduces reliance on manual inspection by highly specialized experts, leading to quicker responses and potentially enhancing the overall outcomes of classification management [6].
The integration of ML into biomedical fields is not only revolutionizing diagnostics but also optimizing operations across various industries, including manufacturing, where it has proven effective in cost estimation [7]. The potential of ML to enhance the accuracy of autism predictions is significant, potentially saving time and resources in clinical settings [8].
Recent studies have demonstrated the effectiveness of ML in detecting brain abnormalities through EEG. For instance, Shi et al. established a predictive model differentiating between autism and neurotypical development, achieving an accuracy of 85.4% using k-Nearest Neighbors (kNN) pipeline on the KAU dataset [9]. Further advancements by Aljalal et al. employed Variational Mode Decomposition (VMD) in conjunction with kNN to detect Mild Cognitive Impairment (MCI), achieving a remarkable accuracy of 99.81% [10]. Similarly, Al-Jumaili et al. demonstrated the utility of various ML algorithms with feature extraction methods for detecting epileptic seizures, obtaining an accuracy of 96% with Support Vector Machine (SVM) [11]. In the context of autism detection, Melinda et al. utilized SVM combined with Continuous Wavelet Transform (CWT) to achieve a 95% classification accuracy for autistic children [12]. Chaddad et al. provided a comprehensive analysis of various EEG signal classification techniques, underscoring the potential of ML algorithms to achieve high accuracy [13].
Among the various ML methods, Fisher Linear Discriminant Analysis (FLDA) is particularly notable for its application in statistics and pattern recognition. This technique effectively classifies data while maximizing class separation, making it an ideal choice for our study [14,15]. This method was chosen because it can produce good accuracy and improve system performance [16,17]. FLDA is a supervised learning technique that is used to reduce dimensions while classifying data by maximizing the distance between classes [18]. We extend the prior DWT + LDA pipeline by adopting SWT and FLDA to address two known limitations. First, SWT (undecimated) is shift-invariant, mitigating the translation-variance and down-sampling artifacts of DWT and yielding more stable EEG sub-band features for short, non-stationary segments [13]. Second, we employ FLDA as a supervised projection (≤C−1 components) rather than the LDA classifier, to reduce dimensionality and avoid small-sample covariance instability typical in high-dimensional EEG [19]. This method is combined with the Stationary Wavelet Transform (SWT) method to filter and extract signals at certain frequencies. This research builds upon a previous journal article that utilized Discrete Wavelet Transform (DWT) extraction and Linear Discriminant Analysis (LDA) in processing EEG data from individuals with autism and those without [16]. The difference lies in the component extraction method used, which utilizes only one detailed component, represented by a beta signal. The results of this follow-up research include the classification of three-component extractions and the accuracy of implementing FLDA on autistic and normal EEG signals.
Stationary Wavelet Transform (SWT) is a digital signal processing technique that simultaneously filters and extracts signals to certain frequencies through Low-Pass Filter (LPF) and High-Pass Filter (HPF) processes [20]. SWT extraction will produce three attributes in the form of levels 3, 4, and 6 components, which are gamma, beta, and theta signals, respectively. This component is a feature chosen to test the performance of the FLDA system in classifying signals [13]. The use of this wavelet technique can also help increase noise suppression in EEG signals through denoising techniques. The advantage of using the SWT technique is that it does not require down-sampling or up-sampling. In this case, the number of coefficients and diffusion levels remains constant, or the length of the transformation signal at each level is fixed [21].
Research on ASD EEG is increasingly adopting deep-learning architectures and advanced time–frequency/connectivity pipelines [13], with representative ASD studies using spectro-temporal features and engineered classifiers reporting strong performance [12,22,23]. In this context, we position our SWT–FLDA approach as a compact, interpretable baseline evaluated subject-wise, which complements modern models by (i) reducing overfitting risk in small cohorts, (ii) providing band-level discriminative insight (β vs. γ/θ), and (iii) aligning with resource-constrained clinical deployment.
The objective of this study is to develop a robust classification system for EEG signals from autistic and typically developing children by combining FLDA with SWT. This approach aims to enhance signal extraction and noise suppression, thereby facilitating more accurate classification of ASD and typical EEG data. The study’s major contributions can be summarized as follows. This research presents a novel approach that integrates Fisher Linear Discriminant Analysis (FLDA) with the Stationary Wavelet Transform (SWT) for classifying EEG signals from autistic and normal children. The study offers an in-depth evaluation of the system’s performance based on confusion matrix metrics, including accuracy, precision, recall, and F1-score. A comparative analysis of FLDA’s effectiveness in classifying multiple EEG components (gamma, beta, and theta signals) extracted through SWT highlights the method’s versatility.
This research aims to enhance clinical diagnostics and improve early intervention strategies for children with autism by providing a reliable and automated detection method for ASD through EEG analysis. Our SWT–FLDA pipeline is lightweight and interpretable, suitable for resource-constrained clinics (e.g., in Indonesia) where routine deployment of large deep-learning models is impractical.

2. Materials and Methods

This study applied the SWT and FLDA methods to process the data (see Figure 1). SWT was used to denoise signals and extract signals into several different frequencies. The results of SWT extraction are presented in the form of three frequency attributes, corresponding to components at levels 3, 4, and 6, which represent gamma, beta, and theta signals, respectively. The denoising or noise suppression process in SWT utilized the thresholding feature. The three levels of filtering and decomposition results were later used as features for FLDA classification testing. Meanwhile, FLDA was used to classify signals in autistic and normal syndromes. This research aims to compare the accuracy results of EEG classification for autistic and normal individuals using the FLDA technique, based on head condition in terms of frequencies at each level.

2.1. Datasets

This study used an EEG dataset provided by King Abdulaziz University (KAU), Jeddah, Saudi Arabia [24]. This dataset comprises 16 EEG signals, including 8 EEG samples from autistic children and 8 EEG samples from typically developing children, recorded using the BCI2000 viewer in a .dat format with a 16 × 16 (trial × channel) structure. Each sample contains 16 channels consisting of FP1, F3, F7, T3, T5, O1, C4, FP2, FZ, F4, F8, C3, CZ, PZ, OZ, and O2, following the international 10–20 system. We retained the 16-channel 10–20 montage (Fp1…O2) to bilaterally sample regions implicated in ASD (frontal, temporal, central, parietal, occipital), given prior ASD EEG reports of reduced/atypical frontal and fronto-posterior coherence and altered spectral power (decreased alpha with increases in theta/beta, with reports of elevated high-frequency activity), which motivates this coverage [25,26,27].
The dataset ensures the confidentiality of all participants by excluding any personal identification information; ASD subtypes are not annotated in the publicly available description, and the annotations consist of group labels (ASD vs. control) rather than per-event clinical markers. The recorded EEG signals represent a diverse demographic, including eight subjects with ASD (five males and three females, aged 6 to 20 years; total recorded duration of 4104.2 s) and eight control subjects (all males, aged 9 to 13 years; total recorded duration of 4534.9 s). The KAU Hospital clinical team assigned group labels (ASD vs. control); we used these labels as the ground truth and performed no re-labeling. The public dataset provides group-level annotations only (no per-event clinical markers and no ASD subtypes). The public description does not disclose the specific diagnostic instruments or protocols used to establish these labels (e.g., ADOS/ADI-R, DSM-5), which we note as a limitation of the source data. This inclusion of a range of ages and both genders among the ASD participants increases the validity and generalizability of the research findings. The EEG recordings were captured using Ag/AgCl electrodes in conjunction with a g.tec EEG deviceand a g.tec USB amplifier (g.tec medical engineering GmbH, Schiedlberg, Austria), and BCI2000 (Wadsworth Center, Albany, NY, USA) while the subjects were in a relaxed state. The recordings were digitized at 256 Hz with an acquisition band-pass of 0.1–60 Hz and a 60 Hz notch. For researchers interested in accessing this dataset, inquiries can be directed to Dr. Mohammed Jaffer Alhaddad via email at malhaddad@kau.edu.sa.
Annotations were performed by qualified clinical staff following standard diagnostic procedures for ASD at KAU Hospital, although specific instruments (e.g., ADOS/ADI-R) were not disclosed in the public dataset [24]. As reported in [24], the original EEG data collection at KAU Hospital was conducted under the approval and oversight of the KAU Ethics Committee. Written informed consent was obtained from all participants or their legal guardians prior to data acquisition, and all recordings were fully anonymized before release. The present work involved no new data collection or direct contact with human participants; only secondary analysis of the anonymized dataset was performed.

2.2. Data Input Preprocessing

In this stage, EEG datasets were input with a sampling frequency of 256 Hz. At the data import stage, the function type used is BCI2kReader (the BCI2000 file I/O utility), because the EEG signal recording format has an output in BCI2000 “.dat” format and was acquired using the BCI2000 system. BCI2000 is a modular platform consisting of Operator, Source, Signal Processing, and Application modules. In our setup, the gUSBamp Source module interfaced the g.tec USB amplifier, and the native .dat files included a header with acquisition parameters and channel labels. Of these sixteen recorded channels, fifteen were analyzed (Fp1…Oz); channel O2 was excluded from analysis. All subsequent preprocessing, SWT feature extraction, and FLDA classification were performed on this 15-channel matrix. Then, to obtain information from the signal, event information from each recorded channel was imported, namely, 16 events per data sample. This process produced a data length of 16,000 for 15 EEG samples. Then, the data was exported in double (number) form to produce a signal output in .txt data format.

2.3. Stationary Wavelet Transform (SWT)

Feature extraction and selection techniques significantly contribute to improving BCI capabilities across various applications, including cognitive enhancement and neuroprosthetics [28]. This research utilized SWT to decompose and filter signals into smaller dimensions through the Low-Pass Filter (LPF) process, which filters low-frequency signals, and the High-Pass Filter (HPF) process, which filters high-frequency signals. SWT is a modified method of Discrete Wavelet Transform (DWT), also known as Undecimate Wavelet Transform. SWT does not include down-sampling or up-sampling, so the number of coefficients and decomposition levels remains constant at each level, resulting in faster computation. The filtering process produces subsample results as described in the following formula [29]:
Ψ b a = 1 a ψ * t b a , a > 0 , b ϵ R
where ψ b a is the mother wavelet, a is a scale parameter, b is a shift parameter, and ψ*(t) is the complex conjugate of the mother wavelet. The discrete values and shift parameters are a = 2j and b = 2j k.
SWT utilizes a thresholding function to remove noise by setting small-amplitude wavelet coefficients (representing noise) to zero or a specified threshold value [30]. The choice of threshold value must be optimal. A threshold value that is too small can leave noise, while one that is too large can erase valid signal information. Therefore, selecting an optimal threshold value is crucial to minimize the Mean Squared Error (MSE) [31]. There are two basic wavelet shrinkage functions, namely, hard threshold and soft threshold [32]. We selected SWT because it is undecimated/translation-invariant, yielding shift-robust multi-scale representations and enabling wavelet-shrinkage denoising without training data, which is appropriate for small-sample EEG. Prior work also showed SWT’s advantage over DWT for EEG artifact removal [33,34]. While deep-learning super-resolution methods (e.g., Wang et al., 2023) can enhance time–frequency concentration, they typically require larger datasets and careful tuning; for our small-sample EEG cohort, SWT provided a reproducible, non-learned alternative with lower overfitting risk [35]. We restricted features to SWT levels corresponding to θ (level 6), β (level 4), and γ (level 3) to balance physiological relevance with our small-sample setting and to limit dimensionality (reducing overfitting risk). Recent ASD EEG studies have indicated robust alterations in higher-frequency features (e.g., decreased α and increased β/γ), whereas δ is highly state-dependent and prone to low-frequency/ocular contamination in short 4 s epochs. Accordingly, α/δ coefficients were decomposed but not used as predictive inputs in the present model. We note this as a limitation and a target for future work (e.g., δ/α ratios) [36,37].

2.4. Hard Threshold

A hard threshold is a linear function that removes coefficients below a threshold value determined by the noise variance. The hard threshold filter coefficient is obtained using
Δ λ H x = x ,   i f   x > λ   k e e p 0 ,   o t h e r w i s e   k i l l
This wavelet is known as a “keep or kill” procedure. Thresholding is better known because the thresholding function includes discontinuity; therefore, x values above the lambda threshold are not affected. Conversely, if they are below the lambda threshold, they will be changed to 0.

2.5. Soft Threshold

The soft thresholding function is typically used because it remains continuous, coefficients with d j n > λ j are shrunk toward zero by λ j , while those with d j n < λ j are set to zero. The soft threshold filter can be stated as follows:
d j ^ n = sign d j n max d j n λ j , 0
where d j ^ n denotes the SWT detail coefficient at level j , λ j is the level-dependent threshold, and d j ^ n is the thresholded coefficient. The primary difference between the hard and soft threshold functions is that the coefficient below the threshold is set to zero in the hard threshold. Thus, this will weaken the signal oscillations, and the reconstructed signal will have poor smoothness [38]. Based on the graph, the signal created will follow the shape of the original signal; however, if the coefficient is less than or equal to the threshold value, it will be set to zero, which can compromise the interpretation of the data. Meanwhile, with soft thresholding, the graph created will follow the shape of the signal without destroying the data. The following is an SWT chart for decomposing and extracting signals into smaller frequencies.
Based on Figure 2, the EEG signal will be extracted, decomposed, and filtered into 6 levels through a High-Pass Filter (detail component) and Low-Pass Filter (approximation component) process. The mother wavelet is Daubechies-4 (db4), selected for its compact support, orthogonality, four vanishing moments, and strong time–frequency localization, properties widely leveraged in EEG denoising and sub-band feature extraction [39]. Prior Q1/Q2 works have commonly used db4 for EEG decomposition/rhythm features, and recent analyses emphasized that the mother wavelet substantially influences the informative content of wavelet features. This wavelet has proven to be very suitable for processing raw EEG signals and to produce the best signal filtering [40]. Consistent with these reports, db4 provided stable coefficients for our data and yielded robust filtering performance [39]. Next, the three detailed components produced in the form of signals at levels 3, 4, and 6 were utilized, corresponding to gamma, beta, and theta signals, respectively. These three levels were chosen because they can represent the conditions of autistic and normal children when they are sleeping, active, and focused.
The denoising process was carried out by utilizing threshold and sampling functions through the Low-Pass Filter and High-Pass Filter processes. This type of soft thresholding was chosen to eliminate noise because it can smooth the reconstructed signal without distorting it, thereby preserving the signal integrity [20]. Therefore, the resulting signal graph will retain the same information as the original signal graph.

2.6. Fisher Linear Discriminant Analysis (FLDA)

We used Fisher Linear Discriminant Analysis (FLDA), a supervised linear method that learns a projection w , maximizing between-class separation relative to within-class variance. For two classes, the Fisher criterion is
J w = w S b w w S w w
where S b   and S w are the between- and within-class scatter matrices; the solution reduces to the generalized eigenproblem S b v = λ S w v , and classification proceeds with a linear decision boundary in the projected space [41,42,43]. Here, we used FLDA explicitly as a supervised projection (feature extractor). In contrast, the classical LDA classifier models class-conditional Gaussians with a shared, full-rank covariance estimated in the original feature space; under small-sample and correlated features, this pooled covariance can be ill-conditioned/singular, degrading LDA classification. FLDA sidesteps this by learning C 1   discriminant components that maximize S b / S w and then applying a linear rule in the projected space [19].
FLDA projects feature vectors into a direction that pushes class means apart while compressing within-class spread. In binary cases, this yields a one-dimensional representation where classes are maximally separated, and in multi-class cases, up to C 1   discriminant components are obtained [41,42,43]. In practice, features were z-normalized before FLDA to align scales across SWT features. This projection-first design is motivated by the Small-Sample-Size (SSS) regime typical in EEG studies (n ≪ d), for which recent work refines LDA/trace-ratio formulations precisely to address SSS pathologies, supporting our choice of FLDA projection for stability.
In our pipeline, SWT features served as inputs. We used a subject-wise, stratified 70/30 train–test split (Figure 3). Given the small cohort and the absence of hyperparameter tuning (FLDA has a closed-form solution), we did not allocate a separate validation split. All preprocessing (e.g., z-normalization) was fit on the training set only to prevent leakage, and performance on the held-out test set was summarized using accuracy, precision, recall, and F1-score. This follows recommended practice in small-sample biomedical ML (hold-out when no tuning; nested cross-validation when tuning is required) [44,45].
In EEG classification, commonly used models include linear (FLDA/LDA, logistic regression), SVM, k-NN, tree ensembles, and deep networks (CNN/LSTM/graph) [13] In view of our small-sample setting and hand-crafted SWT features, we adopted FLDA as a low-variance, interpretable baseline with minimal hyperparameter tuning and a closed-form solution, which is advantageous for reproducibility [14]. Classical ML remains competitive for ASD EEG when data are limited, as shown in recent open-access studies [8]. Exploration of SVM/ensembles/DL with larger cohorts remains for future work [13].
We intentionally employed a single, closed-form FLDA classifier to obtain a stable, interpretable subject-wise baseline in this small cohort. Broader model families (k-NN/SVM/RF/CNN/LSTM) have been well documented in prior work [8,12,13,14], but direct numerical comparisons are not meaningful here due to different datasets and validation protocols; therefore, we treat them as context rather than benchmarks.

3. Results

This section reports results for classifying autism vs. normal EEG using SWT features and FLDA.

3.1. Data Input Preprocessing Results

In processing normal and autistic EEG data, only 15 of the 16 active channels were read, namely, FP1, F3, F7, T3, T5, O1, C4, FP2, FZ, F4, F8, C3, CZ, PZ, and OZ. Channel O2 is not read because it is considered a grounding channel. The input process involved 16 iterations for 16 patients with 15 channels, a total of 240,000 features. The dataset comprises 16,000 data points, divided into 8000 normal EEG data points and 8000 autistic EEG data points. Labels on EEG channels indicate electrode positions on the head, such as F (Frontal), Fp (Frontopolar), C (Central), O (Occipital), and T (Temporal).

3.2. Stationary Wavelet Transform (SWT) Result

In this study, SWT was used to extract and filter signals to certain frequencies using Daubechies 4 (db-4) as the mother wavelet. Extraction yielded three detailed components, namely, levels 3, 4, and 6, which represent the gamma signal, beta, and theta, respectively. The results obtained from feature extraction and denoising for the normal and ASD classes are presented in Table 1, Table 2 and Table 3 below.
The results of the SWT data in the frequency range of 32–64 Hz (level 3) indicate that the patient’s brain activity is in a focused state (Table 1 and Table 4). The frequency range of 16–32 Hz (level 4) shows normal brain activity (Table 2 and Table 5). The frequency range of 4–8 Hz (level 6) shows brain activity in sleep (Table 3 and Table 6). From the results obtained, the normal EEG is predominantly higher and more stable compared to the autistic EEG. The electrical activity of normal children experiences greater fluctuations, meaning that the channels in the normal EEG are more active. Meanwhile, frequency fluctuations produced by autistic children experience inconsistent up-and-down changes and tend to be smaller. These conditions can affect the accuracy of FLDA classification performance.

3.3. Fisher Linear Discriminant Analysis (FLDA) Result

FLDA’s primary goal is to effectively separate different features by transforming data into a lower-dimensional space, maximizing the distance between classes, and minimizing the distance within the same classes. FLDA takes normal and autistic EEG datasets, each of which has 15 active channel parameters, and converts them into one representative value. FLDA creates new axes based on data characteristics, which reduces variance and maximizes the distance between classes. The results of normal and autistic EEG classification show good separation based on each feature, as shown in Figure 4.
Building on this, FLDA projects the SWT-based EEG feature vectors onto a single discriminant axis that maximizes between-class separation (autism vs. normal) while minimizing the within-class variance. Each sample is mapped to a discriminant score and classified in a binary setting with autism as the positive class and normal as the negative class.
Figure 4 visualizes the FLDA outputs by SWT component—level 3 (γ, red), level 4 (β, green), and level 6 (θ, blue). The normal (–) class shows clear separation; autism (+) exhibits some overlap due to non-linear structure, but this does not preclude accurate discrimination, as quantified below.
Table 4 reports the test-set confusion matrices (subject-wise, stratified 70/30 split; n = 4800 segments per level; autism = 2400, normal = 2400) with autism as the positive class. Performance is highest for level 4 (β), with accuracy = 0.950 (TP = 2280; TN = 2280), followed by level 3 (γ), with accuracy = 0.920 (TP = 2208; TN = 2208), and level 6 (θ), with accuracy = 0.850 (TP = 2040; TN = 2040). These results confirm that β-band features are the most discriminative on this dataset.

3.4. Hyperparameters

Signals were processed as described in Section 2; Stationary Wavelet Transform (SWT) features were extracted at levels 3, 4, and 6. All features were z-normalized (mean 0, unit variance) using statistics computed on the training set only.
We employed Fisher’s Linear Discriminant (FLDA) as a supervised projection; for C classes, up to C 1 discriminant components were retained. A linear decision rule with equal class priors was applied in the projected space. FLDA is closed-form and does not require iterative hyperparameter tuning.
A subject-wise, stratified 70 / 30   train–test split was used (no subject overlap). Preprocessing transforms were fit on the training set and applied to the test set to avoid leakage. The random seed was fixed (e.g., 42) to ensure reproducibility. No early stopping or model selection was performed.
All experiments were run using standard scientific Python libraries (Python 3.10); code-level settings (e.g., data loaders, batch sizing if applicable) follow defaults, as no iterative optimization was required by FLDA.

3.5. Evaluation Criteria

Table 5 summarizes the class-wise metrics computed under the same convention (autism = positive): recall/sensitivity (autism), specificity (normal), precision (autism), F1-score (autism), and accuracy. In line with the confusion matrices, level 4 (β) attains precision = 0.950, recall = 0.950, F1 = 0.950, and accuracy = 0.950; level 3 (γ) yields precision = 0.908, recall = 0.932, F1 = 0.920, and accuracy = 0.920; and level 6 (θ) yields precision = 0.846, recall = 0.854, F1 = 0.850, and accuracy = 0.850.

4. Discussion

This study aimed to evaluate the application of the FLDA system in classifying EEG signals from individuals with autism and those without autism, based on accuracy. This classification system separates signals based on detailed components at each level. The results show an accuracy of above 90%, indicating a high level of classification accuracy. FLDA is a suitable choice for ML techniques on autistic and normal EEG signals.
Although the autism and normal score distributions overlap in the FLDA-projected space (Figure 4), linear separability was not required. FLDA chooses a projection that maximizes between-class separation relative to pooled within-class variance, after which a single threshold yields a robust linear decision; this explains the high β-band performance (precision/recall/accuracy = 0.95; Table 4 and Table 5) despite partial overlap. Residual errors near the overlap likely reflect non-linear structure and will be explored with kernel/regularized discriminants under the same subject-wise protocol.
Neurophysiological interpretation: From a neurophysiology perspective, the relatively higher and more stable activity observed in the normal group versus the smaller, more inconsistent fluctuations in the autism group may reflect atypical regulation of cortical oscillations in ASD. Prior work reports group differences in spectral power and connectivity across alpha–beta–gamma rhythms in ASD (with notable heterogeneity), supporting the idea that oscillatory dynamics can serve as informative features for classification. Proposed mechanisms include imbalances in excitatory–inhibitory (E/I) signaling and thalamocortical circuit dysfunction, which can alter the stability and amplitude of resting rhythms. Therefore, we interpret our findings as compatible with these mechanisms while acknowledging cohort size and age/sex differences. Targeted neurophysiological studies will be required to confirm causality [36].
Our study built on existing research in the field of EEG-based classification, aiming to improve ASD detection. Several recent studies related to this topic have identified advantages and drawbacks, which can be compared with the findings of this study. Compared with previous studies, which focused on various feature extraction methods and classifiers, our approach uses SWT for feature extraction and FLDA for classification. This combination offers several advantages, including effective dimensionality reduction of EEG data and accurate ASD classification. By utilizing SWT, important features can be extracted from EEG signals, thereby enabling a better representation of ASD-related brain activity patterns. However, this approach may have some drawbacks, including the use of relatively small datasets and limited FLDA capabilities. Despite these challenges, this study makes a significant contribution to the classification of ASD EEG signals with competitive accuracy while also emphasizing the efficiency of the simpler FLDA method compared to several other complex techniques used in previous studies. This research supports the development of effective and efficient automated diagnosis systems in the context of neurological disorders such as ASD. This resource-aware design supports practical adoption in low-compute clinics. Future work will explore kernel/regularized discriminants and deep learning on larger, balanced cohorts.
To situate our findings within the literature, Table 6 synthesizes representative ASD-EEG classification studies already cited in Section 1, reporting the feature/classifier pair, dataset context, and the authors’ reported accuracy. Because cohorts, acquisition protocols, and evaluation schemes differ across studies, these figures are not directly comparable; the table serves as context rather than a head-to-head benchmark. On the same KAU dataset, the Butterworth → ICA → KNN baseline reports 85.4% accuracy [9], CWT → SVM achieves 95% [12], and spectrogram/STFT pipelines reach 95.25% [22]. Under an explicit subject-wise split on KAU, our SWT (levels 3/4/6) → FLDA pipeline attains 95% accuracy in the β-band. This side-by-side view underscores the interpretability and computational efficiency of our approach alongside competitive performance.
Limitations. Our findings should be viewed in light of sex/age imbalance (ASD: five males/three females, 6–20 years; controls: eight males, 9–13 years) and a small subject count (n = 16). We used a subject-wise 70/30 hold-out to avoid leakage, but broader validation will benefit from sex/age-matched recruitment, subject-wise k-fold/LOSO and external testing. Analyses used 15 of 16 channels (O2 excluded), and future work will examine robust channel/reference choices and non-linear/regularized discriminants within the same evaluation protocol. A single-classifier design constrains the breadth of comparison; future work will perform subject-wise benchmarking with SVM/k-NN/RF and lightweight deep models as the cohort size and balance improve.

5. Conclusions

This study demonstrates that the combination of the SWT and FLDA methods significantly enhances the accuracy of EEG signal processing for both autistic and typically developing subjects, achieving an accuracy rate of over 90%. The key classification parameters included gamma signals (32–64 Hz) at level 3, representing focused brain activity, beta signals (16–32 Hz) at level 4, indicating normal brain activity, and theta signals (4–8 Hz) at level 6, corresponding to brain activity during sleep. These parameters were extracted through SWT using the Daubechies 4 (db-4) mother wavelet, along with soft thresholding for effective filtering. The system demonstrated high accuracy in classifying normal EEG signals at 96%, while the classification accuracy for autistic EEG signals reached 93%, with the slight decrease attributed to overlapping scatter in the data. These results were validated using the confusion matrix method during training and testing. The findings of this research can be applied in practical scenarios, particularly in improving the early detection and classification of autism through EEG analysis.
Additionally, this framework can be extended to more complex brain signal analyses in other neurological disorders [46,47,48], with potential wearable applications [49]. However, there are limitations, including the overlap of certain EEG patterns in autistic subjects, which slightly reduces accuracy. Future research could explore more advanced filtering techniques or a larger, more diverse dataset to address this challenge. Further exploration into more sophisticated classification algorithms may also yield higher accuracy and more robust results.

Author Contributions

Conceptualization, F.F. and M.M.; methodology, M.M.; software, F.F.; validation, F.F., M.M. and P.D.P.; formal analysis, F.F. and E.E.; data curation, M.M.; writing—original draft preparation, M.M. and A.R.; writing—review and editing, F.F., E.E. and A.R.; funding acquisition, F.F., M.M. and P.D.P. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Ministry of Higher Education, Research and Development of Indonesia, through the RKI 2025 scheme of Universitas Sumatera Utara, Universitas Syiah Kuala, and Universitas Indonesia, grant number 28/UN5.4.10.K/PT.01.03/RKI/2025, dated 14 April 2025.

Institutional Review Board Statement

The data comes from the dataset, so it is not applicable.

Informed Consent Statement

Written informed consent was obtained from all participants or their legal guardians by KAU Hospital prior to EEG acquisition; all data were fully anonymized before sharing with researchers.

Data Availability Statement

Restrictions apply to the availability of these data. Data were obtained from King Abdulaziz University (KAU) Hospital, Jeddah, Saudi Arabia and are available from the data owner with permission (contact: Dr. Mohammed Jaffer Alhaddad, malhaddad@kau.edu.sa).

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
ASDAutism Spectrum Disorder
EEGElectroencephalography
SWTStationary Wavelet Transform
DWTDiscrete Wavelet Transform
FLDAFisher’s Linear Discriminant Analysis
LDALinear Discriminant Analysis
Db4Daubechies-4 (mother wavelet)
BCIBrain–Computer Interface
BCI2000Brain–Computer Interface 2000 platform (software)
STFTShort-Time Fourier Transform
CWTContinuous Wavelet Transform
CSPCommon Spatial Pattern
LBPLocal Binary Pattern
k-NNk-Nearest Neighbors
SVMSupport Vector Machine
RFRandom Forest
CNNConvolutional Neural Network
RNNRecurrent Neural Network
LSTMLong Short-Term Memory (network)
GRUGated Recurrent Unit
ANNArtificial Neural Network
VMDVariational Mode Decomposition
AUCArea Under the Curve
ROC-AUCArea Under the Receiver Operating Characteristic Curve
LOSOLeave One Subject Out
MCIMild Cognitive Impairment
KAUKing Abdulaziz University

References

  1. Spinazzi, N.A.; Velasco, A.B.; Wodecki, D.J.; Patel, L. Autism Spectrum Disorder in Down Syndrome: Experiences from Caregivers. J. Autism Dev. Disord. 2024, 54, 1171–1180. [Google Scholar] [CrossRef] [PubMed]
  2. Derbali, M.; Jarrah, M. Randhawa. Autism Spectrum Disorder Detection: Video Games based Facial Expression Diagnosis using Deep Learning. Int. J. Adv. Comput. Sci. Appl. 2023, 14, 110–119. [Google Scholar] [CrossRef]
  3. Zeidan, J.; Fombonne, E.; Scorah, J.; Ibrahim, A.; Durkin, M.S.; Saxena, S.; Yusuf, A.; Shih, A.; Elsabbagh, M. Global prevalence of autism: A systematic review update. Autism Res. 2022, 15, 778–790. [Google Scholar] [CrossRef]
  4. Nurfirdausi, A.F.; Apsari, R.A.; Wijaya, S.K.; Prajitno, P.; Ibrahim, N. Wavelet Decomposition and Feedforward Neural Network for Classification of Acute Ischemic Stroke based on Electroencephalography. Int. J. Technol. 2022, 13, 1745–1754. [Google Scholar] [CrossRef]
  5. Kumar, P.; Chauhan, S.; Awasthi, L.K. Artificial Intelligence in Healthcare: Review, Ethics, Trust Challenges & Future Research Directions. Eng. Appl. Artif. Intell. 2023, 120, 105894. [Google Scholar] [CrossRef]
  6. Golla, S.; Suman, M. Automated Seizure Detection in Neonatal EEG using Signal Processing Algorithms. J. Adv. Res. Appl. Sci. Eng. Technol. 2023, 31, 220–227. [Google Scholar] [CrossRef]
  7. Ma’ruf, A.; Nasution, A.A.R.; Leuveano, R.A.C. Machine Learning Approach for Early Assembly Design Cost Estimation: A Case from Make-to-Order Manufacturing Industry. Int. J. Technol. 2024, 15, 1037–1047. [Google Scholar] [CrossRef]
  8. Farooq, M.S.; Tehseen, R.; Sabir, M.; Atal, Z. Detection of autism spectrum disorder (ASD) in children and adults using machine learning. Sci. Rep. 2023, 13, 9605. [Google Scholar] [CrossRef]
  9. Shi, L.; Ma, L.; Zhao, J.; Kuang, Z.; Wang, S.; Yang, H.; Wang, H.; Han, Q.; Sun, L. TFSNet: A Time–Frequency Synergy Network Based on EEG Signals for Autism Spectrum Disorder Classification. Brain Sci. 2025, 15, 684. [Google Scholar] [CrossRef] [PubMed]
  10. Aljalal, M.; Molinas, M.; Aldosari, S.A.; AlSharabi, K.; Abdurraqeeb, A.M.; Alturki, F.A. Mild cognitive impairment detection with optimally selected EEG channels based on variational mode decomposition and supervised machine learning. Biomed. Signal Process Control 2024, 87, 105462. [Google Scholar] [CrossRef]
  11. Al-Jumaili, S.; Duru, A.D.; Ibrahim, A.A.; Uçan, O.N. Investigation of Epileptic Seizure Signatures Classification in EEG Using Supervised Machine Learning Algorithms. Trait. Signal 2023, 40, 43–54. [Google Scholar] [CrossRef]
  12. Melinda, M.; Juwono, F.H.; Enriko, I.K.A.; Oktiana, M.; Mulyani, S.; Saddami, K. Application of Continuous Wavelet Transform and Support Vector Machine for Autism Spectrum Disorder Electroencephalography Signal Classification. Radioelectron. Comput. Syst. 2023, 3, 73–90. [Google Scholar] [CrossRef]
  13. Chaddad, A.; Wu, Y.; Kateb, R.; Bouridane, A. Electroencephalography Signal Processing: A Comprehensive Review and Analysis of Methods and Techniques. Sensors 2023, 23, 6434. [Google Scholar] [CrossRef]
  14. Chen, D. Improved empirical mode decomposition bagging RCSP combined with Fisher discriminant method for EEG feature extraction and classification. Heliyon 2024, 10, e28235. [Google Scholar] [CrossRef] [PubMed]
  15. Helm, H.; De Silva, A.; Vogelstein, J.T.; Priebe, C.E.; Yang, W. Approximately Optimal Domain Adaptation with Fisher’s Linear Discriminant. Mathematics 2024, 12, 746. [Google Scholar] [CrossRef]
  16. Melinda, M.; Oktiana, M.; Yunidar, Y.; Nabila, N.H.; Enriko, I.K.A. Classification of EEG Signal using Independent Component Analysis and Discrete Wavelet Transform based on Linear Discriminant Analysis. Int. J. Inform. Vis. 2023, 7, 830–838. [Google Scholar] [CrossRef]
  17. Mabrouk, B.; Hamida, A.B.; Mabrouki, N.; Bouzidi, N.; Mhiri, C. A Novel Approach to Perform Linear Discriminant Analyses for a 4-way Alzheimer’s Disease Diagnosis Based on an Integration of Pearson’ s Correlation Coe cients and Empirical Cumulative Distribution Function. Multimed. Tools Appl. 2024, 83, 76687–76703. [Google Scholar] [CrossRef]
  18. Kächele, F.; Schneider, N. Cluster Validation Based on Fisher’s Linear Discriminant Analysis. J. Classif. 2024, 42, 54–71. [Google Scholar] [CrossRef]
  19. Qu, L.; Pei, Y. A Comprehensive Review on Discriminant Analysis for Addressing Challenges of Class-Level Limitations, Small Sample Size, and Robustness. Processes 2024, 12, 1382. [Google Scholar] [CrossRef]
  20. Hermawan, I.; Sevani, N.; Abka, A.F.; Jatmiko, W. Denoising Ambulatory Electrocardiogram Signal Using Interval- Dependent Thresholds-based Stationary Wavelet Transform. Int. J. Inform. Vis. 2024, 8, 742–750. [Google Scholar] [CrossRef]
  21. Geng, X.; Li, D.; Chen, H.; Yu, P.; Yan, H.; Yue, M. An improved feature extraction algorithms of EEG signals based on motor imagery brain-computer interface. Alex. Eng. J. 2022, 61, 4807–4820. [Google Scholar] [CrossRef]
  22. Tawhid, M.N.A.; Siuly, S.; Wang, H. Diagnosis of autism spectrum disorder from EEG using a time-frequency spectrogram image-based approach. Electron. Lett. 2020, 56, 1372–1375. [Google Scholar] [CrossRef]
  23. Alturki, F.A.; Aljalal, M.; Abdurraqeeb, A.M.; Alsharabi, K.; Al-Shamma’a, A.A. Common Spatial Pattern Technique with EEG Signals for Diagnosis of Autism and Epilepsy Disorders. IEEE Access 2021, 9, 24334–24349. [Google Scholar] [CrossRef]
  24. Alhaddad, M.J.; Kamel, M.I.; Malibary, H.M.; Alsaggaf, E.A.; Thabit, K.; Dahlwi, F.; Hadi, A.A. Diagnosis autism by Fisher Linear Discriminant Analysis FLDA via EEG. Int. J. Bio-Sci. Bio-Technol. 2012, 4, 45–54. [Google Scholar]
  25. Murias, M.; Webb, S.J.; Greenson, J.; Dawson, G. Resting state cortical connectivity reflected in EEG coherence in individuals with autism. Biol. Psychiatry 2007, 62, 270–273. [Google Scholar] [CrossRef] [PubMed]
  26. Orekhova, E.V.; Stroganova, T.A.; Nygren, G.; Tsetlin, M.M.; Posikera, I.N.; Gillberg, C.; Elam, M. Excess of High Frequency Electroencephalogram Oscillations in Boys with Autism. Biol. Psychiatry 2007, 62, 1022–1029. [Google Scholar] [CrossRef] [PubMed]
  27. Coben, R.; Clarke, A.R.; Hudspeth, W.; Barry, R.J. EEG power and coherence in autistic spectrum disorder. Clin. Neurophysiol. 2008, 119, 1002–1009. [Google Scholar] [CrossRef]
  28. Talha, A.Z.; Eissa, N.S.; Shapiai, M.I. Applications of Brain Computer Interface for Motor Imagery Using Deep Learning: Review on Recent Trends. J. Adv. Res. Appl. Sci. Eng. Technol. 2024, 40, 96–116. [Google Scholar] [CrossRef]
  29. Elshekhidris, I.H.; MohamedAmien, M.B.; Fragoon, A. Wavelet Transforms for Eeg Signal Denoising and Decomposition. Int. J. Adv. Signal Image Sci. 2023, 9, 11–28. [Google Scholar] [CrossRef]
  30. Aburakhia, S.; Shami, A.; Karagiannidis, G.K. On the Intersection of Signal Processing and Machine Learning: A Use Case-Driven Analysis Approach. arXiv 2024, arXiv:2403.17181. [Google Scholar] [CrossRef]
  31. Ming, F.; Long, H. Partial discharge de-noising based on hybrid particle swarm optimization SWT adaptive threshold. IEEE Access, 2023; early access. [Google Scholar] [CrossRef]
  32. Hagiwara, K. Bridging between Soft and Hard Thresholding by Scaling. IEICE Trans. Inf. Syst. 2022, 105, 1529–1536. [Google Scholar] [CrossRef]
  33. Grobbelaar, M.; Phadikar, S.; Ghaderpour, E.; Struck, A.F.; Sinha, N.; Ghosh, R.; Ahmed, Z.I. A Survey on Denoising Techniques of Electroencephalogram Signals Using Wavelet Transform. Signals 2022, 3, 577–586. [Google Scholar] [CrossRef]
  34. Soroush, M.Z.; Tahvilian, P.; Nasirpour, M.H.; Maghooli, K.; Sadeghniiat-Haghighi, K.; Harandi, S.V.; Abdollahi, Z.; Ghazizadeh, A.; Dabanloo, N.J. EEG artifact removal using sub-space decomposition, nonlinear dynamics, stationary wavelet transform and machine learning algorithms. Front. Physiol. 2022, 13, 910368. [Google Scholar] [CrossRef] [PubMed]
  35. Wang, Z.; Chen, L.; Xiao, P.; Xu, L.; Li, Z. Enhancing time-frequency resolution via deep-learning framework. IET Signal Process. 2023, 17, e12210. [Google Scholar] [CrossRef]
  36. Neo, W.S.; Foti, D.; Keehn, B.; Kelleher, B. Resting-state EEG power differences in autism spectrum disorder: A systematic review and meta-analysis. Nature 2023, 13, 389. [Google Scholar] [CrossRef] [PubMed]
  37. Proteau-Lemieux, M.; Knoth, I.S.; Davoudi, S.; Martin, C.-O.; Bélanger, A.-M.; Fontaine, V.; Côté, V.; Agbogba, K.; Vachon, K.; Whitlock, K.; et al. Specific EEG resting state biomarkers in FXS and ASD. J. Neurodev. Disord. 2024, 16, 53. [Google Scholar] [CrossRef]
  38. Li, C.; Deng, H.; Yin, S.; Wang, C.; Zhu, Y. sEMG signal filtering study using synchrosqueezing wavelet transform with differential evolution optimized threshold. Results Eng. 2023, 18, 101150. [Google Scholar] [CrossRef]
  39. Aljalal, M.; Aldosari, S.A.; Molinas, M.; AlSharabi, K.; Alturki, F.A. Detection of Parkinson’s disease from EEG signals using discrete wavelet transform, different entropy measures, and machine learning techniques. Sci. Rep. 2022, 12, 22547. [Google Scholar] [CrossRef]
  40. Shah, S.Y.; Larijani, H.; Gibson, R.M.; Liarokapis, D. Epileptic Seizure Classification Based on Random Neural Networks Using Discrete Wavelet Transform for Electroencephalogram Signal Decomposition. Appl. Sci. 2024, 14, 599. [Google Scholar] [CrossRef]
  41. Brihadiswaran, G.; Haputhanthri, D.; Gunathilaka, S.; Meedeniya, D.; Jayarathna, S. EEG-based processing and classification methodologies for autism spectrum disorder: A review. J. Comput. Sci. 2019, 15, 1161–1183. [Google Scholar] [CrossRef]
  42. Alturki, F.A.; Alsharabi, K.; Abdurraqeeb, A.M.; Aljalal, M. EEG signal analysis for diagnosing neurological disorders using discrete wavelet transform and intelligent techniques. Sensors 2020, 20, 2505. [Google Scholar] [CrossRef]
  43. Chang, C.-C. Fisher’s Linear Discriminant Analysis With Space-Folding Operations. IEEE Trans. Pattern Anal. Mach. Intell. 2023, 45, 9233–9240. [Google Scholar] [CrossRef] [PubMed]
  44. Rosenblatt, M.; Tejavibulya, L.; Jiang, R.; Noble, S.; Scheinost, D. Data leakage inflates prediction performance in connectome-based machine learning models. Nat. Commun. 2024, 15, 1829. [Google Scholar] [CrossRef]
  45. Bradshaw, T.J.; Huemann, Z.; Hu, J.; Rahmim, A. A Guide to Cross-Validation for Artificial Intelligence in Medical Imaging. Radiol. Artif. Intell. 2023, 5, e220232. [Google Scholar] [CrossRef] [PubMed]
  46. Yazid, M.; Fahmi, F.; Sutanto, E.; Shalannanda, W.; Shoalihin, R.; Horng, G.-J. Simple Detection of Epilepsy from EEG Signal Using Local Binary Pattern Transition Histogram. IEEE Access 2021, 9, 150252–150267. [Google Scholar] [CrossRef]
  47. Siregar, B.; Florence, G.; Seniman; Fahmi; Mubarakah, N. Classification of Human Concentration Levels Based on Electroencephalography Signals. Int. J. Inform. Vis. 2024, 8, 923–930. [Google Scholar] [CrossRef]
  48. Sutanto, E.; Purwanto, T.W.; Fahmi, F.; Yazid, M.; Shalannanda, W.; Aziz, M. Implementation of Closing Eyes Detection with Ear Sensor of Muse EEG Headband using Support Vector Machine Learning. Int. J. Intell. Eng. Syst. 2023, 16, 460–473. [Google Scholar] [CrossRef]
  49. Silitonga, G.M.; Melinda, M.; Purnamasari, P.D.; Sinulingga, E.; Nasution, A.R.; Fahmi, F. Wearable Microstrip Patch Antenna to Develop Wireless Body Area Network for Electroencephalogram (EEG) Application. In Proceedings of the ELTICOM 2024: 8th International Conference on Electrical, Telecommunication and Computer Engineering: Tech-Driven Innovations for Global Organizational Resilience, Medan, Indonesia, 21–22 November 2024; pp. 193–199. [Google Scholar] [CrossRef]
Figure 1. The proposed framework, different colors illustrates normal and ASD classification.
Figure 1. The proposed framework, different colors illustrates normal and ASD classification.
Diagnostics 15 02291 g001
Figure 2. SWT processing method.
Figure 2. SWT processing method.
Diagnostics 15 02291 g002
Figure 3. FLDA system performance settings.
Figure 3. FLDA system performance settings.
Diagnostics 15 02291 g003
Figure 4. Results of FLDA classifications: (a) normal class and (b) autism class.
Figure 4. Results of FLDA classifications: (a) normal class and (b) autism class.
Diagnostics 15 02291 g004
Table 1. Extraction and denoising results of EEG features at level 3.
Table 1. Extraction and denoising results of EEG features at level 3.
DataFP1 (Hz)F3 (Hz)F7 (Hz)T3 (Hz)OZ (Hz)Class
120.7051521.7110420.391436.959006−10.174Normal
222.1963724.0943822.961412.63111−15.6247Normal
324.5350125.3933124.7127116.85653−17.9162Normal
8.00025.9493826.0588124.216812.1017−15.0655Normal
80016.332813−22.1256−12.83417.16984519.57307ASD
80020.626899−21.0605−12.16563.10819512.88358ASD
8003−0.77898−17.5857−17.1613.07586715.92139ASD
16.00014.84071−6.8033−7.828976.83399416.34017ASD
Table 2. Extraction and denoising results of EEG features at level 4.
Table 2. Extraction and denoising results of EEG features at level 4.
DataFP1 (Hz)F3 (Hz)F7 (Hz)T3 (Hz)OZ (Hz)Class
120.1743921.0824719.676967.651534−10.0963Normal
221.3883123.3411621.8794813.52822−15.165Normal
323.4592324.529123.3433517.9128−17.1389Normal
800024.6060325.1229322.678313.30784−13.937Normal
80016.586297−22.0628−12.56538.1775919.45576ASD
80021.226708−21.3885−11.54334.32700513.16418ASD
80030.124645−18.2795−16.25144.39052716.547ASD
16.00016.14102−8.07165−6.581697.99275117.45375ASD
Table 3. Extraction and denoising results of EEG features at level 6.
Table 3. Extraction and denoising results of EEG features at level 6.
DataFP1 (Hz)F3 (Hz)F7 (Hz)T3 (Hz)OZ (Hz)Class
120.2908420.732119.226917.560314−9.21138Normal
221.5413822.6531821.2318113.1909−13.8999Normal
323.5884523.7983122.5349817.41969−15.6655Normal
8.00024.0966424.4845521.5525412.73626−12.8648Normal
80016.57135−22.081−12.57817.77139619.67485ASD
80021.582899−21.7472−11.03734.02314113.54433ASD
80030.970091−18.8331−14.64274.92918217.00702ASD
16.00017.42998−8.73002−5.080229.285977−18.1207ASD
Table 4. Confusion matrices (counts) for ASD vs. normal by SWT level (autism = positive (+); n = 4800 per level).
Table 4. Confusion matrices (counts) for ASD vs. normal by SWT level (autism = positive (+); n = 4800 per level).
Prediction
Level 3Level 4Level 6
ASDNormalASDNormalASDNormal
ActualASD220816022801202040350
Normal224220812022803702040
Table 5. Binary classification performance by SWT level under FLDA (autism = positive).
Table 5. Binary classification performance by SWT level under FLDA (autism = positive).
SWT LevelMetrics
AccuracySpecificityRecallPrecisionF1 Score
Level 30.9200.9080.9320.9080.920
Level 40.9500.9500.9500.9500.950
Level 60.8500.8460.8540.8460.850
Table 6. Overview of prior ASD-EEG classification approaches.
Table 6. Overview of prior ASD-EEG classification approaches.
Ref.MethodDatasetResult
[9]Butterworth → ICA → KNNKAU EEG signal (ASD)Accuracy 85.4%
[12]CWT features → SVMKAU EEG signal (ASD)Accuracy 95%
[22]Spectrogram/STFT features → classical MLKAU EEG signal (ASD)-related EEG taskAccuracy 95.25%
This workSWT (Levels 3/4/6) → FLDAKAU EEG signal (ASD) → β-bandAccuracy 95%
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Fahmi, F.; Melinda, M.; Purnamasari, P.D.; Elizar, E.; Rafiki, A. Recognition of EEG Features in Autism Disorder Using SWT and Fisher Linear Discriminant Analysis. Diagnostics 2025, 15, 2291. https://doi.org/10.3390/diagnostics15182291

AMA Style

Fahmi F, Melinda M, Purnamasari PD, Elizar E, Rafiki A. Recognition of EEG Features in Autism Disorder Using SWT and Fisher Linear Discriminant Analysis. Diagnostics. 2025; 15(18):2291. https://doi.org/10.3390/diagnostics15182291

Chicago/Turabian Style

Fahmi, Fahmi, Melinda Melinda, Prima Dewi Purnamasari, Elizar Elizar, and Aufa Rafiki. 2025. "Recognition of EEG Features in Autism Disorder Using SWT and Fisher Linear Discriminant Analysis" Diagnostics 15, no. 18: 2291. https://doi.org/10.3390/diagnostics15182291

APA Style

Fahmi, F., Melinda, M., Purnamasari, P. D., Elizar, E., & Rafiki, A. (2025). Recognition of EEG Features in Autism Disorder Using SWT and Fisher Linear Discriminant Analysis. Diagnostics, 15(18), 2291. https://doi.org/10.3390/diagnostics15182291

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop