Next Article in Journal
Preoperative Tunnel Measurement in Hidradenitis Suppurativa: Comparison of Palpation and Ultrasound
Previous Article in Journal
Sarcopenia as a Risk Factor in Patients Undergoing Transjugular Intrahepatic Portosystemic Shunt (TIPS) Implantation
Previous Article in Special Issue
Contribution of the EEG in the Diagnostic Workup of Patients with Transient Neurological Deficit and Acute Confusional State at the Emergency Department: The EMINENCE Study
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Connectogram-COH: A Coherence-Based Time-Graph Representation for EEG-Based Alzheimer’s Disease Detection

Department of Computer Engineering, Karabuk University, Karabuk 78050, Turkey
*
Author to whom correspondence should be addressed.
Diagnostics 2025, 15(11), 1441; https://doi.org/10.3390/diagnostics15111441
Submission received: 10 April 2025 / Revised: 22 May 2025 / Accepted: 23 May 2025 / Published: 5 June 2025
(This article belongs to the Special Issue EEG Analysis in Diagnostics)

Abstract

Background: Alzheimer’s disease (AD) is a neurological disorder that affects the brain in the elderly, resulting in memory loss, mental deterioration, and loss of the ability to think and act, while being a cause of death, with its rates increasing dramatically. A popular method to detect AD is electroencephalography (EEG) signal analysis thanks to its ability to reflect neural activity, which helps to identify abnormalities associated with the disorder. Originating from its multivariate nature, EEG signals are generally handled as multidimensional time series, and the related methodology is employed. Methods: This study proposes a new transformation strategy that generates a graph representation with time resolution, which handles EEG recordings as relatively small time windows and converts these segments into a similarity graph based on signal coherence between available channels. The retrieved adjacency matrices are further flattened to form a 1-pixel image column, which represents the coherence activity from the available electrodes within the given time window. These pixel columns are concatenated horizontally for all available sliding time windows with 50% overlap, resulting in a grayscale image representation that can be input to well-known deep learning architectures specialized for images. We name this representation Connectogram-COH, a coherence-based version of the previously proposed time graph representation, Connectogram. Results: The experimental results demonstrate that the proposed Connectogram-COH representation effectively captures the coherence dynamics of multichannel EEG data and achieves high accuracy in detecting Alzheimer’s disease. The time graph images serve as robust input for deep learning classifiers, outperforming traditional EEG representations in terms of classification performance. Conclusions: Connectogram-COH offers a powerful and interpretable approach for transforming EEG signals into image representations that are well suited for deep learning. The method not only improves the detection of AD but also shows promise for broader applications in EEG-based and general time series classification tasks.

1. Introduction

Alzheimer’s disease (AD) is a chronic neurodegenerative disorder that disrupts memory, cognition, and behavior and is the leading cause of dementia in older adults. It presents as progressive loss of neural function, eventually impairing a person’s ability to perform daily tasks [1,2]. AD has become one of the leading causes of death globally [3], and the socioeconomic burden of the disease continues to rise. The World Alzheimer Report projects the number of people living with AD to reach 75 million by 2030 and a staggering 131 million by 2050 [4], underscoring the urgency of developing scalable, reliable, and non-invasive diagnostic tools.
Currently, diagnosis often relies on a combination of clinical evaluations, neuropsychological tests such as the Mini-Mental State Examination (MMSE) and Montreal Cognitive Assessment (MoCA), and structural or molecular imaging [5]. However, these assessments are typically initiated after symptom onset and are limited by subjective judgment, variability in symptom presentation, and dependence on human expertise. The invasive nature and high cost of imaging methods like PET and MRI further limit their accessibility, particularly for population-wide screening in low-resource settings
Given these constraints, researchers are shifting their attention to non-invasive, inexpensive, and scalable alternatives, notably electroencephalography (EEG). EEG is uniquely suited for early diagnosis due to its superior temporal resolution, ease of use, and sensitivity to subtle neural dynamics [6,7,8]. Importantly, EEG allows for repeated measurements over time, making it ideal for tracking disease progression and therapeutic response. Conventional EEG analysis often involves decomposing signals into canonical frequency bands—delta (0.5–4 Hz), theta (4–8 Hz), alpha (8–13 Hz), beta (13–25 Hz), and gamma (25–45 Hz)—each associated with distinct cognitive and physiological functions [9,10]. Abnormalities in these bands, particularly alpha and theta rhythms, are widely reported in AD patients. Transform methods like fast Fourier transform (FFT) [11], Welch’s power spectral density (PSD) [12], discrete wavelet transform (DWT) [13], and empirical mode decomposition (EMD) [14] have been extensively used to analyze these frequency components and extract informative features for classification.
More recently, a paradigm shift toward functional connectivity (FC) and network science has emerged [15,16]. These approaches focus not only on local activity but also on inter-regional communication within the brain. Metrics such as phase-locking value (PLV), coherence, and amplitude envelope correlation are used to quantify synchrony between EEG channels [17,18]. The resulting connectivity matrices represent the brain as a complex graph, where nodes are electrode sites (or brain regions) and edges represent statistical dependencies or coupling strength. Notably, studies show that AD patients often exhibit disrupted global efficiency and reduced clustering in their functional brain networks [19,20], reflecting a breakdown in integrative cognitive processes.
This graph-based conceptualization has promoted the application of graph theory and network neuroscience in AD research. Measures such as global efficiency, betweenness centrality, and clustering coefficients have shown significant differences between healthy and AD-affected individuals [17,21,22]. For instance, Kabbara et al. demonstrated that alpha band connectivity disruptions in resting-state EEGs were positively correlated with cognitive impairment levels in AD, indicating potential use as a biomarker.
These findings have aligned with advances in deep learning (DL), which allow models to automatically learn patterns from high-dimensional data without manual feature engineering. Early work, such as that by Alves et al. (2021), utilized deep neural networks (DNNs) on EEG-derived connectivity matrices to classify AD and schizophrenia with impressive accuracy, outperforming traditional EEG-based classifiers [23]. Building on this, Demir et al. (2021) introduced graph neural networks (GNNs), which respect the spatial relationships between electrodes and incorporate structural information into learning, allowing models to exploit topological features of the brain network—a significant advancement over CNNs, which assume regular grid-like inputs [24].
Beyond static graph representations, the need to capture temporal evolution in brain connectivity has gained traction. Dynamic functional connectivity—where network properties change over time—can reveal transient synchronization events and early biomarkers missed by static analyses. Studies like that by Gupta et al. (2022) have employed tensor decomposition to model EEG data across time, space, and frequency dimensions, recovering latent temporal patterns that align with disease stages [25]. Similarly, the LEAD model proposed by Wang et al. (2025) introduced a scalable pretraining strategy using contrastive learning across large EEG datasets, achieving state-of-the-art performance in subject-independent AD classification [26].
Another noteworthy development is the move toward interpretable AI. Ajra et al. (2023) emphasized the use of shallow CNNs and interpretable connectivity metrics like amplitude envelope correlation (AEC) to classify EEG signals in AD and frontotemporal dementia (FTD), achieving over 94% accuracy [27]. Their work highlights the diagnostic potential of frequency-specific network features. Similarly, Fruehwirt et al. (2018) employed Bayesian neural networks to provide not only predictions but also uncertainty estimates, which are critical for clinical decision making [28].
Despite these successes, challenges remain, especially regarding signal artifacts, inter-subject variability, and generalizability. Recent approaches attempt to address these via subject-independent validation, data augmentation, or by focusing on subcortical source-localized signals [29]. Some models, like the one proposed by Sunkara et al. (2024), even explore alternative neural architectures (e.g., Kolmogorov–Arnold networks) to enhance classification robustness under noisy conditions [30].
The current study builds upon this body of work by proposing a novel methodology for representing time-evolving EEG graphs as 2D images. This transformation preserves the temporal dynamics of functional connectivity while enabling the application of powerful CNN architectures originally developed for image classification. While previous methods primarily utilized static graphs or flattened features [9,11,12,13,14,17,31,32], our approach allows for the concatenation of dynamic connectivity matrices across sliding time windows. This results in a “graph sequence image” that captures both spatial and temporal fluctuations in brain synchronization.
Such methodology aligns with trends in both multimodal EEG representation and efficient deep learning, enabling reduced model complexity without compromising classification performance. It also provides a framework that is generalizable to other time series domains, suggesting broad applicability beyond AD diagnosis. The integration of neuroscience-informed signal processing and interpretable machine learning underscores the potential of our approach to bridge the gap between research and clinical utility.

2. Data and Methodology

2.1. Dataset

The proposed methodology was evaluated in an openly available EEG dataset from the OpenNeuro data repository [33], comprising recordings from 88 participants. EEGs were recorded using 19 scalp electrodes placed according to the international 10–20 system (Fp1, Fp2, F7, F3, Fz, F4, F8, T3, C3, Cz, C4, T4, T5, P3, Pz, P4, T6, O1, and O2) and two mastoid reference electrodes (A1 and A2). All recordings followed a standardized clinical protocol, with participants seated comfortably and instructed to keep their eyes closed. The data were sampled at 500 Hz with a resolution of 10 μV/mm.
The dataset included three groups: 36 individuals with Alzheimer’s disease (AD; 20 female), 23 individuals with frontotemporal dementia (FTD; 9 female), and 29 cognitively normal (CN) participants (18 female). The average ages were 66.4 years (SD = 7.9) for the AD group, 63.6 years (SD = 8.2) for the FTD group, and 67.9 years (SD = 5.4) for the CN group. Cognitive functioning was assessed using the Mini-Mental State Examination (MMSE), with mean scores of 17.75 (SD = 4.5) for AD, 22.17 (SD = 8.22) for FTD, and 30 for CN, indicating moderate impairment in the dementia groups and normal cognition in controls. The AD group had a median disease duration of 25 months (IQR: 24–28.5).
Each EEG recording lasted approximately 13.5 min for AD participants (range: 5.1–21.3 min), 12 min for the FTD group (range: 7.9–16.9 min), and 13.8 min for the CN group (range: 12.5–16.5 min) [33,34]. These demographic and clinical details help to contextualize the stage of disease progression and support the interpretation of EEG patterns identified through machine learning analysis.

2.2. Data Processing for Time Graph Conversion

To enable classification using deep learning, the raw EEG recordings were first transformed into time graph image representations, referred to as Connectogram-COH. This transformation involved segmenting multichannel EEG signals into fixed-length, non-overlapping segments and subsequently converting each into a series of functional brain connectivity graphs.
First of all, the data consisting of multichannel EEG recordings from 19 electrodes were subdivided into fixed length segments (10, 20, or 30 s) without any overlap to generate a pool of independent EEG signals available for the classification task. This process provided the augmentation of the EEG dataset from 88 recordings into 6938, 3446, or 2285 segments, respectively, where each segment was labeled the same as its original full-length recording and had the same multivariate nature. These independent segments were the raw sources to be converted into time graph (grayscale image) representations.
We strictly ensured that all data segments derived from a given subject were assigned exclusively to either the training or the testing set within each fold. To enforce this, we adopted a subject-wise k-fold cross-validation strategy, where the data were split at the subject level, not at the segment level. The segmentations were performed after subject-level splitting, ensuring complete isolation between subjects in the training and testing sets. This meant that no segments from the same subject appeared in both the training and testing sets in any cross-validation fold. Applying zero overlap between segments also secured independence between segments and avoided data leakage. All of these steps collectively ensured the validity of our evaluation protocol and the reliability of the reported classification performance.
Next, each fixed-length segment was handled as a separate and independent data recording. To capture the fine-grained temporal dynamics of brain activity, a sliding window approach was applied. A window length of 0.4 s and overlap of 0.2 s (50% overlap) were employed to subdivide the data into time windows. This resulted in a sequence of short-duration multichannel EEG signals within each segment, ready for graph conversion.
Then, each 0.4 s window was transformed into a functional connectivity graph by computing coherence values between every pair of EEG channels. Coherence is a frequency domain measure that quantifies the degree of synchronization between two signals, providing insight into functional interactions between brain regions. The coherence metric represents the level of mutual information or synchronization of neural activity across the nodes, while the nodes correspond to physical brain areas. The mathematical background of the coherence metric between equal-length signals x(t) and y(t) can be defined using the cross-spectral density function Pxy and the power spectral density functions Pxx and Pyy, as given in Equation (1) [18].
C o h = P x y 2 P x x P y y
This process produces a symmetric 19 × 19 coherence matrix called the adjacency matrix for each window, representing the functional network of brain regions. Figure 1 presents a 19 × 19 coherence matrix, which visualizes the functional connectivity for a 0.4 s EEG window.
In the adjacency matrix, each element represents the coherence between a pair of electrodes defining a connection weight between EEG channels (signals from independent electrodes), as visualized in Figure 1. In this functional brain graph, the nodes represent distinct brain regions or areas, often defined by anatomical or functional parcellation (e.g., regions from the cerebral cortex, subcortical structures), while the edges represent functional connections between these nodes, reflecting statistical relationships or interactions. Such graphs offer insight into the dynamic interactions among brain areas during different cognitive or physiological states.
The achieved adjacency matrix has a symmetric nature caused by the equal coherence values achieved in two directions for a signal pair. Therefore, the upper triangle of this matrix forms a meaningful representation for the corresponding time window. This upper triangle is later flattened to achieve a vector of size 171 × 1, which represents the connectivity pattern derived from a single time window.
For each segment (10, 20, or 30 s), these flattened representation vectors are tiled horizontally, which results in 149 pixels of horizontal resolution. Therefore, the resulting time graph representation shown in Figure 2 has a resolution of 171 × 149 pixels for a 30 s segment. We named this representation Connectogram-COH, as a coherence-based variety of the recently proposed Connectogram for time series [35].
These image-like representations are finally passed to suitable deep learning architectures such as CNN or transfer learning models, leading to successful classification results, as presented in the coming sections. The entire methodology is illustrated in Figure 3.

2.3. Experimental Setup

The experimental setup consisted of trials for various lengths of segmentation of the original recordings as 10, 20, or 30 s. For each segment, we provided the same class label as the original recording, and we generated a separate Connectogram-COH image for each of these data segments. Having applied this set of various segmentation lengths, we achieved 6938, 3446, and 2285 Connectogram-COH images and their corresponding class values. The segment length did not affect the dimension of the adjacency matrix, since it was determined by the number of EEG channels, which was 19 for the dataset used. The vertical dimension of the resulting Connectogram-COH remained the same, accordingly.
However, the horizontal dimension scaled linearly with the size of the segment length, which determined the number of available sliding fixed-length windows. As a result, Connectogram-COH had the resolution of 171 × 149, 171 × 99, and 171 × 49 pixels for the 30, 20, and 10 s segments, respectively, as illustrated in Figure 4.
We tested the classification performance of the Connectogram-COH images across a variety of deep learning architectures, such as CNN, ResNet, VVG16, InceptionV3, EfficientNetB7, and DenseNet121. We also performed experiments for a range of batch sizes (4 to 512) and epochs (20 to 100) for the best performer learning model, as explained in the next section. The learning models used are briefly described below.

2.3.1. Convolutional Neural Network (CNN)

A convolutional neural network (CNN) is a deep learning architecture specifically designed to process structured data, particularly images and visual inputs. It emulates the human brain’s visual processing through convolutional layers that automatically identify features such as edges, textures, and patterns. These layers utilize filters (kernels) that move across the input to generate feature maps, emphasizing key characteristics. Pooling layers allows to downsample the spatial dimensions, enhancing computational efficiency and resilience to minor input variations. Finally, fully connected layers integrate the extracted features to perform predictions. CNNs are extensively applied in image classification, object detection, and segmentation tasks due to their capability to learn hierarchical data representations. The custom CNN architecture employed in this study is detailed in Table 1.

2.3.2. Residual Network (ResNet)

ResNet is a deep learning model designed to overcome the issue of vanishing gradients in very deep networks. It incorporates the concept of ‘residual learning’ by utilizing skip connections, which allow the model to bypass intermediate layers and transmit the input directly to deeper layers. This technique enables the network to learn the residual, or the difference, between the input and the output, rather than attempting to directly learn the output. As a result, ResNet prevents performance degradation in deeper networks, allowing the training of extremely deep models with hundreds or even thousands of layers. ResNet has proven highly effective in image recognition tasks and forms the basis for numerous advanced architectures [36]. There exists a variety of ResNet models, such as ResNet-50, ResNet-18, etc., but our model does not match any of these exact versions. It shares elements with smaller ResNet models (like ResNet-18/34) in terms of basic residual blocks, which makes it a “shallow ResNet” model, as described in Table 2.

2.3.3. VGG-16

VGG-16 is a convolutional neural network (CNN) architecture that is mostly preferred for its simplicity and effectiveness in image recognition tasks. It comprises 16 layers with trainable parameters, including 13 convolutional layers and 3 fully connected layers. The network employs small 3 × 3 filters in all convolutional layers, stacked to deepen the network and capture intricate features. Pooling layers are interspersed to reduce spatial dimensions, while the fully connected layers at the end are responsible for classification. Despite having a relatively large number of parameters, the straightforward design of VGG-16 has established it as a benchmark model in computer vision, widely applied in transfer learning and feature extraction [37].

2.3.4. Inception v3

Inception v3 is an advanced deep convolutional neural network designed for efficient and accurate image classification. It integrates innovative architectural features from the Inception family, optimizing performance while reducing computational cost. It leverages Inception modules, which stack filters of different sizes (e.g., 1 × 1, 3 × 3, and 5 × 5) to capture multi-scale features, all while maintaining low computational costs. The model introduces techniques like factorized convolutions (splitting larger convolutions into smaller ones, e.g., 3 × 3 into two 1 × 3 and 3 × 1), label smoothing to prevent overfitting, and an auxiliary classifier for gradient flow improvement. With 42 layers and approximately 23 million parameters, Inception v3 provides a balance between depth and efficiency, delivering state-of-the-art performance on tasks such as ImageNet classification [37].

2.3.5. EfficientNet-B7

EfficientNet-B7, the largest model in the EfficientNet family, optimizes performance and computational efficiency through a compound scaling approach. This method systematically scales the network’s depth, width, and resolution, allowing the model to achieve state-of-the-art accuracy while requiring fewer parameters than traditional architectures. It incorporates advanced techniques such as squeeze-and-excitation blocks to improve feature extraction, swish activation functions for smoother gradients, and depthwise separable convolutions to reduce computational costs. With approximately 66 million parameters, EfficientNet-B7 achieves remarkable accuracy in image classification tasks, such as ImageNet, while maintaining a balance between performance and computational cost [37].

2.3.6. DenseNet-121

DenseNet-121 is a convolutional neural network that introduces dense connectivity between layers to improve feature propagation and reduce redundancy. Unlike traditional architectures, where each layer connects only to the next, DenseNet establishes direct connections between every layer in a feed-forward manner. This dense connectivity enhances gradient flow, optimizes parameter usage, and mitigates the vanishing gradient problem. DenseNet-121 features 121 layers, organized into dense blocks and transition layers, with each layer receiving inputs from all preceding layers. This structure reduces the number of parameters while maintaining high accuracy. DenseNet-121 is particularly efficient for image classification tasks, achieving improved performance on benchmarks like ImageNet with fewer parameters compared to other models [37].

3. Results and Discussion

We first tested the classification performance of the generated time graph representations across the mentioned deep learning models to determine the best performing model. These experiments were conducted on the available three classes: control (CN), Alzheimer’s disease (AD), and frontotemporal dementia (FTD). For each experimental setup, a set of 5 experiments was conducted to achieve reliable results, which are presented in Table 3.
However, some of the experiments could not be executed since the aspect ratio of the images did not match the requirements of the learning models, being far from a square-like output. The experiments implied that the ResNet model outperformed the rest of the models, with 99.49% mean accuracy, while the custom CNN model scored the second-best accuracy (98.63%) after ResNet. Other models failed to compete with these models. The second major outcome of these experiments was that better accuracy results were achieved for lower-length data segments, which resulted in smaller connectograms on the horizontal scale.
For the two-class case, ResNet achieved 99.53% accuracy for AD vs. CN, 99.50% for FTD vs. CN, and 99.45% for AD vs. FTD. Finally, the 5-fold mean accuracy for ResNet in the three-class scenario was 99.33%, with a standard deviation of 0.25. The learning curves for CNN and ResNet shown in Figure 5 demonstrate good generalization capability, with close alignment between training and validation accuracy, indicating no evidence of overfitting.
Confusion matrices for the first two models, CNN and ResNet, are also presented in Figure 6. The two models showed similar patterns of classification accuracy while their differences originated from mislabeled instances by CNN in the CN and AD classes. Very few misclassified instances in ResNet side demonstrated that the model had a very strong and robust performance for Connectogram-COH images.
The classification reports for the models tested are shown in Table 4, while Table 5 and Figure 7 present the ROC curves and the areas under the ROC curves for each algorithm employed.
Having determined the best performing model as ResNet, we fixed the architecture as ResNet and tried to fine-tune the framework. We repeated each experimental setup by subdividing the original data into different equal-length segments as 10, 20, and 30 s without any overlap, in compliance with the strategy commonly used for detecting brain disorders using EEG signals [38,39]. This segmentation strategy, beyond providing equal-size data inputs to the learning models, has the advantage of extending the available number of samples for a given dataset, which in turn meets the demand for high data volume by a deep learning model. The dataset employed in this study originally included 485.5 min of AD, 276.5 min of FTD, and 402 min of CN recordings which are extended into a total number of 6938, 3446 and 2285 equal-size 410 bins for 10, 20 and 30-s segment lengths. As a result, a sufficient number of independent samples was achieved to feed the deep neural network model. The zero-overlap segmentation procedure also enhanced the integrity and reliability of both percentage split and k-fold validation due by securing non-overlapping training and test sets.
Table 6 presents the average accuracy achieved for a variety of parametric values, ranging from 4 to 512 for batch size and 20 to 100 for epoch size, after 5 trials for each setup were performed.
The experiments yielded promising results for different batch sizes from one set to another, generally appearing for small-to-moderate batch sizes in the range 4 to 64. This was because choosing a small batch size such as (4, 8, 16, 32) may lead to fuzzy estimates of gradients, as the weights are updated based on a small number of examples each time. But it may be useful in avoiding overfitting and it contributes to speeding up in some cases. It helps to improve the generalizability of the model. On the other hand, choosing a large batch size such as (64, 128, 256, 512) may provide more stable and accurate estimates of gradients because a larger number of examples is used at each step. It can also reduce the “noise” in the gradients and speed up convergence, but it requires more memory and computational power. Sometimes, it may lead to poor generalizability if the size is not set appropriately. The same goes for the number of epochs; using too few epochs means training for too short a time and the model may not learn the data well enough, leading to poor performance (this is called underfitting). Conversely, if we specify too many epochs, the model may start to overfit the training data, meaning it will overlearn small details in the data, leading to poor performance on the test data.
Another important factor to note is the change in the number of images from one segment length to another. This was clearly shown when using a length of 30 s, where the number of images was 2285 and the resolution of each image was 171 × 149 pixels. By contrast, when using a length of 10 s, the number of images was up to 6638 with a relatively low image resolution of 171 × 49 pixels. Therefore, the quality of the test differed slightly. Finally, we can recommend the use of small-to-moderate batch size values and epochs, since the generated time graph representations had the ability to serve as robust input over a wide range of parameters; however, limiting the segment length values to 10 or 20 s would give better accuracy regardless of the other parameters tested. The robustness of the methodology also makes it suitable for detecting a wide range of brain disorders, such as autism, Parkinson’s, epilepsy, etc.
Unlike most studies performed recently that handled binary classification tasks such as AD-FTD, AD-CN, or FTD-CN, our study handles a three-class problem, which critically affected the classification metrics. The results presented, which include the classification of frontotemporal dementia in addition to Alzheimer’s disease, make the approach valuable and promising for several multiclass classification tasks, such as cardiac disorders that range from 4 to 20 or more classes, in case multi-lead recordings are available.
Recent research on neurodegenerative disease detection using electroencephalography (EEG) signals has introduced a variety of innovative approaches tailored for specific classification tasks. Dogan et al. [40] proposed a binary classification model for Alzheimer’s disease (AD) detection, employing a graph-based methodology inspired by the primate brain connectome to extract discriminative features from EEG signals. Miltiadous et al. [10,41] extended this work by addressing the differentiation between AD and frontotemporal dementia (FTD), leveraging robust classification techniques and novel neural architectures, such as DICE-Net, which combines convolutional and transformer-based models. Araujo et al. [42] introduced a smart data-driven system for AD detection, demonstrating the utility of advanced feature extraction in EEG-based studies, while Gomez et al. [43] focused on multiclass classification tasks involving AD, mild cognitive impairment (MCI), and healthy controls, highlighting the potential of spontaneous EEG activity for accurate classification. Safi et al. [11] emphasized the importance of early detection, utilizing Hjorth parameters to effectively distinguish AD from control groups. Khatun et al. [44] developed a single-channel EEG approach based on speech-evoked brain responses for detecting mild cognitive impairment, showing the feasibility of simplified setups. Beyond EEG, Gordon et al. [45] explored clinical and MRI data to measure disease progression in frontotemporal lobar degeneration.
Despite these advancements, most prior studies are limited to binary classification tasks or smaller datasets, with minimal focus on visualizing connectivity patterns over time. Our study bridges these gaps by introducing a time graph representation methodology for a three-class classification task, distinguishing AD, FTD, and control groups. This approach generates a unique output image where the horizontal axis represents time and the vertical axis reflects dynamic connectivity patterns across brain regions, providing an innovative visualization of neural activity. Furthermore, our work leverages a significantly larger EEG dataset than many existing studies, enabling better generalization and robustness. By combining time graph imaging with advanced classification techniques, our study offers a novel perspective on understanding and differentiating neurodegenerative diseases, paving the way for more comprehensive and scalable methodologies in EEG-based diagnostics.
In comparison to prior studies employing three-class classification approaches, including those by Lopes et al. [46], Gomez et al. [43], Ieracitano et al. [47], and Xia et al. [48], which achieved competitive results in differentiating Alzheimer’s disease (AD), mild cognitive impairment (MCI), and healthy controls, our proposed strategy demonstrates superior performance in classification accuracy and robustness. Lopes et al. [46] combined convolutional neural networks (CNNs) with saliency maps and EEG modulation spectra to enhance interpretability and diagnostic accuracy for AD. Gomez et al. [43] utilized spontaneous EEG activity for multiclass classification, emphasizing the efficacy of automated feature extraction methods. Similarly, Ieracitano et al. [47] developed a CNN-based model that employed two-dimensional spectral representations of EEG recordings to effectively classify dementia stages. Xia et al. [48] introduced a deep pyramid CNN architecture tailored for EEG signals, showcasing its potential in AD detection with high precision. While these studies have made significant contributions, they primarily focus on either specific architectures or unique representations of EEG data.
Both the CNN and residual network (ResNet) architectures integrated into our framework offer significant advantages in handling complex EEG patterns, which are converted into grayscale image representations. However, a key limitation lies in the reliance on a fixed number of electrodes, which constrains the vertical dimension of the resulting image. A central design principle of the proposed methodology is the creation of square-like images to enable compatibility with certain deep learning architectures. This limitation necessitates careful balancing of parameters that affect the vertical and horizontal dimensions of the image—particularly the selection of appropriate segmentation lengths and windowing strategies—to ensure comparable dimensions across both axes. However, there lies a key strength of the methodology if the pixel aspect ratio constraint is loosened, introducing an inherent flexibility in adapting to different EEG setups. Various window-length values, resulting in an imbalanced width/height aspect ratio, would also represent the original signal activity with various temporal characteristics embedded into the graph representation. With the aid of concurrent advancements in the AI domain, these output images with imbalanced aspect ratios can also be processed efficiently by deep learning models. For example, when working with 20, 32, or 64 channels (c), the resulting coherence (or adjacency) matrices scale accordingly to 20 × 20, 32 × 32, and 64 × 64. These matrices are then flattened into vectors of sizes 190 × 1, 496 × 1, and 2016 × 1, respectively (can be calculated as c(c − 1)/2 to reflect the size of the upper triangle of the adjacency matrix), effectively capturing the connectivity patterns within each time window regardless of the montage. This same adaptable framework applies to changes in segmentation and windowing parameters. By adjusting these inputs, the method maintains compatibility with various EEG acquisition setups, enabling generalization across datasets with different channel counts and recording configurations. Overall, while the current implementation is optimized for a specific configuration, the underlying principles can be extended through scaling strategies, dynamic preprocessing, and architectural adjustments. This flexibility enhances the broader applicability of our approach in both research and clinical EEG contexts.
Another potential concern is that flattening the upper triangle of the coherence matrix may discard essential topological properties of spatiotemporal graphs. To address this, we employed a sliding window approach, segmenting the data into discrete time windows and computing a coherence matrix for each segment. This strategy yields a time series of matrices rather than a single static representation, thereby enabling the capture of dynamic fluctuations in connectivity patterns over time. To reduce redundancy while retaining key information, we extracted only the upper triangular portion of each coherence matrix. This conservative linear representation preserves the exact pairwise connectivity values and their ordering—features that are often the most informative for classification tasks in clinical applications. Although this approach does not explicitly preserve certain abstract topological properties (e.g., spatial node configurations), it maintains the core connectivity structure necessary for effective downstream analysis.
The reason for not implementing and testing the Connectogram-COH images on graph-based learning models is that these models expect pure graph structures (adjacency matrices) as input, while our model outputs flattened graphs concatenated on the horizontal axis, resulting in time graph representations. These images rather resemble the structure of power spectrogram images that are well known in the literature, which are commonly input to CNN-like architectures for classification tasks.
Importantly, we enforced strict subject-wise cross-validation to eliminate the risk of data leakage, which could otherwise lead to artificially inflated classification performance. Specifically, all data segments from a given subject were assigned exclusively to either the training or the testing set within each fold. We adopted a subject-wise k-fold cross-validation strategy, where data splitting was performed at the subject level, followed by segmentation. This ensured that no data from the same subject appeared in both the training and testing sets in any fold. Additionally, to further mitigate the risk of data leakage, all preprocessing and segmentation procedures were conducted independently within each fold. We also avoided using overlapping segments, which helps to reduce the likelihood of high similarity between samples in the training and testing sets. These precautions ensured the validity and generalizability of our model evaluation.
It is important to note that we recognize that recent literature increasingly explores transformer models, graph neural networks, and multimodal data fusion. Spatiotemporal GNNs (ST-GNNs), in particular, provide a powerful framework for modeling dynamic and spatial interactions by leveraging the complete graph structure over time, often resulting in superior performance in both predictive and interpretive tasks. However, these models typically require significant design effort, including the construction of the graph itself and the manual definition of node features.
By contrast, our study aims to demonstrate that a well-structured convolutional neural network (CNN)-based architecture—specifically ResNet—can achieve competitive, and in some cases superior, performance when combined with biologically meaningful input features, such as coherence-based connectograms. With appropriate preprocessing and rigorous validation, this approach proves highly effective, as evidenced by its performance compared to more complex transformer-based models such as that in [10], despite using the same dataset. Our contribution lies in showing that architectural simplicity, when paired with strong signal representation, can rival or even surpass more elaborate pipelines under certain conditions. Moreover, our method integrates seamlessly with traditional high-performance classifiers such as XGBoost. Notably, despite its streamlined structure, our approach achieved an excellent classification accuracy of 99.49%.
Table 7 presents a list of recent studies that have addressed the same problem together with the details of the datasets and methodology in comparison with the current study to provide insights about the improvements provided. It is important to note that the performance comparisons should be interpreted in the context of varying study designs, dataset sizes, and classification tasks. The given results indicate that our methodology can outperform all of the recent studies handling the same task. Meanwhile, some referenced works use much smaller datasets (fewer subjects, such as in [35,44,49,50]), which can lead to overfitting or inflated performance. The success of the methodology is more apparent among the studies handling relatively large datasets, including more than a hundred subjects, and multiclass classifications. We can conclude that the repeatability of success for all pairs of classes is mostly caused by the high representative capability of the generated time graph representations.

4. Model Deployment

As shown in Table 8, this compact deep learning model features just under 1 million trainable parameters, achieving a balanced trade-off between capability and efficiency. Its lightweight architecture enables deployment across a wide range of platforms—including desktops, mobile devices, and embedded edge hardware—without the need for substantial computational resources.
With a small memory footprint, the model is especially well suited for storage-constrained environments such as mobile applications, embedded systems, and specialized domains like medical, industrial IoT, and mobile computing. The minimal storage requirements also facilitate cloud-based deployment, allowing for fast downloads and seamless updates via APIs.
In terms of performance, the model demonstrates high throughput, making it ideal for batch inference tasks such as offline data processing, bulk analytics, and backend pipelines where real-time responsiveness is not critical. The inference latency supports near real-time processing, which suffices for many interactive applications. However, further optimization may be required for ultra-low-latency scenarios, such as brain–computer interfaces (BCIs) or real-time biomedical signal analysis, where milliseconds can be mission-critical.

5. Limitations and Future Directions

While our methodology has demonstrated effectiveness, several limitations merit attention:
  • Fixed Electrode Count: The current design supports only 19-channel EEG. For high-density arrays, the image size may become unwieldy. Future work could use graph pooling, dimensionality reduction, or adaptive montages to scale effectively.
  • Aspect Ratio Issues: Some CNN models (e.g., InceptionV3, EfficientNetB7) struggle with the elongated shape of connectograms. Padding, resizing, or using models that accept non-square inputs can improve compatibility.
  • Explainability: Explainable AI (XAI) techniques applied to Connectogram-COH images to highlight which regions most influence the model’s predictions would give valuable insights to the study. Additionally, for EEG coherence graphs, graph-specific explainability approaches could offer insight into which connections or brain regions are most critical to the classification. These additions would help clinicians to better understand and trust the model’s decisions, and we view them as a vital direction for future work.
The need for more flexible data representations would encourage future work that would:
  • Implement dynamic window sizing and normalization to better handle varying EEG lengths, sampling rates, and channel counts.
  • Benchmark the method on multiple public datasets, using fine-tuning or domain adaptation as needed.
  • Explore graph-based models (such as GNN) to retain spatial and temporal structures for improved performance.
  • Apply transfer learning to adapt pretrained models to smaller, clinical datasets.
In the long term, we aim to translate this work into real-world clinical applications by partnering with clinicians to conduct prospective studies, compare model predictions with expert diagnoses, and ensure alignment between model outputs and established diagnostic criteria. These steps will help to bridge the gap between research and practice, fostering trust and utility in AI-assisted neurological diagnostics.

6. Conclusions

The current study proposes a novel time graph representation technique that can be applied to any multidimensional time series data. The graph conversion methodology relies on assuming each channel of EEG data as a node, while the signal coherence values between each signal pair represent the weight between corresponding nodes. In contrast to recent approaches converting EEG recordings into a static graph, the proposed transformation technology also captures the time dependency of the functional connectivity graph, resulting in a time graph representation that can be handled as a grayscale image and input to image-related deep learning architectures. The methodology is tested on a task related to the detection of Alzheimer’s disease by employing a variety of deep learning architectures. The promising results indicate the potential of the transformation strategy to be employed in any type of multidimensional time series data by extending the static graph representation approach into a time-dependent variant by enhancing a suitable graph conversion strategy that is coherent to the signal’s nature. The proposed framework was applied to a dataset available in the OpenNeuro data repository [19], labelled as Alsheimer’s disease (AD), frontotemporal dementia (FTD), or control (CN) groups, and overperformed current state-of-art methodologies in the field at a significant level. As a result, this study upgrades the capability of accurate classification of multidimensional time series data, specifically focusing on the diagnosis of Alzheimer’s disease but has similar potential for detecting several disorders, such as autism, Parkinson’s, epilepsy, etc. We also believe that introducing the time dimension to brain connectivity graphs not only captures the connectivity patterns of signal activity between brain regions but also represents the time-dependent temporal activity, providing a trustworthy basis for future research and clinical diagnosis.

Author Contributions

Conceptualization, İ.T.; data curation, E.A.; formal analysis, E.A. and İ.T.; investigation, E.A. and İ.T.; methodology, İ.T.; writing—original draft preparation, E.A.; writing—review and editing, İ.T.; supervision, İ.T. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Data collection was conducted according to the guidelines of the Declaration of Helsinki and approved by the Ethics Committee of University of Ioannina, as declared by the owners of the dataset in ref. [41].

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study, declared by the owners of the dataset in ref. [41].

Data Availability Statement

The original data presented in the study are openly available in OpenNeuro.org at https://doi.org/10.18112/openneuro.ds004504.v1.0.1.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Puri, D.V.; Nalbalwar, S.L.; Nandgaonkar, A.B.; Gawande, J.P.; Wagh, A. Automatic detection of Alzheimer’s disease from EEG signals using low-complexity orthogonal wavelet filter banks. Biomed. Signal Process. Control 2023, 81, 104439. [Google Scholar] [CrossRef]
  2. Fernández, M.; Gobartt, A.L.; Balañá, M. Behavioural symptoms in patients with Alzheimer’s disease and their association with cognitive impairment. BMC Neurol. 2010, 10, 87. [Google Scholar] [CrossRef] [PubMed]
  3. Atri, A. The Alzheimer’s disease clinical spectrum. Med. Clin. N. Am. 2019, 103, 263–293. [Google Scholar] [CrossRef] [PubMed]
  4. Ding, Y.; Chu, Y.; Liu, M.; Ling, Z.; Wang, S.; Li, X.; Li, Y. Fully automated discrimination of Alzheimer’s disease using resting-state electroencephalography signals. Quant. Imaging Med. Surg. 2022, 12, 1063–1078. [Google Scholar] [CrossRef]
  5. Dubois, B. Research criteria for the diagnosis of Alzheimer’s disease: Revising the NINCDS-ADRDA criteria. Lancet Neurol. 2007, 6, 734–746. [Google Scholar] [CrossRef]
  6. Miltiadous, A.; Tzimourta, K.D.; Giannakeas, N.; Tsipouras, M.G.; Glavas, E.; Kalafatakis, K.; Tzallas, A.T. Machine learning algorithms for epilepsy detection based on published EEG databases: A systematic review. IEEE Access 2023, 11, 564–594. [Google Scholar] [CrossRef]
  7. Geraedts, V.J.; Boon, L.I.; Marinus, J.; Gouw, A.A.; van Hilten, J.J.; Stam, C.J.; Tannemaat, M.R.; Contarino, M.F. Clinical correlates of quantitative EEG in Parkinson disease. Neurology 2018, 91, 871–883. [Google Scholar] [CrossRef]
  8. Alhassan, S.; Soudani, A. Energy-aware EEG-based Scheme for early-age Autism detection. In Proceedings of the 2022 2nd International Conference of Smart Systems and Emerging Technologies (SMARTTECH), Riyadh, Saudi Arabia, 9–11 May 2022; IEEE: New York, NY, USA, 2022; pp. 97–102. [Google Scholar]
  9. Kulkarni, N.N.; Bairagi, V.K. Extracting salient features for EEG-based diagnosis of Alzheimer’s disease using support vector machine classifier. IETE J. Res. 2017, 63, 11–22. [Google Scholar] [CrossRef]
  10. Miltiadous, A.; Gionanidis, E.; Tzimourta, K.D.; Giannakeas, N.; Tzallas, A.T. DICE-Net: A Novel Convolution-Transformer Architecture for Alzheimer Detection in EEG Signals. IEEE Access 2023, 11, 71840–71858. [Google Scholar] [CrossRef]
  11. Safi, M.S.; Safi, S.M.M. Early detection of Alzheimer’s disease from EEG signals using Hjorth parameters. Biomed. Signal Process. Control 2021, 65, 102338. [Google Scholar] [CrossRef]
  12. Şeker, M.; Özbek, Y.; Yener, G.; Özerdem, M.S. Complexity of EEG dynamics for early diagnosis of Alzheimer’s disease using permutation entropy neuromarker. Comput. Methods Programs Biomed. 2021, 206, 106116. [Google Scholar] [CrossRef] [PubMed]
  13. Tzimourta, K.D.; Giannakeas, N.; Tzallas, A.T.; Astrakas, L.G.; Afrantou, T.; Ioannidis, P.; Grigoriadis, N.; Angelidis, P.; Tsalikakis, D.G.; Tsipouras, M.G. EEG window length evaluation for the detection of Alzheimer’s disease over different brain regions. Brain Sci. 2019, 9, 81. [Google Scholar] [CrossRef] [PubMed]
  14. Roy, Y.; Banville, H.; Albuquerque, I.; Gramfort, A.; Falk, T.H.; Faubert, J. Deep learning-based electroencephalography analysis: A systematic review. J. Neural Eng. 2019, 16, 051001. [Google Scholar] [CrossRef] [PubMed]
  15. Olgun, N.; Özkaynak, E. Complex Network Analysis of EEG Signals of Epilepsy Patients. In Proceedings of the 32nd Signal Processing and Communications Applications Conference (SIU), Mersin, Turkey, 15–18 May 2024; pp. 1–4. [Google Scholar] [CrossRef]
  16. Olgun, N.; Özkaynak, E. A novel approach to detecting epileptic patients: Complex network-based EEG classification. J. Complex Netw. 2024, 12, cnae044. [Google Scholar] [CrossRef]
  17. Zhang, D.; Yin, J.; Zhu, X.; Zhang, C. Network representation learning: A survey. IEEE Trans. Big Data 2020, 6, 3–28. [Google Scholar] [CrossRef]
  18. Bastos, A.M.; Schoffelen, J.M. Investigation of functional connectivity using MEG: Assessing the influence of source modeling approaches. NeuroImage 2016, 186, 370–380. [Google Scholar] [CrossRef]
  19. Kabbara, A.; El Falou, R.; Khalil, C.; Hassan, M. Scalp-EEG Network-Based Analysis of AD. arXiv 2017, arXiv:1706.03839. [Google Scholar]
  20. Wang, Y.; Zhang, J.; Chen, L.; Wang, S. LEAD: A Large Foundation Model for EEG-Based Alzheimer’s Disease Detection. arXiv 2025, arXiv:2502.01678. [Google Scholar]
  21. Stam, C.J. Modern network science of cognitive dynamics: A review. Neurosci. Biobehav. Rev. 2014, 48, 32–42. [Google Scholar]
  22. Gollo, L.L.; Breakspear, M. The human connectome: A structural description of the human brain network. Brain Cogn. 2014, 88, 95–109. [Google Scholar]
  23. Alves, C.L.; Vigário, R.; Rodrigues, P.M. EEG Functional Connectivity and Deep Learning for Automatic Diagnosis. arXiv 2021, arXiv:2110.06140. [Google Scholar]
  24. Demir, A.; Ertam, F.; Cetin, A.E. EEG-GNN: Graph Neural Networks for EEG Classification. arXiv 2021, arXiv:2106.09135. [Google Scholar]
  25. Gupta, T.; Zhang, W.; Wu, D. Tensor Decomposition of Large-Scale EEGs. arXiv 2022, arXiv:2211.13793. [Google Scholar]
  26. Wang, Y.; Zhang, J.; Chen, L.; Wang, S. Flexible and Explainable Graph Analysis for EEG-Based Alzheimer’s. arXiv 2025, arXiv:2504.01329. [Google Scholar]
  27. Ajra, Z.; Belkacem, S.; El Khiat, H.; El Ansari, M. Shallow Neural Networks With Functional Connectivity. Front. Neurol. 2023, 14, 1270405. [Google Scholar] [CrossRef]
  28. Fruehwirt, W.; Steinheimer, J.M.; Scherer, R. Bayesian DNNs for AD Severity Estimation. arXiv 2018, arXiv:1812.04994. [Google Scholar]
  29. Ranjan, S.; Jaiswal, S.; Kumar, R. Subcortical EEG-Based AD and FTD Classification. arXiv 2024, arXiv:2408.10816. [Google Scholar]
  30. Sunkara, A.; Chandak, S.; Varshney, P.K. ANNs vs. KANs in EEG Alzheimer Detection. arXiv 2024, arXiv:2409.05989. [Google Scholar]
  31. Sharma, N.; Kolekar, M.H.; Jha, K. EEG based dementia diagnosis using multi-class support vector machine with motor speed cognitive test. Biomed. Signal Process. Control 2021, 63, 102102. [Google Scholar] [CrossRef]
  32. Praveena, D.M.; Sarah, D.A.; George, S.T. Deep learning techniques for EEG signal applications—A review. IETE J. Res. 2022, 68, 3030–3037. [Google Scholar] [CrossRef]
  33. Miltiadous, A. A dataset of 88 EEG recordings from: Alzheimer’s disease, Frontotemporal dementia and healthy subjects. OpenNeuro 2023, 1, ds004504. [Google Scholar] [CrossRef]
  34. Miltiadous, A.; Tzimourta, K.D.; Afrantou, T.; Ioannidis, P.; Grigoriadis, N.; Tsalikakis, D.G.; Angelidis, P.; Tsipouras, M.G.; Glavas, E.; Giannakeas, N.; et al. A dataset of scalp EEG recordings of Alzheimer’s disease, frontotemporal dementia and healthy subjects from routine EEG. Data 2023, 8, 95. [Google Scholar] [CrossRef]
  35. Türker, I.; Aksu, S. Connectogram—A graph-based time dependent representation for sounds. Appl. Acoust. 2022, 191, 108660. [Google Scholar] [CrossRef]
  36. Sarwinda, D.; Paradisa, R.H.; Bustamam, A.; Anggia, P. Deep Learning in Image Classification using Residual Network (ResNet) Variants for Detection of Colorectal Cancer. Procedia Comput. Sci. 2021, 179, 423–431. Available online: https://www.sciencedirect.com/science/article/pii/S1877050921000284 (accessed on 1 December 2024). [CrossRef]
  37. Eskandari, S.; Eslamian, A.; Munia, N.; Alqarni, A.; Cheng, Q. Evaluating Deep Learning Models for Breast Cancer Classification: A Comparative Study. arXiv 2025, arXiv:2408.16859v2. Available online: https://arxiv.org/pdf/2408.16859 (accessed on 1 December 2024). [Google Scholar]
  38. Ozdenizci, O.; Eldeeb, S.; Demir, A.; Erdoğmuş, D.; Akçakaya, M. EEG-based texture roughness classification in active tactile exploration with invariant representation learning networks. Biomed. Signal Process. Control 2021, 67, 102507. [Google Scholar] [CrossRef]
  39. Seo, J.; Laine, T.H.; Oh, G.; Sohn, K.A. EEG-based emotion classification for Alzheimer’s disease patients using conventional machine learning and recurrent neural network models. Sensors 2020, 20, 7212. [Google Scholar] [CrossRef]
  40. Dogan, S.; Baygin, M.; Tasci, B.; Loh, H.W.; Barua, P.D.; Tuncer, T.; Tan, R.-S.; Acharya, U.R. Primate brain pattern-based automated Alzheimer’s disease detection model using EEG signals. Cogn. Neurodynamics 2022, 17, 647–659. [Google Scholar] [CrossRef]
  41. Miltiadous, A.; Tzimourta, K.D.; Giannakeas, N.; Tsipouras, M.G.; Afrantou, T.; Ioannidis, P.; Tzallas, A.T. Alzheimer’s disease and frontotemporal dementia: A robust classification method of EEG signals and a comparison of validation methods. Diagnostics 2021, 11, 1437. [Google Scholar] [CrossRef]
  42. Araujo, T.; Teixeira, J.P.; Rodrigues, P.M. Smart-data-driven system for Alzheimer disease detection through electroencephalographic signals. Bioengineering 2022, 9, 141. [Google Scholar] [CrossRef]
  43. Gomez, S.R.; Gomez, C.; Poza, J.; Tobal, G.G.; Arribas, M.T.; Cano, M.; Hornero, R. Automated multiclass classification of spontaneous EEG activity in Alzheimer’s disease and mild cognitive impairment. Entropy 2018, 20, 35. [Google Scholar] [CrossRef] [PubMed]
  44. Khatun, S.; Morshed, B.I.; Bidelman, G.M. A single-channel EEG-based approach to detect mild cognitive impairment via speech-evoked brain responses. IEEE Trans. Neural Syst. Rehabil. Eng. 2019, 27, 1063–1070. [Google Scholar] [CrossRef] [PubMed]
  45. Gordon, E.; Rohrer, J.D.; Kim, L.G.; Omar, R.; Rossor, M.N.; Fox, N.C.; Warren, J.D. Measuring disease progression in frontotemporal lobar degeneration: A clinical and MRI study. Neurology 2010, 74, 666–673. [Google Scholar] [CrossRef] [PubMed]
  46. Lopes, M.; Cassani, R.; Falk, T.H. Using CNN saliency maps and EEG modulation spectra for improved and more interpretable machine learning-based Alzheimer’s disease diagnosis. Comput. Intell. Neurosci. 2023, 2023, 3198066. [Google Scholar] [CrossRef]
  47. Ieracitano, C.; Mammone, N.; Bramanti, A.; Hussain, A.; Morabito, F.C. A Convolutional Neural Network approach for classification of dementia stages based on 2D-spectral representation of EEG recordings. Neurocomputing 2019, 323, 96–107. [Google Scholar] [CrossRef]
  48. Xia, W.; Zhang, R.; Zhang, X.; Usman, M. A novel method for diagnosing Alzheimer’s disease using deep pyramid CNN based on EEG signals. Heliyon 2023, 9, e14858. [Google Scholar] [CrossRef]
  49. Xiaojun, B.; Haibo, W. Early Alzheimer’s disease diagnosis based on EEG spectral images using deep learning. Neural Netw. 2019, 114, 119–135. [Google Scholar]
  50. Siuly, S.; Alçin, Ö.F.; Kabir, E.; Sengür, A.; Wang, H.; Zhang, Y.; Whittaker, F. A New Framework for Automatic Detection of Patients With Mild Cognitive Impairment Using Resting-State EEG Signals. IEEE Trans. Neural Syst. Rehabil. Eng. 2020, 28, 1966–1976. [Google Scholar] [CrossRef]
  51. Ismail, M.; Hofmann, K.; Abd El Ghany, M.A. Early diagnoses of Alzheimer using EEG data and deep neural networks classification. In Proceedings of the 2019 IEEE Global Conference on Internet of Things (GCIoT), Dubai, United Arab Emirates, 4–7 December 2019. [Google Scholar] [CrossRef]
  52. Wen, D.; Zhou, Y.; Li, P.; Zhang, P.; Li, J.; Wang, Y.; Li, X.; Bian, Z.; Yin, S.; Xu, Y. Resting-state EEG signal classification of amnestic mild cognitive impairment with type 2 diabetes mellitus based on multispectral image and convolutional neural network. J. Neural Eng. 2020, 17, 036005. [Google Scholar] [CrossRef]
  53. Cassani, R.; Falk, T.H. Alzheimer’s Disease Diagnosis and Severity Level Detection Based on Electroencephalography Modulation Spectral ‘Patch’ Features. IEEE J. Biomed. Health Inform. 2020, 24, 1982–1993. [Google Scholar] [CrossRef]
  54. Huggins, C.J.; Escudero, J.; Parra, M.A.; Scally, B.; Anghinah, R.; De Arajujo, A.V.L.; Basile, L.F.; Abasolo, D. Deep learning of resting-state electroencephalogram signals for three-class classification of Alzheimer’s disease, mild cognitive impairment and healthy ageing. J. Neural Eng. 2021, 18, 046087. [Google Scholar] [CrossRef] [PubMed]
  55. Amini, M.; Pedram, M.M.; Moradi, A.R.; Ouchani, M. Diagnosis of Alzheimer’s Disease by Time-Dependent Power Spectrum Descriptors and Convolutional Neural Network Using EEG Signal. Comput. Math. Methods Med. 2021, 2021, 5511922. [Google Scholar] [CrossRef] [PubMed]
  56. Wu, L.; Zhao, Q.; Liu, J.; Yu, H. Efficient identification of Alzheimer’s brain dynamics with Spatial-Temporal Autoencoder: A deep learning approach for diagnosing brain disorders. Biomed. Signal Process. Control 2023, 86, 104917. [Google Scholar] [CrossRef]
  57. Zhou, H.; Yin, L.; Su, R.; Zhang, Y.; Yuan, Y.; Xie, P.; Li, X. STCGRU: A hybrid model based on CNN and BiGRU for mild cognitive impairment diagnosis. Comput. Methods Programs Biomed. 2024, 248, 108123. [Google Scholar] [CrossRef]
  58. Parra, C.R.; Torres, A.P.; Reolid, R.S.; Sotos, J.M.; Borjab, A.L. Inter-hospital moderate and advanced Alzheimer’s disease detection through convolutional neural networks. Heliyon 2024, 10, e26298. [Google Scholar] [CrossRef]
Figure 1. The visualization of an adjacency matrix that represents the functional brain network, derived from a 0.4 s time window of a 19-channel EEG recording.
Figure 1. The visualization of an adjacency matrix that represents the functional brain network, derived from a 0.4 s time window of a 19-channel EEG recording.
Diagnostics 15 01441 g001
Figure 2. The Connectogram-COH image derived from a 30 s multichannel recording. Each pixel column represents a flattened graph representation of the 0.4 s time window, forming the vertical dimension of the image. These 1-pixel flattened vectors, tiled horizontally for each time window, form the time resolution of the image on the horizontal axis.
Figure 2. The Connectogram-COH image derived from a 30 s multichannel recording. Each pixel column represents a flattened graph representation of the 0.4 s time window, forming the vertical dimension of the image. These 1-pixel flattened vectors, tiled horizontally for each time window, form the time resolution of the image on the horizontal axis.
Diagnostics 15 01441 g002
Figure 3. Illustration of the entire methodology presented. The time graph conversion is applied to equal-length EEG recordings, specifically 10, 20, or 30 s segments in this study. These recordings are handled in sliding time windows with a length of 0.4 s and 50% overlap (0.2 s). For each time window, a graph representation is generated based on the signal coherence values between EEG channels, corresponding to the edge weights, where each EEG channel (electrode) is a single node. Since the adjacency matrix of this graph is symmetrical, we extract the upper triangle and flatten it to obtain a vertical vector. For each sliding time window, a vertical graph representation is achieved, serving as a 1-pixel column that will be concatenated horizontally to generate a time graph representation. The resulting image, named Connectogram-COH, can be input to a well-known convolutional neural network model to perform the classification task.
Figure 3. Illustration of the entire methodology presented. The time graph conversion is applied to equal-length EEG recordings, specifically 10, 20, or 30 s segments in this study. These recordings are handled in sliding time windows with a length of 0.4 s and 50% overlap (0.2 s). For each time window, a graph representation is generated based on the signal coherence values between EEG channels, corresponding to the edge weights, where each EEG channel (electrode) is a single node. Since the adjacency matrix of this graph is symmetrical, we extract the upper triangle and flatten it to obtain a vertical vector. For each sliding time window, a vertical graph representation is achieved, serving as a 1-pixel column that will be concatenated horizontally to generate a time graph representation. The resulting image, named Connectogram-COH, can be input to a well-known convolutional neural network model to perform the classification task.
Diagnostics 15 01441 g003
Figure 4. The Connectogram-COH for “30”, “20”, and “10” s segments.
Figure 4. The Connectogram-COH for “30”, “20”, and “10” s segments.
Diagnostics 15 01441 g004
Figure 5. Learning curves for (a) CNN and (b) ResNet classifiers.
Figure 5. Learning curves for (a) CNN and (b) ResNet classifiers.
Diagnostics 15 01441 g005
Figure 6. Confusion matrices for CNN and ResNet classifiers.
Figure 6. Confusion matrices for CNN and ResNet classifiers.
Diagnostics 15 01441 g006
Figure 7. ROC curves of each of the algorithms used in this study.
Figure 7. ROC curves of each of the algorithms used in this study.
Diagnostics 15 01441 g007aDiagnostics 15 01441 g007b
Table 1. The custom CNN model used in this study.
Table 1. The custom CNN model used in this study.
LayerTypeOutput ShapeDetails
InputInput Layer(None, 171, 149, 1)-
Conv2DConvolutional Layer(None, 169, 147, 32)Filters: 32, Kernel: 3, Stride: 1
MaxPooling2DPooling Layer(None, 84, 73, 32)Pool Size: 2
Conv2DConvolutional Layer(None, 82, 71, 64)Filters: 64, Kernel: 3, Stride: 1
MaxPooling2DPooling Layer(None, 41, 35, 64)Pool Size: 2
Conv2DConvolutional Layer(None, 39, 33, 128)Filters: 128, Kernel: 3, Stride: 1
MaxPooling2DPooling Layer(None, 19, 16, 128)Pool Size: 2
FlattenFlatten Layer(None, 38,912)-
DenseFully Connected Layer(None, 128)Units: 128
DenseFully Connected Layer(None, 3)Units: 3 (Output classes)
Table 2. The custom ResNet model used in this study.
Table 2. The custom ResNet model used in this study.
LayerTypeOutput ShapeDetails
InputInput Layer(None, 171, 149)-
Conv1DConvolutional Layer(None, 86, 64)Filters: 64, Kernel: 3, Stride: 2
BatchNormBatch Normalization(None, 86, 64)-
ActivationReLU Activation(None, 86, 64)-
MaxPooling1DPooling Layer(None, 43, 64)Pool Size: 2
Residual Block 12x Conv + Add(None, 43, 64)Skip connection, Filters: 64, Kernel: 3
Residual Block 22x Conv + Add(None, 22, 128)Strided conv for downsampling, Filters: 128
Residual Block 32x Conv + Add(None, 11, 256)Strided conv for downsampling, Filters: 256
GlobalAvgPoolingGlobal Avg Pooling(None, 256)-
DenseFully Connected Layer(None, 3)Units: 3 (Output classes)
Table 3. Mean accuracy metrics for the learning models tested, performed for segmentations into 10, 20, and 30 s scenarios. For each scenario, the duration of the sliding window (0.4 s) and overlap (50%) remained the same. Some of the experiments could not be handled and are noted as N/A because the aspect ratio of the images was not suitable for the corresponding learning models.
Table 3. Mean accuracy metrics for the learning models tested, performed for segmentations into 10, 20, and 30 s scenarios. For each scenario, the duration of the sliding window (0.4 s) and overlap (50%) remained the same. Some of the experiments could not be handled and are noted as N/A because the aspect ratio of the images was not suitable for the corresponding learning models.
Model/Segment Len.30 s20 s10 s
CNN96.2897.8298.63
ResNet98.5999.0999.49
VVG 1677.6873.0473.55
InceptionV371.33N/AN/A
EfficientNetB741.7941.7341.71
DenseNet12169.3669.42N/A
Table 4. Mean classification report for the learning models tested, performed for segmentations into 30-s scenario. For each scenario, the duration of the sliding windows (0.4 s) and overlap (50%) remain the same.
Table 4. Mean classification report for the learning models tested, performed for segmentations into 30-s scenario. For each scenario, the duration of the sliding windows (0.4 s) and overlap (50%) remain the same.
AccuracyPrecisionRecallF1-ScoreSupport
ResNetCN0.98590.98750.98750.9875158
AD0.98590.98430.97410.9792191
FTD0.98590.96230.98080.9714108
CNNCN0.96720.96250.97470.9686158
AD0.96720.97370.96860.9711191
FTD0.96720.96260.95370.9581108
VVG16CN0.78770.72680.89240.8011158
AD0.78770.84710.75390.7978191
FTD0.78770.80650.69440.7463108
InceptionV3CN0.71550.71430.72780.7210158
AD0.71550.74040.80630.7719191
FTD0.71550.65910.53700.5918108
EfficientNetB7CN0.41790.00000.00000.0000158
AD0.41790.41791.00000.5895191
FTD0.41790.00000.00000.0000108
DenseNet121CN0.69800.70660.74680.7262158
AD0.69800.71980.68590.7024191
FTD0.69800.64810.64810.6481108
Table 5. The ROC curves for the learning models tested, performed for segmentations into 30 s scenario. For each scenario, the duration of the sliding window (0.4 s) and overlap (50%) remained the same.
Table 5. The ROC curves for the learning models tested, performed for segmentations into 30 s scenario. For each scenario, the duration of the sliding window (0.4 s) and overlap (50%) remained the same.
CNADFTD
ResNet0.99950.99890.9999
CNN0.99630.99240.9934
VVG160.92530.90660.9160
InceptionV30.87570.86580.8489
EfficientNetB70.55930.52880.5429
DenseNet1210.87160.83200.8726
Table 6. Mean accuracy metrics of the ResNet model with our approach using different segment lengths, epochs, and batch sizes, with each cell representing the average output of 5 executions.
Table 6. Mean accuracy metrics of the ResNet model with our approach using different segment lengths, epochs, and batch sizes, with each cell representing the average output of 5 executions.
Epochs\Batch_Size48163264128256512
10 s segments2098.7799.4998.9998.4197.6297.4794.5271.54
5099.6399.4299.0699.4999.4998.4195.2488.61
7099.3599.2799.2799.4999.5699.4296.5889.21
10099.4299.2799.2099.4999.6398.9997.8191.18
20 s segments2093.1895.9497.8298.6399.2789.2771.1550.28
5099.4299.4298.4098.8298.8497.2489.1970.43
7099.1399.2798.9899.2799.4298.8189.2779.56
10098.2699.5699.4298.9997.3995.7989.5677.82
30 s segments2096.4998.9098.0398.0393.9381.1662.4558.29
5097.3797.5998.2497.8198.0395.1889.2770.32
7098.0397.5998.0398.2498.2496.1487.2371.33
10098.9098.8698.4198.4198.5994.2188.5571.99
Table 7. Comparison of the current study with recent studies. (AD: Alzheimer’s disease, MCI: mild cognitive impairment, CN: healthy control, FTD: frontotemporal dementia, CV: cross-validation.
Table 7. Comparison of the current study with recent studies. (AD: Alzheimer’s disease, MCI: mild cognitive impairment, CN: healthy control, FTD: frontotemporal dementia, CV: cross-validation.
Author(s)YearClassifierSize of the DatasetNo. of Channels Segment Length (s)Folds for CVAccuracy
Gomez et al. [43]2018MLP11119--AD-MCI-CN: 78.43
Xiaojun & Haibo [49]2019CNN1264-1–1095.04
Ieracitano et al. [47]2019CNN1891958AD-CN: 92.95
AD-MCI: 84.61
MCI-CN: 91.88
AD-MCI-CN: 83.33
Khatun et al. [44]2019ERP SVM231--87.9
Ismail et al. [51]2019CNN601016-AD-CN: 92.52
MCI-CN: 90.36
Wen et al. [52]2020CNN3919-592.92
Siluy et al. [50]2020ELM
SVM
KNN
2719210ELM: 98.78
SVM: 97.41
KNN: 98.19
Cassani et al. [53]2020SVM54208-78.7
Safi & Safi [11]2021SVM
KNN
RLDA
86208-SVM: 95.79
KNN: 97.64
RLDA: 97.02
Miltiadous et al. [41]2021Meny2819510AD-CN: 78.58
FTD-CN: 86.30
Huggins et al. [54]2021CNN14120 51099.3 AD
98.3 MCI
98.8 CN
Amini et al. [55]2021CNN19219--82.3
Dogan et al. [40]2022KNN2316-1092.1
Araujo et al. [42]2022SVM38195-AD-CN: 81
MCI-CN: 79
Ding et al. [4]2022Meny30160155AD-CN: 72.43
AD-MCI: 69.11
MCI-CN: 59.91
Xia et al. [48]2023CNN10019-5AD-MCI-CN: 97.10
Wu et al. [56]2023STAE5316 1-96.30
Lopes et al. [46]2023CNN SVM54208-87.3
Miltiadous et al. [10]2023DICE-net8819 305AD-CN: 83.28
FTD-CN: 74.96
Zhou et al. [57]2024STCGRU2719 5 10MCI: 99.95
Parra et al. [58]2024CNN66832-5CN-ADA: 97.49
CN-ADM: 97.03
Our study2024ResNet8819ch30
20
10
5AD-CN: 99.53
AD-FTD: 99.45
FTD-CN: 99.50
AD-FTD-CN: 99.41
Table 8. An overview of performance outputs.
Table 8. An overview of performance outputs.
MetricValueComment
Trainable Parameters979,715Medium-sized model. Can run on desktop, mid-level mobile, or embedded edge devices.
Model Size11.45 MBCompact enough for mobile apps or cloud API deployment.
Inference Latency147.4 ms (1 sample)Good for near real-time processing, but may need optimization for ultra low-latency apps (e.g., BCI, live EEG).
Throughput234.4 samples/sVery efficient batch processing—good for offline or background inference.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Aljanabi, E.; Türker, İ. Connectogram-COH: A Coherence-Based Time-Graph Representation for EEG-Based Alzheimer’s Disease Detection. Diagnostics 2025, 15, 1441. https://doi.org/10.3390/diagnostics15111441

AMA Style

Aljanabi E, Türker İ. Connectogram-COH: A Coherence-Based Time-Graph Representation for EEG-Based Alzheimer’s Disease Detection. Diagnostics. 2025; 15(11):1441. https://doi.org/10.3390/diagnostics15111441

Chicago/Turabian Style

Aljanabi, Ehssan, and İlker Türker. 2025. "Connectogram-COH: A Coherence-Based Time-Graph Representation for EEG-Based Alzheimer’s Disease Detection" Diagnostics 15, no. 11: 1441. https://doi.org/10.3390/diagnostics15111441

APA Style

Aljanabi, E., & Türker, İ. (2025). Connectogram-COH: A Coherence-Based Time-Graph Representation for EEG-Based Alzheimer’s Disease Detection. Diagnostics, 15(11), 1441. https://doi.org/10.3390/diagnostics15111441

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop