Next Article in Journal
Socioeconomic Covariate-Dependent Bayesian Nonparametric Mixture Model for Household Spending Patterns to Identify Multidimensional Vulnerability
Previous Article in Journal
Parameter-Efficient Fine-Tuning via General Linear Structural Regularization for High-Rank Adaptation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Early Mild Cognitive Impairment Diagnosis via Resting-State fMRI Brain Networks Using a Region-Specific Hierarchical Fusion Graph Neural Network

1
College of Information Engineering, Shanghai Maritime University, Shanghai 201306, China
2
Faculty of Engineering, The Chinese University of Hong Kong, Hong Kong 999077, China
*
Author to whom correspondence should be addressed.
Information 2026, 17(5), 461; https://doi.org/10.3390/info17050461
Submission received: 24 March 2026 / Revised: 19 April 2026 / Accepted: 28 April 2026 / Published: 9 May 2026
(This article belongs to the Section Biomedical Information and Health)

Abstract

Early mild cognitive impairment (EMCI) is the earliest intervenable stage of Alzheimer’s disease (AD). Although graph neural networks (GNNs) have begun to exploit brain network topology, traditional fMRI-based diagnostic methods often neglect these structural patterns by relying on vectorized features. Furthermore, existing GNNs frequently disregard inter-regional functional heterogeneity and group-level discriminative patterns, leading to limited accuracy and biomarker interpretability. To address these challenges, we propose HF-BrainGNN, an end-to-end hierarchical graph learning framework for EMCI identification. Our method introduces a functional affinity region convolution (FAR-Conv) layer to learn region-adaptive kernels, a Differential Focus Pooling (DF-Pool) module to identify disease-salient brain regions by maximizing inter-group distinctiveness, and a hierarchical integration classifier (HIC) to fuse multi-level graph representations. The framework is optimized using classification, focus separation, and consistency regularization losses. Experiments on the ADNI dataset (104 EMCI, 114 Cognitively Normal) show that HF-BrainGNN achieves 86.78% accuracy, outperforming the best baseline (Hi-GCN) by 4.64%. Furthermore, the automatically identified regions, such as the bilateral hippocampus and default mode network hubs, align with established EMCI biomarkers. Ultimately, HF-BrainGNN provides an efficient, interpretable artificial intelligence tool for precise brain network characterization and early AD intervention.

1. Introduction

Alzheimer’s disease (AD) is a progressive neurodegenerative disorder affecting more than 50 million individuals worldwide, and the number is projected to reach 152 million by 2050 [1]. AD typically develops from mild cognitive impairment (MCI), whereas early mild cognitive impairment (EMCI) represents the earliest clinically detectable and potentially intervenable stage [2]. Although EMCI manifests only a subtle cognitive decline that does not severely impact daily life, its timely and accurate identification is crucial for delaying the progression toward dementia [3]. Due to the mild and non-specific nature of EMCI symptoms, diagnosis based solely on clinical evaluation remains challenging, highlighting the need for objective neuroimaging-based biomarkers [4].
Functional magnetic resonance imaging (fMRI) provides a non-invasive means of quantifying brain functional activity and connectivity, offering critical insight into early neural abnormalities associated with AD [5]. Resting-state fMRI (rs-fMRI), which requires no cognitive task, is particularly suitable for clinical and research settings involving patients with varying levels of cognitive ability. The brain functional connectivity networks derived from rs-fMRI reflect both local and global neural interactions and have been widely used to investigate the full AD continuum [6,7].
Traditional approaches for rs-fMRI analysis commonly rely on seed-based correlation mapping, independent component analysis (ICA), or graph-theoretic metrics [8,9]. These methods typically follow a two-stage pipeline: handcrafted feature extraction followed by classification via conventional machine-learning algorithms. For instance, Khazaee et al. [10] employed graph-theoretic features combined with support vector machines (SVMs) for AD classification; Wee et al. [11] integrated volumetric and diffusion tensor imaging (DTI) data in multimodal models; and Tijms et al. [12] summarized graph-based biomarkers across the AD spectrum. Although informative, such two-stage strategies depend heavily on manual feature design and prior knowledge, limiting end-to-end optimization and interpretability [13].
Recently, deep learning—especially graph neural networks (GNNs)—has offered a powerful framework for modeling brain connectivity data [14]. Through message passing on graph structures, GNNs jointly learn node attributes and topological relations in an end-to-end manner, enabling automated discovery of discriminative connectivity patterns. Notable examples include population-graph GNNs for autism spectrum disorder [15], the BrainGNN architecture for fMRI-based biomarker detection [16], and dynamic attention networks for capturing temporal connectivity variations [17]. These advances demonstrate that GNNs can extract complex, informative representations of brain disorders.
However, several critical challenges remain in applying GNNs to fMRI-based cognitive impairment analysis. First, most models adopt unified convolution kernels for all nodes, disregarding the functional specialization and anatomical diversity among brain regions. Second, existing graph pooling methods predominantly assess node importance within individual graphs, neglecting systematic differences between subject groups—thereby limiting their ability to identify disease-specific regions. Third, many GNN frameworks depend only on final-layer features for classification, ignoring useful multi-level representations obtained during hierarchical propagation. Incorporating multi-hop neighborhood information across layers would allow more comprehensive integration of local and global connectivity patterns.
Specifically, to address regional heterogeneity, we introduce the functional affinity region convolution (FAR-Conv) layer. It constructs region-specific kernels by integrating functional similarity with anatomical proximity, enabling the model to capture diverse inter-regional dependencies. Simultaneously, to uncover group-level biomarkers, the differential focus pooling (DF-Pool) layer compares inter-group focus score distributions to identify disease-relevant regions, performing dimensionality reduction via a soft mask-based mechanism. Finally, a hierarchical integration classifier (HIC) is designed to adaptively fuse multi-level features with learnable weights, ensuring a robust diagnosis by combining local and global perspectives.
Furthermore, the framework optimizes three complementary loss functions—classification, focus separation, and consistency regularization—to strike a balance between diagnostic accuracy and biological interpretability.
The proposed HF-BrainGNN was evaluated on resting-state fMRI data from the ADNI cohort, including 104 EMCI patients and 114 cognitively normal individuals. Experimental results demonstrate superior classification performance compared with existing methods, achieving improved accuracy, sensitivity, and specificity. Moreover, the key brain regions identified by DF-Pool correspond closely to EMCI-associated areas reported in previous neuroscience studies [18], underscoring the physiological plausibility and clinical interpretability of the proposed model.

2. Method

This section elaborates on the HF-BrainGNN framework for EMCI classification. As illustrated in Figure 1, the overall workflow consists of three distinct stages: (1) brain network construction (Section 2.1), where fMRI data is processed via atlas-based parcellation to generate functional connectivity networks; (2) HF-BrainGNN learning, which extracts hierarchical features and identifies discriminative regions (Section 2.2, Section 2.3, Section 2.4, Section 2.5 and Section 2.6); and (3) diagnosis and interpretation, where the model outputs classification results and highlights potential biomarkers. The following subsections detail each component of this pipeline.

2.1. Brain Network Construction

In this study, the Automated Anatomical Labeling (AAL90) template is used to parcellate the brain into N anatomical regions ( N = 90 ), providing a standardized anatomy-based reference widely adopted in neuroimaging studies [19]. Each region of interest (ROI) corresponds to one node in the brain graph; namely, V = { v 1 , v 2 , , v N } .
For each subject, a Pearson correlation matrix R R N × N is first computed based on the mean time series of all ROIs. Each element R i j reflects the functional connectivity (FC) strength between regions v i and v j . The adjacency matrix A is then derived from R to define the graph topology. To emphasize the most significant functional connections, we threshold the correlation matrix by retaining only the top 20% of absolute correlation coefficients. Specifically, elements within the top 20th percentile are set to 1 (indicating an existing edge), and the remainder are set to 0, thereby forming a binary adjacency matrix A.
After constructing the brain graph G = { V , A } , node features are defined from the same FC matrix. The feature of node v i is represented by its corresponding FC vector x i = R i , where R i denotes the i-th row of R. Thus, the initial node feature matrix can be expressed as X ( 0 ) = [ x 1 , x 2 , , x N ] T . In this study, only FC information is used to form node features, without introducing additional morphological or clinical covariates [20]. All notations used in this section are summarized in Table 1.

2.2. Overview of the HF-BrainGNN Method

The proposed hierarchical focused brain graph neural network (HF-BrainGNN) is illustrated in Figure 2a. It comprises three main components: the FAR-Conv layer, the differential focus pooling layer, and the hierarchical integration classifier. The model takes a single-subject brain graph G = ( V , A ) and its initial node features X ( 0 ) (formed by FC row vectors) as input.
As shown in Figure 2a, HF-BrainGNN contains two cascaded processing layers, each integrating a FAR-Conv layer, a differential focus pooling (DF-Pool) layer, and a global pooling operation (Mean/Max Pool) to obtain layer-specific graph representations ( g 1 , g 2 ). These representations are adaptively fused in the hierarchical integration module and fed into a multilayer perceptron (MLP) for EMCI/CN classification.
Within each hierarchical brain processing unit (HBPU), FAR-Conv computes regional affinity vectors to generate region-specific convolution kernels and performs two-hop message passing, enabling multi-level neighborhood aggregation to yield representations H ( l ) for the two cascaded processing layers ( l = 1 , 2 ). DF-Pool then calculates inter-group-enhanced focus scores to select the top-k disease-relevant brain regions and outputs the induced subgraph for subsequent layer-wise processing.
Finally, the hierarchical integration classifier performs weighted fusion of the readout vectors from all layers and outputs the final class probabilities through the MLP. The entire pipeline proceeds from feature extraction (FAR-Conv) to region selection (DF-Pool), and then to feature integration and classification. Three complementary loss functions are jointly optimized to ensure both strong predictive performance and interpretability, as detailed in Section 2.6.

2.3. Functional Affinity Region Convolutional Layer

Graph neural networks have shown promising potential in brain network analysis; however, there is a notable limitation in their application: existing models typically use a unified convolution kernel for feature extraction across all nodes. This approach may fail to fully capture the unique functional characteristics and anatomical positional differences of distinct brain regions.
Neuroscience studies have demonstrated that brain regions exhibit clear functional specialization and complex interaction patterns [21]. For instance, the prefrontal cortex primarily mediates executive functions, while the temporal lobe is involved in auditory and language processing. This functional differentiation and regional specificity are critical for a comprehensive understanding of brain network dynamics, and thus require special consideration in computational model design.
To address this issue, we propose the FAR-Conv layer, which accounts for the functional similarity and anatomical positional characteristics of brain regions, assigning region-specific embedding strategies to different brain areas. The core idea of FAR-Conv is to enhance information exchange between functionally similar brain regions while preserving the anatomical specificity of each region, as illustrated in Figure 2b. For each brain region node v i , we introduce a regional affinity vector α i R K + 1 , which represents the membership degree of node v i to each functional region [22]. In this study, K = 8 functional networks are defined according to the commonly adopted eight-network division widely used in neuroscience research [23]. To incorporate multi-dimensional positional information p i , the regional affinity vector is defined as α i = [ d i 1 , d i 2 , , d i K , p i ] , where d i k = c i C k 2 denotes the Euclidean distance from brain region i to the centroid of the k-th functional network; c i R 3 represents the normalized 3D coordinate of the brain region’s center, p i is the positional encoding of node v i .
The positional encoding adopts the following form:
p i = sin c i τ , cos c i τ
where c i = ( x i , y i , z i ) is the 3D coordinate vector of the brain region, and τ is a scale parameter. This formulation provides a 6-dimensional positional embedding (3 sine and 3 cosine components), enabling the model to distinguish brain regions located in different anatomical positions even if they exhibit similar FC patterns.
The centroid coordinate of the k-th functional network is calculated as follows:
C K = 1 | n K | v i n K pos ( v i )
where C K is the centroid coordinate of network K, n K is the set of brain regions in the k-th functional network, and pos ( v i ) denotes the 3D coordinate of brain region v i .
Based on the regional affinity, the region-specific convolution kernel W i ( l ) for each node is computed as follows:
W i ( l ) = W share ( l ) + k = 1 K softmax ( d i k ) B k ( l )
where W share ( l ) is the shared base weight, and B k ( l ) is the base weight matrix of the k-th functional network. Here, d i k represents the Euclidean distance between brain region v i and the centroid of network k. To ensure that brain regions located closer to a functional network receive higher aggregation weights, we apply the negative distance ( d i k ) within the softmax function, effectively converting the distance metric into a similarity measure. The superscript l indexes the hierarchical processing layer, with each layer comprising one FAR-Conv and one DF-Pool operation. Two such cascaded layers are stacked in the proposed architecture.
Information transfer between functionally similar brain regions is realized via an edge weight enhancement factor:
e i j = e i j 0 · σ α i T α j
Here, σ denotes the sigmoid function, which ensures the enhancement factor is constrained to the range ( 0 , 1 ) , and e i j 0 is the original edge weight. This design strengthens message passing between node pairs with similar regional affinity, while weakening message passing between dissimilar node pairs.
The final forward propagation of FAR-Conv is given by the following:
h i ( l ) = ReLU W i ( l ) x i ( l ) + j N ( l ) ( i ) e i j · W j ( l ) x j ( l )
where x i ( l ) represents the feature representation of node v i at the l-th layer, W i ( l ) is the region-specific graph convolution kernel of node v i , e i j is the edge weight enhancement factor based on regional affinity, N ( l ) ( i ) denotes the neighbor set of node v i at the l-th layer, and W j ( l ) is the region-specific graph convolution kernel of the neighbor node v j .

2.4. Differential Focus Pooling

Graph pooling, a fundamental operation in graph neural networks, plays a crucial role in reducing computational complexity and extracting salient features [24,25]. However, conventional pooling methods mainly evaluate node importance within individual graphs or datasets, without explicitly modeling inter-group discriminative patterns. In brain network analysis, this limitation reduces the ability to identify disease-specific key regions [26].
Motivated by these observations, we propose the differential focus pooling (DF-Pool) mechanism (Figure 3), which automatically identifies disease-related brain regions by comparing node-level representations across subject groups: it first computes node focus scores for each subject, then estimates inter-group differences to highlight highly discriminative regions, and finally performs threshold-based pooling to retain the most informative nodes, enabling targeted feature extraction and effective dimensionality reduction. Grounded in a widely accepted neuroimaging principle—that brain regions with high diagnostic value typically exhibit significant group-wise differences between patients and healthy controls [27] (a principle validated across multiple neuropsychiatric disorders and providing a theoretical basis for discovering informative differential biomarkers)—we design a three-step procedure for DF-Pool implementation: (1) node focus score calculation, (2) inter-group difference estimation, and (3) differential node selection.
First, we compute the focus score for each node in a brain graph to quantify its contribution to the classification objective:
f m = H m ( l ) · v ( l ) / v ( l ) 2 ,
where H m ( l ) is the hidden feature matrix output by FAR-Conv at the l-th layer, v ( l ) is a learnable projection vector, and v ( l ) 2 is its normalization factor. The focus score f m measures the projection strength of node features along the direction of v ( l ) , reflecting node saliency with respect to the learned representation.
We then apply min–max normalization to constrain the focus scores within the interval [ 0 , 1 ] :
f ˜ m = f m min ( f m ) / max ( f m ) min ( f m ) ,
which ensures a consistent range and preserves the relative ranking of node importance, preventing bias from extreme values.
Next, we estimate inter-group differences among subjects. For each class c, we compute its group-level average focus score F c (Equation (8)), and derive the inter-group difference matrix D (Equation (9)):
F c = 1 | S c | m S c f ˜ m ,
D = | F c 1 F c 2 | F c 1 + F c 2 2 ,
where S c denotes the set of subjects belonging to class c, and | S c | is the number of subjects. The Hadamard product operator (⊙) jointly encodes two complementary factors [28]: the absolute difference | F c 1 F c 2 | quantifies the inter-group discrepancy, while the average activation level ( F c 1 + F c 2 ) / 2 captures the overall node activity. This design emphasizes brain regions that show both strong group-wise discriminability and high activation, which are likely to correspond to disease-specific biomarkers. The group-level average focus F c and difference matrix D serve as intermediate representations that guide the subsequent node selection, emphasizing nodes with stronger inter-group discriminability.
Finally, threshold-based pooling is performed to retain the most informative nodes:
s m pool = f ˜ m D ,
During the implementation, the group-level average focus F c and the difference matrix D are computed solely using the subjects in the training set to prevent data leakage. In the inference phase, the matrix D remains fixed as a learned prior. For a single unseen subject, the pooling operation (Equation (10)) relies only on the subject’s individual focus score and the pre-trained D, ensuring the diagnosis is independent of ground-truth labels or other test data.
X m ( l + 1 ) = ( H m ( l ) f ˜ m ) i d x ,
A m ( l + 1 ) = A m ( l ) ( i d x , i d x ) ,
where i d x denotes the spatial index set of the selected brain regions, and the top - k operation returns the indices corresponding to the k highest values in the comprehensive score vector s m pool . We set k = N / 2 , meaning each pooling layer preserves 50% of the nodes [29]. Here, X m ( l + 1 ) and A m ( l + 1 ) represent the pooled node feature matrix and corresponding adjacency matrix.
A soft mask mechanism is employed to smooth transitions in feature importance before node selection [30]. By weighting features with the focus scores ( H m ( l ) f ˜ m ) prior to top - k pooling, this approach (1) reduces abrupt information loss by retaining partial contributions from low-focus nodes, and (2) stabilizes training through continuous feature importance adjustment.

2.5. Hierarchical Integration Classification

Existing brain network analysis methods typically rely on the feature representation from the final layer for classification [31]. However, prior studies have shown that such a single-layer classification strategy may constrain diagnostic accuracy, as it overlooks the complementary information contained in features across different layers. Recent work further suggests that representations derived from multiple network depths offer diverse and complementary diagnostic cues for various neurological diseases [32,33].
Motivated by these findings, multi-level feature fusion has emerged as a critical strategy for improving brain disease classification. Inspired by this concept, we propose a hierarchical integration classification (HIC) framework. Specifically, we stack FAR-Conv layers and DF-Pool modules to construct a two-level architecture that captures hierarchical brain FC patterns (i.e., multi-hop interactions among ROI nodes).
The proposed network architecture comprises multiple regional processing modules connected sequentially (as illustrated in Figure 2). Each regional processing module consists of one FAR-Conv layer followed by one DF-Pool operation. The core objective of this design is to enable each module to extract brain network features at different representational levels and then systematically integrate these multi-level features for disease classification.
For the output graph G = ( V , A ) of the l-th regional processing module, with the corresponding node feature matrix X ( l ) , a feature aggregation function is applied to obtain a fixed-size graph-level representation:
S ( l ) = mean X ( l ) max X ( l )
Here, mean X ( l ) computes the average node features to capture global distribution patterns, while max X ( l ) extracts the most salient feature values in each dimension. The concatenation operator (‖) then combines these two representations, preserving both global statistical and local discriminative information to produce a comprehensive graph-level feature embedding.
Unlike conventional methods that merely concatenate representations from different layers, we introduce an adaptive weighted integration mechanism:
α l = exp ( e l ) j = 1 L exp ( e j ) ,
Ψ = l = 1 L α l · Proj l S ( l ) ,
where α l represents the adaptive layer-wise weight computed via a softmax normalization, e l is a learnable scalar indicating the relative importance of the l-th layer, and Proj l ( · ) denotes a projection function that maps features from different layers into a unified embedding space. All parameters are jointly optimized through backpropagation, enabling the network to automatically determine the contribution of each layer in the overall classification process.
Finally, the integrated feature vector Ψ is passed through a multilayer perceptron (MLP) followed by a softmax activation to obtain the final predicted label:
y ^ = softmax MLP Ψ

2.6. Loss Function Design

To optimize our proposed HF-BrainGNN model—particularly to guide effective learning of the hierarchical integration classifier introduced in Section 2.3—we design three complementary loss functions. These loss functions target classification accuracy, feature selection clarity, and intra-group consistency, respectively, together forming a multi-objective optimization framework.

2.6.1. Classification Loss

As the primary optimization objective, we adopt the standard binary cross-entropy loss function [34], which effectively addresses the class imbalance issue common in brain disease datasets:
L C E = 1 M m = 1 M y m log y ^ m + ( 1 y m ) log 1 y ^ m
Here, y m { 0 , 1 } denotes the true class label of subject m, and y ^ m represents the model’s predicted probability that the subject belongs to the positive class (EMCI). y ^ m is the predicted probability output by the hierarchical integration classifier (introduced in Section 2.5).

2.6.2. Focus Separation Loss

To enhance the feature selection capability of the DF-Pool layer and guide the hierarchical integration classifier toward learning distinct and discriminative node representations, we design the focus separation (FS) loss:
L F S = 1 M m = 1 M 1 k i = 1 k f ˜ m , ( i ) 1 N k j = k + 1 N f ˜ m , ( j )
The core idea of this loss is to maximize the difference between the average focus scores of the selected (top-k) nodes and the unselected nodes. For each subject m, we first sort the normalized focus scores f ˜ m in descending order. Here, f ˜ m , ( i ) denotes the i-th highest normalized focus score. We then compute the mean score of the top-k selected nodes and that of the remaining ( N k ) unselected nodes, taking the negative of their difference. Minimizing L F S thus explicitly enforces the pooling layer to produce a sharper contrast between informative and less informative brain regions.

2.6.3. Consistency Regularization Loss

To balance the individual variability and within-class consistency of brain disease patterns [35], we introduce the consistency regularization (CR) loss:
L C R = c = 1 C 1 | I c | 2 m , n I c f ˜ m f ˜ n 1 L F S = 1 M m = 1 M 1 k i = 1 k f ˜ m , ( i ) 1 N k j = k + 1 N f ˜ m , ( j )
Here, I c denotes the index set of subjects belonging to class c, and | I c | is the number of subjects in that class. This term computes the average L1-distance between the top-k node focus scores of different subjects within the same class. The L1 norm is adopted to reduce the influence of outliers and improve robustness.
By constraining subjects of the same class to maintain similar focus distributions, this regularization promotes group-level consistency. It enables the hierarchical integration classifier to form stable, class-specific attention weight patterns, thereby strengthening discriminative feature representations and improving classification performance.

2.6.4. Total Loss

We combine the three loss functions into a total loss function, where the weights of each component are adjusted via hyperparameters:
L total = L C E + λ 1 L F S + λ 2 L C R
This multi-objective optimization design enables the model to pursue multiple goals simultaneously: L C E ensures prediction accuracy by directly optimizing the output of the hierarchical integration classifier; L F S enhances the clarity of node selection; L C R balances individual specificity and group consistency, thereby improving the stability of hierarchical integration.
The hyperparameters λ 1 and λ 2 play a key balancing role: by tuning their values, we can trade off between the model’s classification performance and interpretability [36].

3. Experiments and Results

3.1. Datasets

We used public data from the Alzheimer’s Disease Neuroimaging Initiative (ADNI; https://adni.loni.usc.edu, (accessed on 6 March 2025)), which includes resting-state functional magnetic resonance imaging (rs-fMRI) data from 104 EMCI patients and 114 cognitively normal (CN) participants. The demographic and clinical characteristics of the study participants are summarized in Table 2. To quantify the cognitive status of the participants, we utilized the Mini-Mental State Examination (MMSE), a widely recognized 30-point clinical tool used to screen for cognitive impairment and dementia. MMSE scores range from 0 to 30, where higher scores indicate better cognitive function. Typically, a score of 27–30 is considered cognitively normal, while lower scores reflect varying degrees of cognitive decline. All data were acquired using a Philips Medical Systems 3.0T MRI scanner, with the following scan parameters: repetition time (TR) = 3000 ms, echo time (TE) = 30 ms, flip angle = 80°, slice thickness = 3.3 mm, and number of slices = 48.
Data preprocessing was performed using the GRETNA software package [37], with key steps including: discarding the first 10 time points, temporal slice correction, head motion correction, spatial normalization to the Montreal Neurological Institute (MNI) space (resolution: 3 × 3 × 3 mm3), spatial smoothing (6 mm full-width at half-maximum (FWHM) Gaussian kernel), linear trend removal, and band-pass filtering (0.01–0.1 Hz). These preprocessing steps conform to the widely accepted standards for rs-fMRI data analysis [38].

3.2. Experimental Setup

All experiments in this study were implemented in the PyTorch framework (version 2.5.1), and training and testing were conducted on a workstation equipped with an NVIDIA RTX 3090Ti GPU (24 GB memory). The model architecture (Figure 2) consists of two affinity-aware regional convolution (FAR-Conv) layers and two differential focus pooling (DF-Pool) layers, forming two complete regional processing modules.
For the EMCI vs. CN classification task, we set the parameters as follows: N = 90 (number of brain regions in the AAL template); K ( 0 ) = K ( 1 ) = 8 (number of functional networks); d ( 0 ) = 90 (initial feature dimension); d ( 1 ) = 32 (output feature dimension of the first layer); d ( 2 ) = 32 (output feature dimension of the second layer); and C = 2 (number of classes).
The pooling rate is set to 0.5, meaning each pooling layer retains 50% of the nodes; after two pooling layers, approximately 25% of the original nodes are preserved as key brain regions. The setting of K = 8 (number of functional communities) is based on the commonly used eight functional network divisions in neuroscience research [23].
We use 5-fold cross-validation to evaluate model performance: the entire dataset is randomly partitioned into five mutually exclusive folds of approximately equal size. In each fold, four folds (80% of the data) are used as the training set (with 20% of the training data further split as the validation set for hyperparameter tuning), and the remaining one fold (20%) is used as the test set. This process is repeated five times (each fold serving as the test set once), ensuring that each subject’s data is only included in the test set once and remains independent across folds. The final performance metrics are the average of the results from the five folds.
The model is trained using the Adam optimizer, with an initial learning rate of 0.001 (halved every 30 epochs), a weight decay of 0.0005, a batch size of 32, and a maximum training epoch of 120. An early stopping strategy (training stops if validation set accuracy does not improve for 15 consecutive epochs) is applied to avoid overfitting.
For the hyperparameters of the loss functions, we tune λ 1 (focus separation loss coefficient) and λ 2 (consistency regularization loss coefficient) using the training and validation sets. Model evaluation metrics include accuracy, sensitivity, specificity, and AUC; all results are reported as the mean and standard deviation across 5-fold cross-validation.

3.3. Hyperparameter Discussion and Ablation Study

3.3.1. Hyperparameter Settings

To examine the impact of hyperparameters on model performance, we conducted a grid search to tune λ 1 and λ 2 in the loss function using the training and validation sets, with the search ranges set as λ 1 { 0 , 0.01 , 0.05 , 0.1 , 0.2 , 0.5 } and λ 2 { 0 , 0.01 , 0.05 , 0.1 , 0.2 , 0.5 , 1.0 } . As defined in Section 2.6, λ 1 controls the focus separation (FS) loss, where larger values encourage more distinct focus scores, and λ 2 controls the consistency regularization (CR) loss, balancing individual diversity and group-level consistency. A larger λ 2 promotes shared group-level patterns, while a smaller one captures individual-specific variations.
To evaluate the proposed affinity-aware regional convolution (FAR-Conv), we conducted an ablation study by replacing the FAR-Conv layer with a standard graph convolution using shared kernels (Standard-Conv). The performance gap between the two reflects the effect of regional adaptivity.
We adopted a heuristic tuning procedure: first fixing λ 1 to 0 or 0.1 while adjusting λ 2 , and then optimizing λ 1 based on the best λ 2 . Figure 4 shows the impact of different hyperparameter combinations on validation accuracy.

3.3.2. Ablation Study

We systematically examine the effects of λ 1 (focus separation, FS loss) and λ 2 (consistency regularization, CR loss), and evaluate the contribution of FAR-Conv compared with the standard convolution version.
Figure 4 illustrates how different loss coefficients affect validation performance for both variants.
When λ 1 = 0 , the accuracy remains relatively stable while λ 2 [ 0 , 0.5 ] , but declines when λ 2 reaches 1.0; Figure 4a. The highest accuracy occurs at λ 2 = 0.1 , indicating that moderate consistency regularization effectively alleviates overfitting, whereas excessive constraints lead to underfitting.
As λ 1 increases to 0.1 (Figure 4b), performance decreases when λ 2 exceeds 0.2. Notably, the results at λ 2 = 0 are higher than those obtained with λ 1 = 0 , showing that FS loss itself introduces a mild regularization effect that stabilizes training.
After fixing λ 2 = 0.1 , varying λ 1 between 0 and 0.5 yields the best accuracy at λ 1 = 0.1 (Figure 4c), while stronger FS constraints ( λ 1 = 0.2 or 0.5) slightly reduce performance, suggesting an optimal trade-off between inter-group differentiation and model flexibility.
Across all hyperparameter settings, FAR-Conv consistently surpasses Standard-Conv, as clearly shown in every subplot. This performance gap highlights FAR-Conv’s ability to assign region-specific convolution kernels, effectively capturing the functional heterogeneity of brain networks that uniform kernels overlook.
Based on the comprehensive analysis, λ 1 = 0.1 and λ 2 = 0.1 are selected as the optimal coefficient combination, achieving the highest validation accuracy. This configuration is therefore adopted for all subsequent experiments, with results evaluated on an independent test set for reliability.

3.4. Comparison with Baseline Methods

To evaluate the effectiveness of the proposed HF-BrainGNN, we compare it with both traditional machine learning models and widely used deep learning approaches on the EMCI vs. CN classification task. Traditional handcrafted methods include ridge classifier [39], support vector machine (SVM) with RBF kernel [40], and random forest with 100 estimators [41]. These models use the vectorized FC matrix as input features and are implemented via the scikit-learn library [42]. For deep learning methods, we adopt a standard deep neural network (DNN) [43]. Graph-based neural network models specifically designed for brain analysis include GCN [44], GAT [45], BrainGNN [16], MVS-GCN [46], PopulationGCN [47], and Hi-GCN [48]. It is important to note that while models like PopulationGCN [47] utilize a population-graph approach (where nodes represent subjects), our HF-BrainGNN and other baselines like Hi-GCN follow a subject-graph paradigm (where nodes represent brain regions). This allows HF-BrainGNN to focus on learning region-adaptive features and hierarchical topology within the individual brain network.
All models were evaluated on the ADNI dataset (218 subjects) under identical experimental configurations to ensure fair and comparable performance evaluation, with 5-fold cross-validation repeated five times and the final results reported as the mean ± standard deviation across all repetitions. Traditional machine learning and DNN methods adopted 8100-dimensional vectorized connectivity features as input, while all GNN-based models used brain graphs constructed from the AAL90 atlas, where edge weights were calculated via Pearson correlation and only the top 20% of the strongest connections were retained for graph topology construction. For all deep learning models, including DNN and various GNN variants, the Adam optimizer with a fixed learning rate of 0.001 was uniformly employed, and a consistent early stopping strategy was applied to mitigate overfitting. Moreover, hyperparameters for all compared models (e.g., regularization coefficients, hidden layer dimensions, top-k pooling ratios) were optimized through a unified grid search on the validation set of each cross-validation fold, ensuring that the hyperparameter tuning criteria and implementation approaches were consistent across all methods for rigorous and unbiased comparison.
Quantitative results are reported in Table 3, including accuracy (Acc), sensitivity (Sen), specificity (Spe), area under the curve (AUC), and F1-score. As shown in the table, traditional handcrafted methods perform poorly, with accuracies below 75% and specificity mostly under 65%. This indicates that vector-based approaches lose topological structure information, making them prone to false positives and misclassification. The DNN achieves slightly better accuracy (76.84%), benefiting from nonlinear feature learning but still suffering from the loss of spatial adjacency information inherent to vectorization.
In contrast, all GNN-based models outperform traditional and general deep learning approaches, confirming the importance of preserving brain network structure. Among these, Hi-GCN achieves 82.14% accuracy—the best among existing baselines—while other GNN variants (GAT 78.52%, BrainGNN 79.81%, PopulationGCN 80.23%) follow closely behind.
From Table 3, the proposed HF-BrainGNN exhibits the highest overall performance: Accuracy = 86.78 % , Sensitivity = 85.20 % , Specificity = 87.30 % , and AUC = 0.867 . The sensitivity improvement is particularly notable, demonstrating superior capability to correctly identify EMCI patients. High sensitivity is essential in clinical diagnosis to minimize false negatives, ensuring that patients with early mild cognitive impairment are detected reliably.
Compared with Hi-GCN, HF-BrainGNN improves accuracy by 4.64% and specificity by 15.82%, while achieving a 21.01% higher specificity than SVM (the best traditional method). Furthermore, a paired t-test evaluated across the cross-validation folds confirms that the overall performance improvement of HF-BrainGNN over the best-performing baseline (Hi-GCN) is statistically significant ( p < 0.05 ). These improvements are mainly attributed to: 1. The region-adaptive convolution (FAR-Conv), capturing distinct connectivity patterns across different brain areas; 2. The disease-oriented differential focus pooling (DF-Pool), emphasizing discriminative regions related to EMCI; 3. The hierarchical feature integration mechanism, which leverages complementary multi-level information.
Overall, HF-BrainGNN consistently outperforms both conventional and existing GNN-based baselines, achieving higher diagnostic sensitivity and reducing false positives—an essential property for reliable clinical applications.

3.5. Interpretability of HF-BrainGNN

3.5.1. Biomarker Detection

To verify the neurobiological significance of the key brain regions identified by HF-BrainGNN, we analyzed the salient ROIs automatically selected by the differential focus pooling (DF-Pool) layer and compared them with previously reported EMCI biomarkers. The average node focus scores of twenty-three salient brain regions were ranked across all EMCI subjects in descending order. Specifically, these scores represent the mean importance values derived from the test subjects across all 5-fold cross-validation runs. And the top ten regions were then selected as core biomarkers for in-depth analysis. Their differential focus scores ranged from 0.52 to 0.88 (Table 4), and the variation of regional importance is shown in Figure 5b.
The ten most discriminative brain regions identified by HF-BrainGNN (Figure 5a) include the left and right hippocampus (HIP.L, HIP.R), left precuneus (PCUN.L), right posterior cingulate gyrus (PCG.R), left angular gyrus (ANG.L), left middle temporal gyrus (MTG.L), right parahippocampal gyrus (PHG.R), left medial superior frontal gyrus (SFGmed.L), right anterior cingulate gyrus (ACG.R), and left inferior parietal lobule (IPL.L). These regions show a function-oriented distribution pattern, with the hippocampus exhibiting the strongest memory-related characteristics.
To examine the stability of DF-Pool across folds, we extracted the top-five nodes per fold in a five-fold cross-validation scheme and constructed a binary “ROI × Fold” selection matrix ( 1 = selected , 0 = not selected ). As shown in Figure 6, regions such as the hippocampus (37 HIP.L, 38 HIP.R) were repeatedly chosen across folds, demonstrating high reproducibility and stable identification of EMCI-related biomarkers. It is worth noting that while HIP.L achieved the highest global differential score (Table 4), local data distributions within individual cross-validation folds cause minor fluctuations in the fold-specific top-5 rankings. This behavior is expected and accurately reflects the natural intersubject variability in functional connectivity.
Most of the top regions are functionally linked to early EMCI pathology. In the hippocampal memory system, the bilateral hippocampus and right parahippocampal gyrus serve as core components for memory encoding and consolidation; their functional disconnection from the posterior cingulate gyrus, as reported by Sorg et al. [49], correlates with memory decline. HF-BrainGNN accurately captured these regions, validating its ability to identify memory-related impairments. Within the Default Mode Network (DMN), the precuneus and posterior cingulate gyrus act as integrative hubs for self-referential processing. Greicius et al. [50] demonstrated that DMN abnormalities distinguish healthy elders from AD patients; our results reveal that such disruptions can already be detected at the EMCI stage. In the semantic network, the left angular gyrus and left middle temporal gyrus are involved in semantic memory storage and retrieval, and their reduced connectivity strength may explain early semantic deficits in EMCI patients.
Overall, these findings show strong agreement with previous EMCI studies, confirming the neurobiological relevance of the identified regions. Reduced hippocampal connectivity [49], DMN hub abnormalities [51], and weakened angular gyrus connectivity [23] align closely with our model outcomes. Additionally, the right parahippocampal gyrus, left medial superior frontal gyrus, right anterior cingulate gyrus, and left inferior parietal lobule identified by HF-BrainGNN have all been reported as important biomarkers in prior EMCI studies [52,53,54,55]. This high consistency validates the scientific rigor and physiological plausibility of our approach, providing reliable imaging biomarkers for early EMCI diagnosis and prognosis evaluation.
Furthermore, to investigate the clinical validity and interpretability of the learned focus scores, we performed a Pearson correlation analysis between the subject-level focus scores (pooled across the top-five regions) and the clinically established MMSE scores. As illustrated in Figure 7, a strong, statistically significant negative correlation was observed ( r = 0.63 , p < 0.001 ). This confirms that higher focus scores assigned by HF-BrainGNN effectively correspond to lower cognitive function (i.e., lower MMSE scores). Ultimately, this robust alignment between our model-derived metrics and standard clinical cognitive assessments underscores the clinical relevance and interpretability of the proposed framework.

3.5.2. Functional Connectivity Analysis of Core Brain Regions

To further validate the pathological significance of the core brain regions identified by HF-BrainGNN, we examined the differential patterns of twelve key functional connections constructed from the ten core biomarker regions between the CN and EMCI groups. As illustrated in Figure 8, these twelve connections can be categorized as follows: intra-connections within the hippocampal memory system (HIP.L–HIP.R, HIP.L–PHG.R); hippocampus–Default Mode Network (DMN) links (HIP.L–PCUN.L, HIP.R–PCG.R); intra-connections of the DMN (PCUN.L–PCG.R, PCUN.L–ANG.L, PCG.R–SFGmed.L, ACG.R–PCUN.L); semantic cognition network connections (ANG.L–MTG.L, MTG.L–PHG.R); and executive control links (SFGmed.L–ACG.R, IPL.L–ANG.L).
Statistical analyses revealed that the FC strength of all twelve pairs exhibited varying degrees of reduction in the EMCI group. The significance of these differences was evaluated using independent-sample t-tests with FDR correction for multiple comparisons (q < 0.05). Among these, connections in the hippocampal memory system showed the most pronounced impairment: the bilateral hippocampal link (HIP.L–HIP.R) was significantly weakened in EMCI subjects ( CN : 0.20 ± 0.15 vs . EMCI : 0.12 ± 0.18 , p < 0.001 ), and the right hippocampus–posterior cingulate gyrus connection (HIP.R–PCG.R) also declined markedly ( CN : 0.25 ± 0.12 vs . EMCI : 0.15 ± 0.14 , p < 0.001 ). Intra-DMN connections were similarly reduced—for example, the precuneus–posterior cingulate gyrus link (PCUN.L–PCG.R) decreased significantly in EMCI ( CN : 0.35 ± 0.10 vs . EMCI : 0.25 ± 0.12 , p < 0.01 ).
In contrast, semantic and executive control connections appeared relatively preserved. Several pairs, such as HIP.L–PCUN.L, PCUN.L–ANG.L, MTG.L–PHG.R, and SFGmed.L–ACG.R, did not show significant group differences, suggesting partial resistance to early-stage impairment or the engagement of compensatory mechanisms in EMCI.
This distinct pattern of connectivity reduction highlights the vulnerability of hippocampal and DMN networks and supports the clinical relevance of HF-BrainGNN-identified regions. Collectively, these results provide compelling network-level evidence for understanding the neuropathological mechanisms underlying early mild cognitive impairment.
To further quantify the magnitude of these differences, we calculated Cohen’s d as a measure of effect size for the twelve identified connections, with values ranging from 0.32 to 0.91. Specifically, the connection between the bilateral hippocampus (HIP.L-HIP.R) showed a medium effect size ( d = 0.49 ), while the precuneus-posterior cingulate gyrus connection (PCUN.L-PCG.R) exhibited a large effect size ( d = 0.91 ). These results indicate that the identified FC reductions are not only statistically significant but also represent substantial biological alterations in the EMCI brain.

4. Discussion

The proposed hierarchical focused brain graph neural network (HF-BrainGNN) demonstrates clear advantages in classifying early mild cognitive impairment (EMCI) versus the control normal (CN), achieving an accuracy of 86.78%, substantially surpassing existing graph-based and conventional machine-learning methods. These improvements arise from the synergistic effects of its three key components.
First, the FAR-Conv layer introduces region-specific kernels by integrating functional similarity and anatomical proximity, effectively addressing the limitation that traditional GNNs neglect regional functional heterogeneity. Ablation experiments confirm that FAR-Conv consistently outperforms standard graph convolution across parameter configurations, validating the benefit of region-adaptive modeling. This mechanism accords with the neuroscientific principle of functional specialization, offering a biologically plausible representation framework for brain connectivity analysis.
Second, the DF-Pool layer performs disease-oriented feature selection by comparing inter-group focus scores to automatically identify EMCI-relevant brain regions. The ten salient biomarkers discovered, covering the hippocampal memory system, the default mode network, and the semantic network, align closely with previous EMCI studies and exhibit negative correlations with MMSE scores ( r = 0.63 , p < 0.001 ). This result furnishes an interpretable neurobiological basis for model predictions.
Beyond biomarker identification, the hierarchical integration classifier (HIC) adaptively fuses multi-level representations through learnable weights, capturing multi-scale functional interactions within the brain network. Together with a multi-objective loss combining classification accuracy, focal separation, and consistency regularization, HF-BrainGNN achieves an optimal balance between performance and interpretability, strengthening its potential for clinical translation.
Furthermore, the results of this study are dependent on the AAL90 parcellation scheme used to define the graph nodes. While AAL90 is a widely accepted anatomical reference, it may introduce parcellation bias, as anatomical boundaries do not always align perfectly with functionally specialized areas. Future research should evaluate the robustness of HF-BrainGNN across diverse brain templates, such as functional or higher-resolution multi-scale atlases, to mitigate the impact of atlas selection on diagnostic performance.
Nevertheless, several directions remain open for future investigation. While HF-BrainGNN demonstrated superior performance on the ADNI dataset, its generalizability across different imaging protocols and diverse populations has not yet been fully established. Expanding the dataset to multi-center cohorts and cross-scanner acquisitions will help evaluate model generalization and ensure the robustness of the identified biomarkers. Furthermore, extending HF-BrainGNN to other neuropsychiatric conditions, such as Parkinson’s disease or depression, may broaden its utility for early diagnostic and prognostic imaging biomarker discovery.

5. Conclusions

This work presents HF-BrainGNN, a hierarchical graph neural network for EMCI diagnosis based on rs-fMRI brain networks, which incorporates novel designs, including FAR-Conv, DF-Pool, and HIC, to address the functional heterogeneity, group-level feature selection, and multi-level feature utilization problems in existing methods. Experiments on the ADNI dataset demonstrate that HF-BrainGNN achieves outstanding performance, with a classification accuracy of 86.78%, sensitivity of 85.2%, and specificity of 87.3%. The identified biomarkers are consistent with existing neuroscience findings, ensuring both algorithmic performance and interpretability. This study provides a new technical framework for neuroimaging-based disease diagnosis and an interpretable AI tool for early EMCI diagnosis, with important implications for AD early intervention. Future work will focus on multi-center dataset validation, MCI progression analysis, and cross-disease application of the HF-BrainGNN framework to advance brain imaging precision medicine.

Author Contributions

Conceptualization, Z.C. and M.S.; methodology, Z.C.; software, Z.C. and N.W.; validation, Z.C. and N.W.; formal analysis, Z.C.; investigation, Z.C. and N.W.; resources, M.S.; data curation, Z.C. and N.W.; writing—original draft preparation, Z.C.; writing—review and editing, Z.C., N.W., and M.S.; visualization, Z.C.; supervision, M.S.; project administration, M.S. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Ethical review and approval were waived for this study, as it involved the secondary analysis of publicly available, de-identified data from the Alzheimer’s Disease Neuroimaging Initiative (ADNI) database. The original ADNI study was conducted in accordance with the Declaration of Helsinki and was approved by the Institutional Review Boards of all participating institutions. Written informed consent was obtained from all participants before protocol-specific procedures were performed.

Informed Consent Statement

Informed consent was obtained from all subjects involved in the ADNI study at the time of their original enrollment. Since the current study performed a secondary analysis of de-identified data obtained from the ADNI database, additional informed consent for this specific analysis was not required. All ADNI participants provided written informed consent for their data to be used in future research studies.

Data Availability Statement

The data used in this study were obtained from the Alzheimer’s Disease Neuroimaging Initiative (ADNI) database (adni.loni.usc.edu, (accessed on 6 March 2025)). The ADNI data are publicly available to qualified researchers upon application and approval through the ADNI website. The authors do not have the authority to share the raw data directly due to the data use agreement with ADNI. All codes developed for the HF-BrainGNN framework and the pre-processed data indices used in this study are available from the corresponding author upon reasonable request.

Acknowledgments

The authors would like to thank the anonymous reviewers for their constructive comments and insightful suggestions, which have significantly improved the quality of this manuscript. Data collection and sharing for this project was funded by the Alzheimer’s Disease Neuroimaging Initiative (ADNI) (National Institutes of Health Grant U01 AG024904) and the DOD ADNI (Department of Defense award number W81XWH-12-2-0012). Data used in preparation of this article were obtained from the ADNI database (adni.loni.usc.edu, (accessed on 6 March 2025)). As such, the investigators within the ADNI contributed to the design and implementation of ADNI and/or provided data but did not participate in the analysis or writing of this report. A complete listing of ADNI investigators can be found at: http://adni.loni.usc.edu/wp-content/uploads/how_to_apply/ADNI_Acknowledgement_List.pdf, (accessed on 6 March 2025).

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. World Health Organization. A Blueprint for Dementia Research; World Health Organization: Geneva, Switzerland, 2022. [Google Scholar]
  2. Petersen, R.C.; Roberts, R.O.; Knopman, D.S.; Boeve, B.F.; Geda, Y.E.; Ivnik, R.J.; Smith, G.E.; Jack, C.R. Mild cognitive impairment: Ten years later. Arch. Neurol. 2009, 66, 1447–1455. [Google Scholar] [CrossRef] [PubMed]
  3. Livingston, G.; Huntley, J.; Liu, K.Y.; Costafreda, S.G.; Selbæk, G.; Alladi, S.; Ames, D.; Banerjee, S.; Burns, A.; Brayne, C.; et al. Dementia prevention, intervention, and care: 2024 report of the Lancet standing Commission. Lancet 2024, 404, 572–628. [Google Scholar] [CrossRef] [PubMed]
  4. Jack, C.R., Jr.; Andrews, J.S.; Beach, T.G.; Buracchio, T.; Dunn, B.; Graf, A.; Hansson, O.; Ho, C.; Jagust, W.; McDade, E.; et al. Revised criteria for diagnosis and staging of Alzheimer’s disease: Alzheimer’s Association Workgroup. Alzheimer’s Dement. 2024, 20, 5143–5169. [Google Scholar] [CrossRef]
  5. Sun, L.; Zhao, T.; Liang, X.; Xia, M.; Li, Q.; Liao, X.; Gong, G.; Wang, Q.; Pang, C.; Yu, Q.; et al. Human lifespan changes in the brain’s functional connectome. Nat. Neurosci. 2025, 28, 891–901. [Google Scholar] [CrossRef]
  6. de Vos, F.; Koini, M.; Schouten, T.M.; Seiler, S.; van der Grond, J.; Lechner, A.; Schmidt, R.; de Rooij, M.; Rombouts, S.A. A comprehensive analysis of resting state fMRI measures to classify individual patients with Alzheimer’s disease. Neuroimage 2018, 167, 62–72. [Google Scholar] [CrossRef]
  7. Du, Y.; Fang, S.; He, X.; Calhoun, V.D. A survey of brain functional network extraction methods using fMRI data. Trends Neurosci. 2024, 47, 608–621. [Google Scholar] [CrossRef]
  8. Wang, J.; Zuo, X.; He, Y. Graph-based network analysis of resting-state functional MRI. Front. Syst. Neurosci. 2010, 4, 1419. [Google Scholar] [CrossRef]
  9. Hallquist, M.N.; Hillary, F.G. Graph theory approaches to functional network organization in brain disorders: A critique for a brave new small-world. Netw. Neurosci. 2018, 3, 1–26. [Google Scholar] [CrossRef]
  10. Zhang, T.; Liao, Q.; Zhang, D.; Zhang, C.; Yan, J.; Ngetich, R.; Zhang, J.; Jin, Z.; Li, L. Predicting MCI to AD conversation using integrated sMRI and rs-fMRI: Machine learning and graph theory approach. Front. Aging Neurosci. 2021, 13, 688926. [Google Scholar] [CrossRef]
  11. Addison, P.S. The Illustrated Wavelet Transform Handbook: Introductory Theory and Applications in Science, Engineering, Medicine and Finance; CRC Press: Boca Raton, FL, USA, 2017. [Google Scholar]
  12. Tijms, B.M.; Wink, A.M.; De Haan, W.; van der Flier, W.M.; Stam, C.J.; Scheltens, P.; Barkhof, F. Alzheimer’s disease: Connecting findings from graph theoretical studies of brain networks. Neurobiol. Aging 2013, 34, 2023–2036. [Google Scholar] [CrossRef] [PubMed]
  13. Farahani, F.V.; Karwowski, W.; Lighthall, N.R. Application of graph theory for identifying connectivity patterns in human brain networks: A systematic review. Front. Neurosci. 2019, 13, 585. [Google Scholar] [CrossRef] [PubMed]
  14. Wu, Z.; Pan, S.; Chen, F.; Long, G.; Zhang, C.; Yu, P.S. A comprehensive survey on graph neural networks. IEEE Trans. Neural Netw. Learn. Syst. 2020, 32, 4–24. [Google Scholar] [CrossRef]
  15. Parisot, S.; Ktena, S.I.; Ferrante, E.; Lee, M.; Guerrero, R.; Glocker, B.; Rueckert, D. Disease prediction using graph convolutional networks: Application to autism spectrum disorder and Alzheimer’s disease. Med. Image Anal. 2018, 48, 117–130. [Google Scholar] [CrossRef]
  16. Li, X.; Zhou, Y.; Dvornek, N.; Zhang, M.; Gao, S.; Zhuang, J.; Scheinost, D.; Staib, L.H.; Ventola, P.; Duncan, J.S. ngnn: Interpretable brain graph neural network for fmri analysis. Med. Image Anal. 2021, 74, 102233. [Google Scholar] [CrossRef]
  17. Shehzad, A.; Zhang, D.; Yu, S.; Abid, S.; Xia, F. Dynamic Graph Transformer for Brain Disorder Diagnosis. IEEE J. Biomed. Health Inform. 2025, 29, 4388–4400. [Google Scholar] [CrossRef]
  18. Bi, X.a.; Zhou, W.; Li, L.; Xing, Z. Detecting risk gene and pathogenic brain region in EMCI using a novel GERF algorithm based on brain imaging and genetic data. IEEE J. Biomed. Health Inform. 2021, 25, 3019–3028. [Google Scholar] [CrossRef]
  19. Long, Z.; Huang, J.; Li, B.; Li, Z.; Li, Z.; Chen, H.; Jing, B. A comparative atlas-based recognition of mild cognitive impairment with voxel-based morphometry. Front. Neurosci. 2018, 12, 916. [Google Scholar] [CrossRef] [PubMed]
  20. Gallo, S.; El-Gazzar, A.; Zhutovsky, P.; Thomas, R.M.; Javaheripour, N.; Li, M.; Bartova, L.; Bathula, D.; Dannlowski, U.; Davey, C.; et al. Functional connectivity signatures of major depressive disorder: Machine learning analysis of two multicenter neuroimaging studies. Mol. Psychiatry 2023, 28, 3013–3022. [Google Scholar] [CrossRef]
  21. Bressler, S.L.; Menon, V. Large-scale brain networks in cognition: Emerging methods and principles. Trends Cogn. Sci. 2010, 14, 277–290. [Google Scholar] [CrossRef]
  22. Wang, Y.; Sun, Y.; Liu, Z.; Sarma, S.E.; Bronstein, M.M.; Solomon, J.M. Dynamic graph cnn for learning on point clouds. ACM Trans. Graph. (TOG) 2019, 38, 1–12. [Google Scholar] [CrossRef]
  23. Seeley, W.W.; Crawford, R.K.; Zhou, J.; Miller, B.L.; Greicius, M.D. Neurodegenerative diseases target large-scale human brain networks. Neuron 2009, 62, 42–52. [Google Scholar] [CrossRef] [PubMed]
  24. Ying, Z.; You, J.; Morris, C.; Ren, X.; Hamilton, W.; Leskovec, J. Hierarchical graph representation learning with differentiable pooling. Adv. Neural Inf. Process. Syst. 2018, 31, 4800–4810. [Google Scholar] [CrossRef]
  25. Gao, H.; Ji, S. Graph u-nets. In Proceedings of the International Conference on Machine Learning, PMLR, Long Beach, CA, USA, 9–15 June 2019; pp. 2083–2092. [Google Scholar]
  26. Abbas, M.; Matta, J.; Le, T.; Bensmail, H.; Obafemi-Ajayi, T.; Honavar, V.; EL-Manzalawy, Y. Biomarker discovery in inflammatory bowel diseases using network-based feature selection. PLoS ONE 2019, 14, e0225382. [Google Scholar] [CrossRef]
  27. Wang, Z.; Xu, Y.; Peng, D.; Gao, J.; Lu, F. Brain functional activity-based classification of autism spectrum disorder using an attention-based graph neural network combined with gene expression. Cereb. Cortex 2023, 33, 6407–6419. [Google Scholar] [CrossRef] [PubMed]
  28. Xu, Y.; Wang, J.; Guang, M.; Yan, C.; Jiang, C. Multistructure graph classification method with attention-based pooling. IEEE Trans. Comput. Soc. Syst. 2022, 10, 602–613. [Google Scholar] [CrossRef]
  29. Ranjan, E.; Sanyal, S.; Talukdar, P. Asap: Adaptive structure aware pooling for learning hierarchical graph representations. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; Volume 34, pp. 5470–5477. [Google Scholar]
  30. Bianchi, F.M.; Grattarola, D.; Alippi, C. Spectral clustering with graph neural networks for graph pooling. In Proceedings of the International Conference on Machine Learning, PMLR, Virtual, 13–18 July 2020; pp. 874–883. [Google Scholar]
  31. Chu, Y.; Wang, G.; Cao, L.; Qiao, L.; Liu, M. Multi-scale graph representation learning for autism identification with functional MRI. Front. Neuroinformatics 2022, 15, 802305. [Google Scholar] [CrossRef]
  32. Ji, J.; Yao, Y. A novel CNN framework to extract multi-level modular features for the classification of brain networks. Appl. Intell. 2022, 52, 6835–6852. [Google Scholar] [CrossRef]
  33. Zhang, J.; Guo, Y.; Zhou, L.; Wang, L.; Wu, W.; Shen, D. Constructing hierarchical attentive functional brain networks for early AD diagnosis. Med. Image Anal. 2024, 94, 103137. [Google Scholar] [CrossRef]
  34. Goodfellow, I.; Bengio, Y.; Courville, A. Deep Learning; MIT Press: Cambridge, MA, USA, 2016. [Google Scholar]
  35. Shaheen, H.; Melnik, R. Brain network dynamics and multiscale modelling of neurodegenerative disorders: A Review. IEEE Access 2025, 13, 33074–33100. [Google Scholar] [CrossRef]
  36. Peitz, S.; Hotegni, S.S. Multi-objective deep learning: Taxonomy and survey of the state of the art. Mach. Learn. Appl. 2025, 21, 100700. [Google Scholar] [CrossRef]
  37. Wang, J.; Wang, X.; Xia, M.; Liao, X.; Evans, A.; He, Y. GRETNA: A graph theoretical network analysis toolbox for imaging connectomics. Front. Hum. Neurosci. 2015, 9, 386. [Google Scholar] [PubMed]
  38. Murphy, K.; Birn, R.M.; Handwerker, D.A.; Jones, T.B.; Bandettini, P.A. The impact of global signal regression on resting state correlations: Are anti-correlated networks introduced? Neuroimage 2009, 44, 893–905. [Google Scholar] [CrossRef] [PubMed]
  39. Friedman, J. The elements of statistical learning: Data mining, inference, and prediction. J. Am. Stat. Assoc. 2009, 99, 567. [Google Scholar]
  40. Chandra, M.A.; Bedi, S. Survey on SVM and their application in image classification. Int. J. Inf. Technol. 2021, 13, 1–11. [Google Scholar] [CrossRef]
  41. Genuer, R.; Poggi, J.M. Random forests. In Random Forests with R; Springer: Berlin/Heidelberg, Germany, 2020; pp. 33–55. [Google Scholar]
  42. Rumelhart, D.E.; Hinton, G.E.; Williams, R.J. Learning representations by back-propagating errors. Nature 1986, 323, 533–536. [Google Scholar] [CrossRef]
  43. Mittal, S. A survey on modeling and improving reliability of DNN algorithms and accelerators. J. Syst. Archit. 2020, 104, 101689. [Google Scholar] [CrossRef]
  44. Kipf, T. Semi-supervised classification with graph convolutional networks. arXiv 2016, arXiv:1609.02907. [Google Scholar]
  45. Veličković, P.; Cucurull, G.; Casanova, A.; Romero, A.; Lio, P.; Bengio, Y. Graph attention networks. arXiv 2017, arXiv:1710.10903. [Google Scholar]
  46. Shang, C.; Liu, Q.; Tong, Q.; Sun, J.; Song, M.; Bi, J. Multi-view spectral graph convolution with consistent edge attention for molecular modeling. Neurocomputing 2021, 445, 12–25. [Google Scholar] [CrossRef]
  47. Parisot, S.; Ktena, S.I.; Ferrante, E.; Lee, M.; Moreno, R.G.; Glocker, B.; Rueckert, D. Spectral graph convolutions for population-based disease prediction. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Quebec City, QC, Canada, 10–14 September 2017; Springer: Berlin/Heidelberg, Germany, 2017; pp. 177–185. [Google Scholar]
  48. Jiang, H.; Cao, P.; Xu, M.; Yang, J.; Zaiane, O. Hi-GCN: A hierarchical graph convolution network for graph embedding learning of brain network and brain disorders prediction. Comput. Biol. Med. 2020, 127, 104096. [Google Scholar] [CrossRef]
  49. Sorg, C.; Riedl, V.; Muhlau, M.; Calhoun, V.D.; Eichele, T.; Läer, L.; Drzezga, A.; Förstl, H.; Kurz, A.; Zimmer, C.; et al. Selective changes of resting-state networks in individuals at risk for Alzheimer’s disease. Proc. Natl. Acad. Sci. USA 2007, 104, 18760–18765. [Google Scholar] [CrossRef]
  50. Greicius, M.D.; Srivastava, G.; Reiss, A.L.; Menon, V. Default-mode network activity distinguishes Alzheimer’s disease from healthy aging: Evidence from functional MRI. Proc. Natl. Acad. Sci. USA 2004, 101, 4637–4642. [Google Scholar] [CrossRef] [PubMed]
  51. Buckner, R.L.; Andrews-Hanna, J.R.; Schacter, D.L. The brain’s default network: Anatomy, function, and relevance to disease. Ann. N. Y. Acad. Sci. 2008, 1124, 1–38. [Google Scholar] [CrossRef]
  52. Dickerson, B.C.; Sperling, R.A. Functional abnormalities of the medial temporal lobe memory system in mild cognitive impairment and Alzheimer’s disease: Insights from functional MRI studies. Neuropsychologia 2008, 46, 1624–1635. [Google Scholar] [CrossRef]
  53. Jones, D.; Machulda, M.; Vemuri, P.; McDade, E.; Zeng, G.; Senjem, M.; Gunter, J.; Przybelski, S.; Avula, R.; Knopman, D.; et al. Age-related changes in the default mode network are more advanced in Alzheimer disease. Neurology 2011, 77, 1524–1531. [Google Scholar] [CrossRef] [PubMed]
  54. Celone, K.A.; Calhoun, V.D.; Dickerson, B.C.; Atri, A.; Chua, E.F.; Miller, S.L.; DePeau, K.; Rentz, D.M.; Selkoe, D.J.; Blacker, D.; et al. Alterations in memory networks in mild cognitive impairment and Alzheimer’s disease: An independent component analysis. J. Neurosci. 2006, 26, 10222–10231. [Google Scholar] [CrossRef] [PubMed]
  55. Wang, L.; Zang, Y.; He, Y.; Liang, M.; Zhang, X.; Tian, L.; Wu, T.; Jiang, T.; Li, K. Changes in hippocampal connectivity in the early stages of Alzheimer’s disease: Evidence from resting state fMRI. Neuroimage 2006, 31, 496–504. [Google Scholar] [CrossRef]
Figure 1. Construction and analysis workflow of the brain functional network.
Figure 1. Construction and analysis workflow of the brain functional network.
Information 17 00461 g001
Figure 2. Hierarchical graph learning framework of HF-BrainGNN and its core functional modules. Arrows in the figures denote data flow paths, colored bars represent regional affinity vectors, and colored nodes represent brain regions with different properties. Layer 1 and Layer 2 implement hierarchical feature extraction via FAR-Conv and DF-Pool for EMCI/CN classification.
Figure 2. Hierarchical graph learning framework of HF-BrainGNN and its core functional modules. Arrows in the figures denote data flow paths, colored bars represent regional affinity vectors, and colored nodes represent brain regions with different properties. Layer 1 and Layer 2 implement hierarchical feature extraction via FAR-Conv and DF-Pool for EMCI/CN classification.
Information 17 00461 g002
Figure 3. Overview of the differential focus pooling (DF-Pool) mechanism. In the figure, red-colored dots represent the brain regions of the HC group, while blue-colored dots represent those of the EMCI group.
Figure 3. Overview of the differential focus pooling (DF-Pool) mechanism. In the figure, red-colored dots represent the brain regions of the HC group, while blue-colored dots represent those of the EMCI group.
Information 17 00461 g003
Figure 4. Comparison between FAR-Conv and Standard-Conv under different loss function coefficients. (a) Set λ 1 to 0 and adjust λ 2 ; (b) Set λ 1 to 0.1 and adjust λ 2 ; (c) Set λ 2 to 0.1 and adjust λ 1 . In each subplot, red bars correspond to Standard-Conv and blue bars to FAR-Conv, with error bars representing the standard deviation across repeated runs. The first two plots show the influence of λ 2 under different fixed values of λ 1 , while the third plot analyzes the variation of λ 1 when λ 2 is fixed at 0.1.
Figure 4. Comparison between FAR-Conv and Standard-Conv under different loss function coefficients. (a) Set λ 1 to 0 and adjust λ 2 ; (b) Set λ 1 to 0.1 and adjust λ 2 ; (c) Set λ 2 to 0.1 and adjust λ 1 . In each subplot, red bars correspond to Standard-Conv and blue bars to FAR-Conv, with error bars representing the standard deviation across repeated runs. The first two plots show the influence of λ 2 under different fixed values of λ 1 , while the third plot analyzes the variation of λ 1 when λ 2 is fixed at 0.1.
Information 17 00461 g004
Figure 5. Key brain regions and their importance scores for EMCI classification identified by HF-BrainGNN. (a) Visualizes the ten core brain regions (marked as blue nodes) on multi-view brain atlases, intuitively presenting their spatial distributions across different brain perspectives. (b) Displays the importance scores (i.e., differential focus scores between CN and EMCI groups) of these brain regions via a heatmap: the color bar on the right indicates the score range (0.50–0.90, darker blue corresponds to higher importance); asterisks in the heatmap denote the statistical significance of inter-group score differences (****: p < 0.0001 , ***: p < 0.001 , **: p < 0.01 , *: p < 0.05 ). All p-values were corrected for multiple comparisons using the Benjamini–Hochberg false discovery rate (FDR) method.
Figure 5. Key brain regions and their importance scores for EMCI classification identified by HF-BrainGNN. (a) Visualizes the ten core brain regions (marked as blue nodes) on multi-view brain atlases, intuitively presenting their spatial distributions across different brain perspectives. (b) Displays the importance scores (i.e., differential focus scores between CN and EMCI groups) of these brain regions via a heatmap: the color bar on the right indicates the score range (0.50–0.90, darker blue corresponds to higher importance); asterisks in the heatmap denote the statistical significance of inter-group score differences (****: p < 0.0001 , ***: p < 0.001 , **: p < 0.01 , *: p < 0.05 ). All p-values were corrected for multiple comparisons using the Benjamini–Hochberg false discovery rate (FDR) method.
Information 17 00461 g005
Figure 6. Top-5 selection matrix across 5 folds. Each row is an ROI (AAL-ID and abbreviation), each column a CV fold; yellow = selected (1), purple = not selected (0). While the DF-Pool layer mathematically retains N / 2 nodes during network forwarding, this matrix visualizes only the absolute top-5 nodes per fold to clearly illustrate the stability of the most extreme discriminative biomarkers.
Figure 6. Top-5 selection matrix across 5 folds. Each row is an ROI (AAL-ID and abbreviation), each column a CV fold; yellow = selected (1), purple = not selected (0). While the DF-Pool layer mathematically retains N / 2 nodes during network forwarding, this matrix visualizes only the absolute top-5 nodes per fold to clearly illustrate the stability of the most extreme discriminative biomarkers.
Information 17 00461 g006
Figure 7. Statistical Validation and Clinical Interpretability of Model Salience. The scatter plot demonstrates a significant negative Pearson correlation between the model-derived normalized focus scores (top-5 pooled) and clinical MMSE scores, with the shaded area representing the 95% confidence band.
Figure 7. Statistical Validation and Clinical Interpretability of Model Salience. The scatter plot demonstrates a significant negative Pearson correlation between the model-derived normalized focus scores (top-5 pooled) and clinical MMSE scores, with the shaded area representing the 95% confidence band.
Information 17 00461 g007
Figure 8. Boxplot illustrating the relationship among different brain regions, with statistical significance evaluated using the independent-sample t-test. Cohen’s d values for these connections are reported in Section 3.5.2.
Figure 8. Boxplot illustrating the relationship among different brain regions, with statistical significance evaluated using the independent-sample t-test. Cohen’s d values for these connections are reported in Section 3.5.2.
Information 17 00461 g008
Table 1. Notations and their descriptions.
Table 1. Notations and their descriptions.
NotationDescription
NNumber of brain regions (nodes)
v i i-th node (brain region) in the graph
N ( l ) ( i ) Neighbor set of node v i at the l-th layer
AAdjacency matrix, A R N × N
X Node feature matrix
x i ( l ) Feature representation of node v i at the l-th layer
H ( l ) , h i ( l ) Hidden feature representation output by FAR-Conv at the l-th layer
W i ( l ) Graph convolution kernel of node v i at the l-th layer
W j ( l ) Graph convolution kernel of neighbors of node v i
α i Regional affinity vector of node v i
e i j Edge weight enhancement factor based on regional affinity
n K Set of brain regions in the K-th functional network
pos ( v i ) 3D coordinates of brain region v i
C K Centroid coordinates of network K
s ( l ) Node focus score at the l-th layer
s ˜ ( l ) Normalized node focus score
λ 1 , λ 2 Hyperparameters of the loss function
Table 2. Demographic and clinical characteristics of study participants.
Table 2. Demographic and clinical characteristics of study participants.
CharacteristicEMCI ( n = 104 )CN ( n = 114 )
Age 74.94 ± 6.11 74.48 ± 6.48
Gender (M/F)56/4860/54
Education (years) 15.19 ± 2.98 15.78 ± 3.02
MMSE Score 24.39 ± 2.48 28.12 ± 1.21
Table 3. Classification results of different methods on the ADNI dataset.
Table 3. Classification results of different methods on the ADNI dataset.
MethodAcc (%)Sen (%)Spe (%)AUCF1 (%)
Ridge Classifier [39] 73.82 ± 1.07 73.18 ± 2.30 65.05 ± 1.42 0.626 ± 0.035 77.59 ± 2.89
SVM [40] 73.95 ± 0.68 80.09 ± 0.86 66.29 ± 2.21 0.675 ± 0.023 82.77 ± 0.82
Random Forest [41] 74.69 ± 1.57 77.48 ± 3.67 63.24 ± 3.07 0.621 ± 0.042 73.17 ± 2.81
DNN [43] 74.61 ± 4.45 78.62 ± 2.75 68.75 ± 3.85 0.719 ± 0.037 83.90 ± 2.30
GCN [44] 76.91 ± 2.58 78.72 ± 3.13 66.95 ± 4.40 0.642 ± 0.024 81.61 ± 1.81
GAT [45] 76.91 ± 1.29 80.04 ± 3.68 67.97 ± 2.15 0.708 ± 0.022 82.96 ± 0.97
BrainGNN [16] 77.66 ± 4.25 81.86 ± 3.04 70.88 ± 3.62 0.636 ± 0.052 82.82 ± 1.96
MVS-GCN [46] 69.28 ± 4.27 72.66 ± 3.21 76.26 ± 2.69 0.602 ± 0.049 71.76 ± 2.64
PopulationGCN [47] 79.90 ± 1.82 79.59 ± 1.29 74.15 ± 3.73 0.523 ± 0.015 86.14 ± 1.09
Hi-GCN [48] 82.14 ± 1.35 83.01 ± 0.86 67.71 ± 4.79 0.791 ± 0.030 86.96 ± 1.07
HF-BrainGNN (Ours) 86.78 ± 1.42 85.20 ± 1.63 87.30 ± 1.53 0.867 ± 0.014 89.13 ± 1.12
Table 4. Ten core brain regions identified by HF-BrainGNN.
Table 4. Ten core brain regions identified by HF-BrainGNN.
RankAAL AbbreviationAAL-IDDifferential ScoreFunctional Network
1HIP.L370.88Hippocampal Memory System
2HIP.R380.84Hippocampal Memory System
3PCUN.L670.79Default Mode Network
4PCG.R360.75Default Mode Network
5ANG.L650.71Semantic Network
6MTG.L850.67Semantic Network
7PHG.R400.63Hippocampal Memory System
8SFGmed.L230.59Executive Control Network
9ACG.R320.55Salience Network
10IPL.L610.52Attention Network
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Chen, Z.; Song, M.; Wu, N. Early Mild Cognitive Impairment Diagnosis via Resting-State fMRI Brain Networks Using a Region-Specific Hierarchical Fusion Graph Neural Network. Information 2026, 17, 461. https://doi.org/10.3390/info17050461

AMA Style

Chen Z, Song M, Wu N. Early Mild Cognitive Impairment Diagnosis via Resting-State fMRI Brain Networks Using a Region-Specific Hierarchical Fusion Graph Neural Network. Information. 2026; 17(5):461. https://doi.org/10.3390/info17050461

Chicago/Turabian Style

Chen, Zhiang, Miao Song, and Ningge Wu. 2026. "Early Mild Cognitive Impairment Diagnosis via Resting-State fMRI Brain Networks Using a Region-Specific Hierarchical Fusion Graph Neural Network" Information 17, no. 5: 461. https://doi.org/10.3390/info17050461

APA Style

Chen, Z., Song, M., & Wu, N. (2026). Early Mild Cognitive Impairment Diagnosis via Resting-State fMRI Brain Networks Using a Region-Specific Hierarchical Fusion Graph Neural Network. Information, 17(5), 461. https://doi.org/10.3390/info17050461

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop