Next Article in Journal
Communication Abilities, Assessment Procedures, and Intervention Approaches in Rett Syndrome: A Narrative Review
Previous Article in Journal
Subtyping Early Parkinson’s Disease by Mapping Cognitive Profiles to Brain Atrophy with Visual MRI Ratings
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

SAGEFusionNet: An Auxiliary Supervised Graph Neural Network for Brain Age Prediction as a Neurodegenerative Biomarker

1
Neural Engineering Lab, Department of Biosciences and Bioengineering, Indian Institute of Technology Guwahati, Guwahati 781039, India
2
Department of Radiology and Imaging, Apollo Hospitals, Guwahati 781005, India
*
Author to whom correspondence should be addressed.
Brain Sci. 2025, 15(7), 752; https://doi.org/10.3390/brainsci15070752
Submission received: 29 May 2025 / Revised: 10 July 2025 / Accepted: 11 July 2025 / Published: 15 July 2025
(This article belongs to the Section Neurorehabilitation)

Abstract

Background: The ability of Graph Neural Networks (GNNs) to analyse brain structural patterns in various kinds of neurodegenerative diseases, including Parkinson’s disease (PD), has drawn a lot of interest recently. One emerging technique in this field is brain age prediction, which estimates biological age to identify ageing patterns that may serve as biomarkers for such disorders. However, a significant problem with most of the GNNs is their depth, which can lead to issues like oversmoothing and diminishing gradients. Methods: In this study, we propose SAGEFusionNet, a GNN architecture specifically designed to enhance brain age prediction and assess PD-related brain ageing patterns using T1-weighted structural MRI (sMRI). SAGEFusionNet learns important ROIs for brain age prediction by incorporating ROI-aware pooling at every layer to overcome the above challenges. Additionally, it incorporates multi-layer feature fusion to capture multi-scale structural information across the network hierarchy and auxiliary supervision to enhance gradient flow and feature learning at multiple depths. The dataset utilised in this study was sourced from the Alzheimer’s Disease Neuroimaging Initiative (ADNI) database. It included a total of 580 T1-weighted sMRI scans from healthy individuals. The brain sMRI scans were parcellated into 56 regions of interest (ROIs) using the LPBA40 brain atlas in CAT12. The anatomical graph was constructed based on grey matter (GM) volume features. This graph served as input to the GNN models, along with GM and white matter (WM) volume as node features. All models were trained using 5-fold cross-validation to predict brain age and subsequently tested for performance evaluation. Results: The proposed framework achieved a mean absolute error (MAE) of 4.24 ± 0.38 years and a mean Pearson’s Correlation Coefficient (PCC) of 0.72 ± 0.03 during cross-validation. We also used 215 PD patient scans from the Parkinson’s Progression Markers Initiative (PPMI) database to assess the model’s performance and validate it. The initial findings revealed that out of 215 individuals with Parkinson’s disease, 213 showed higher and 2 showed lower predicted brain ages than their actual ages, with a mean MAE of 13.36 years (95% confidence interval: 12.51–14.28). Conclusions: These results suggest that brain age prediction using the proposed method may provide important insights into neurodegenerative diseases.

1. Introduction

The structure of the human brain alters significantly over the span of a lifetime, and specific patterns of healthy brain ageing have been noticed [1,2]. These structural changes are not a medical condition but an inevitable consequence of ageing. However, these changes make it more likely for older people to develop neurodegenerative diseases and dementia [3]. Due to neurodegenerative diseases such as Alzheimer’s, Parkinson’s conditions, and schizophrenia, the human brain suffers from an unusually accelerated brain ageing process [4,5,6]. These disorders indicate an altered and sometimes accelerated brain ageing pattern in people suffering from them, as they deviate from the normal patterns of healthy brain ageing. In addition to providing information about the normal ageing process, these deviations commonly referred to as the brain age gap refer to the difference between an individual’s chronological age and brain age, which has considerable promise as a useful biomarker for detecting and investigating neurodegenerative disease, cognitive decline, and other age-related conditions [7,8,9]. The chronological age is the age of a person in years since birth, also known as actual age. Brain age, on the other hand, is determined by considering the structural and functional characteristics of the brain, which is also known as biological age [7,8]. The accurate estimation of brain age from neuroimaging data has emerged as a vital area of research. Although brain age estimation has been extensively explored in healthy populations using T1-weighted structural MRI, its application in Parkinson’s disease (PD) is still relatively limited [7,10]. In Parkinson’s disease (PD), networks that control motor, cognitive, and emotional activities are disrupted by widespread cortical and subcortical atrophy, especially in the frontal cortex, thalamus, brainstem, and basal ganglia [11,12]. According to the results of recent research on brain age in Parkinson’s disease, patients often have altered brain age compared to healthy controls [13,14].
Structural magnetic resonance imaging (sMRI) is one of the several neuroimaging modalities that offer comprehensive anatomical details, making it an effective tool for investigating the structure of the brain [15,16] and how it relates to age. Classical techniques for predicting brain age frequently use machine learning algorithms using features derived from sMRI data [17,18,19]. Despite their good performance, these methods are unable to capture the intricate pairwise relational information between various brain regions. This issue can be addressed by Graph Neural Networks (GNNs), which operate directly on graph-structured data. The human brain can be naturally depicted as a graph, representing a complex network of interconnected brain areas. In a study, different brain regions are represented as nodes, and the physical connections between regions are represented as edges, forming a graph [20]. In recent years, the analysis of graphs using machine learning has garnered significant attention among researchers due to the remarkable expressive capabilities offered by graphs.
Graph Neural Networks (GNNs) are deep learning-based methods that extend existing deep learning techniques to non-Euclidean data by leveraging the graph structure of data to simplify learning tasks effectively [21]. GNNs have shown significant promise in processing graph-structured data by leveraging the underlying relationships within graphs [22]. Their strong performance across various tasks has made them increasingly popular in various domains. There are some studies where GNNs have been successfully applied to sMRI data, including age and gender prediction using GCNs on the cortical surface area [23], and the examination of performance disparities in brain age prediction using multi-volumetric sMRI features [24]. Moreover, semi-supervised node prediction has been performed using population graphs [25], and in a comparative examination of various message-passing and pooling methods, ROI-aware pooling showed significant improvements [26]. Most GNNs employ a sequential neighbourhood aggregation approach to learn a node’s representation vector, often referred to as node embedding, during the training process. This approach involves utilising a first-order graph filtering operation to aggregate representation vectors from a node’s immediate one-hop neighbours. The term “first-order” signifies that this operation considers only direct connections or edges between nodes, usually without considering higher-order connections or paths. These aggregated vectors are then used to compute the node’s representation vector for each iteration, forming a GNN layer [27]. Many common variants of Graph Neural Networks (GNNs) have been developed based on different choices of sequential neighbourhood aggregation strategies. These variants include Graph Convolutional Networks (GCNs) [22], Graph Sample and Aggregated Layers (GraphSAGEs) [28], Graph Attention Networks (GATs) [29], and Graph Isomorphism Networks (GINs) [30]. Each of these approaches employs distinct methods to aggregate features from a node’s neighbours in the graph. These techniques play a crucial role in determining the trajectory of the learning process and the expressiveness of the final GNN model.
Through cascading multiple GNN layers, a variety of tasks, such as node classification [22,31], edge classification [32], graph classification [33], link prediction [34], and various complex problems including those found in social networks [35], have been performed. This sequential cascading, along with non-linear activations, enables Graph Neural Networks (GNNs) to aggregate features beyond one-hop neighbours, leading to enhanced generalisation. For instance, when two such layers are cascaded along with non-linearity, information from two-hops-away neighbours is aggregated. Similarly, cascading three, four, or more layers aggregates features from three-, four-, or more-hops-away neighbours, respectively, thus improving model performance. However, due to successive aggregations, the node embeddings converge towards a non-informative state [36]. When it comes to mean aggregation, node embeddings become increasingly similar and eventually indistinguishable; this phenomenon is called oversmoothing [36,37]. This issue, along with others like the vanishing gradient problem, makes sequential aggregation-based GNN models susceptible to performance degradation due to the cascading of multiple layers and the network growing deeper [22,37,38]. Geometric deep learning research on oversmoothing is quite active, and it would be outside the purview of this article to provide a comprehensive list of works. Research has mostly concentrated on new architectures to alleviate it, like randomly deleting connections and residual techniques [39,40], using regularisations [41], as well as ingenious normalisation [42,43]. These challenges pose significant obstacles to designing deeper GNN models and result in the loss of important information available beyond one-hop neighbours. The motivation of this study is to develop a GNN-based framework that enables the design of deeper models without encountering these issues.
The aim of this study is to develop a novel and enhanced GNN-based framework for brain age prediction. This integrates the expressive capabilities of GNNs at the ROI level with T1-weighted structural MRI (sMRI) data. The suggested approach attempts to represent the brain as a graph in order to capture its rich topological structure. The techniques used in our model, such as node-level pooling [26,44] and multi-depth supervision [45], have already been investigated in other studies. The novel aspect of our work is the strategic integration of these techniques within the proposed SAGEFusionNet architecture, which combines them in a unified framework. Our model incorporates ROI-aware pooling at each layer for identifying important ROIs in brain age prediction, multi-layer feature fusion to capture multi-scale structural information, and auxiliary supervision to enhance feature learning across different network depths. This integration enables the model to enhance its performance by mitigating oversmoothing, enhancing gradient flow, and improving representational diversity, as supported by the comparative results and Dirichlet energy analysis. Additionally, in order to confirm the clinical applicability of our approach, we tested the proposed model on Parkinson’s disease (PD) datasets. While some of these techniques have been used individually in prior studies, to the best of our knowledge, this particular integration strategy and its application to brain age prediction in Parkinson’s Disease using GNNs have not been previously reported. This manuscript is divided into the following sections and subsections: Section 2 provides an outline of the proposed approach, including information about the dataset, preprocessing, construction of the anatomical network, model description, training, and testing. The results and the discussion are presented in Section 3 and Section 4. The conclusion of the proposed study is finally described in Section 5.

2. Materials and Methods

2.1. Dataset Description

The dataset utilised in this study was sourced from the Alzheimer’s Disease Neuroimaging Initiative (ADNI) database (https://adni.loni.usc.edu/, accessed on 11 January 2024), a large public repository that offers neuroimaging, neuropsychological, clinical, and genetic data aimed at tracking the development of Alzheimer’s disease dementia. The ADNI datasets are multisite datasets, containing data obtained from multiple MRI scanners and locations across various research sites. Our study incorporated a total of 580 T1-weighted MRI scans obtained from healthy subjects, comprising 233 male and 347 female scans. The subjects included in our study were carefully selected to ensure age and gender matching, with ages ranging between 51 and 95 years. The average age of male subjects was 74.99 ± 7.08 years, while that of female subjects was 72.70 ± 6.94 years. As per the ADNI Data Manual, these subjects were chosen based on the following criteria: no significant memory complaints beyond age-related norms; normal memory function confirmed by Logical Memory II scores meeting education-specific cutoffs (≥9 for 16+ years, ≥5 for 8–15 years, ≥3 for 0–7 years of education); Mini-Mental State Exam score between 24 and 30; Clinical Dementia Rating of 0 with a Memory Box score of 0; and overall cognitive normality with no notable impairments in cognition or daily functioning. Another 215 Parkinson’s disease (PD) patient scans were used in this study, obtained from the Parkinson’s Progression Markers Initiative (PPMI) database (www.ppmi-info.org/access-data-specimens/download-data, accessed on 30 July 2023) to assess brain ageing patterns related to PD. Among these, 129 were male and 88 were female, with an average age of 58.47 ± 7.76 years. As per the PPMI Clinical Protocol Manual, these PD patients were chosen based on the following criteria: if they were at least 30 years old, had a clinical diagnosis of PD, and displayed at least two motor symptoms (such as stiffness, bradykinesia, or resting tremor), with either bradykinesia or resting tremor being necessary. Prior to receiving a DaTscanTM injection, participants had to meet UPSIT requirements, give informed consent, pause specific drugs before SPECT imaging, and, if they were female, test negative for pregnancy.

2.2. Preprocessing

The raw T1-weighted structural sMRI images obtained from the Alzheimer’s Disease Neuroimaging Initiative (ADNI) were initially in Digital Imaging and Communications in Medicine (DICOM) format. These images were converted to Neuroimaging Informatics Technology Initiative (NIfTI) format using the dcm2nii tool from MRIcron (Version:v1.0.2019092) [46]. Subsequently, these scans underwent preprocessing with the Computational Anatomy Toolbox (CAT12, Version: vCAT12.8.2) [47], within Statistical Parametric Mapping (SPM12) (https://www.fil.ion.ucl.ac.uk/spm/software/spm12/, accessed on 8 August 2023), using MATLAB 2022a [48]. Major preprocessing steps are depicted in Figure 1; the NIfTI scans were denoised using a spatial-adaptive Non-Local Means (SANLM) denoising filter, which enhances image quality by removing noise while preserving edges. Bias field correction was then performed to mitigate intensity inhomogeneities by minimising the effects of the bias field. Following this, the scans were segmented using tissue probability maps, dividing them into three primary tissues: grey matter (GM), white matter (WM), and cerebrospinal fluid (CSF). The segmented scans were then spatially normalised to Montreal Neurological Institute (MNI152) standard space. Subsequent to normalisation, the images were further processed using the LONI Probabilistic Brain Atlas (LPBA40) for parcellation, which divided the brain into 56 regions of interest (ROIs) [49]. Each ROI represents a specific anatomical area within the brain. The regional grey matter G M MN and white matter W M MN volumes for each subject were extracted, where G M MN denotes the grey matter volume and W M MN denotes the white matter volume of the N th ROI for the M th subject. These volumes of each ROI were normalised by the subject’s intracranial volume to account for individual head size differences. Subsequently, the mean and standard deviation of each normalised ROI volume were computed across subjects and used to standardise the corresponding ROI volumes, resulting in features with zero mean and unit variance. Therefore, for M subjects, each having N ROIs, we obtained subject vs. ROI volume matrices X GM and X WM , as shown in Equations (1) and (2), each with dimensions M × N . Only the X GM volume feature was used for further construction of the anatomical network.
Brainsci 15 00752 i001
Brainsci 15 00752 i002

2.3. Construction of Anatomical Network

A graph is a mathematical structure consisting of a set of nodes ( V ) and edges ( E ) , represented as G = ( V , E ) . Nodes denote specific entities or points, while edges indicate the relationships or connections between them. An edge ( u , v ) in graph G illustrates a connection between node u V and node v V . In this study, we constructed an anatomical network using a correlation-based method with LASSO optimisation, a graph learning technique derived from the literature of graph signal processing (GSP) [50,51]. The network’s nodes represent various brain regions of interest (ROIs), and the edges show the associations of regional grey matter G M MN volume among these ROIs obtained from structural MRI (sMRI) data [26]. The specific dependencies depicted by the edges can vary based on the graph construction method and the inherent constraints of the optimisation problem. An anatomical network was constructed from data of M subjects, each having N brain ROI volumes, by applying LASSO optimisation [26,51].
For an arbitrary node i, the edge weights w i from the node i to all other nodes m = 1 , 2 , , N 1 , m i , can be derived by minimising the optimisation function J i . In matrix form, it is given by
J i = | | y i Y i w i | | 2 2 + λ | | w i | | 1
where w i = [ w i 2 , w i 3 , w i 4 , , w i m ] T , i = 1 , 2 , 3 , , N (ROIs), and m = 1 , 2 , , N 1 , m i . y i is the i-th column of X G M , which represents the i-th ROI across all the subjects, Y i is the X G M matrix after removing the i-th column, and λ is the sparsity parameter. The first term encourages correlation among connected nodes, whereas the second term encourages sparsity. A sparser graph results from a larger value of λ , and its optimal value λ = 0.06 was selected based on multiple experiments given in Table 1.
The optimisation problem mentioned above needs to be solved individually for every node. This process produces a set of edge weights or coefficients, denoted as w i , for each node i, resulting in N sets of weights corresponding to N nodes. In order to maintain dimensional consistency, a zero is padded at the i-th position of w i since the i-th ROI is eliminated during regression. These sets of weights are then concatenated together to form the matrix W, which contains all the edge weights for the entire network. However, this process does not guarantee that W will be a symmetric matrix. An adjacency matrix, by definition, is symmetric. To confirm symmetry, the matrix W is multiplied by its transpose W T , and then the square root is taken to obtain the adjacency matrix depicted in Equation (4). The size of the obtained adjacency matrix (A) will be N × N .
A N × N = W · W T
Using the aforementioned technique, an unweighted anatomical graph is constructed, where N represents the number of nodes in the network depicted in Figure 2. The nodes of these graphs are initialised with regional G M MN and W M MN volumes features instead of structural or location-based attributes [52].

2.4. Model Description

The GNN model proposed in this study consists of multiple layers working together to generate the final output depicted in Figure 3. The anatomical graph generated from sMRI using Equations (3) and (4) was utilised as the input for the proposed architecture. Every node in this graph has two features representing the regional grey matter G M MN and white matter W M MN volumes of a specific brain region extracted from sMRI. These features are encapsulated within the feature matrix X R N × F , where N symbolises the total brain ROIs for each subject and F represents the number of features associated with each node or ROI. The initial layer comprises the Graph Sample and Aggregation (GraphSAGE) layer, which is a fundamental element of the proposed model introduced by William L. Hamilton [28]. GraphSAGE is a flexible inductive framework designed for generating node embeddings by sampling and aggregating features from a node’s local neighbourhood. Instead of considering the entire neighbourhood, GraphSAGE uniformly samples a fixed-size subset of neighbours for aggregation. The function A G G ( · ) represents the aggregation function, and GraphSAGE proposes three types of aggregators: mean aggregator, LSTM aggregator, and pooling aggregator. In the proposed model, we used a mean aggregator in GraphSAGE that can be considered an inductive version of Graph Convolutional Networks (GCNs). In contrast, the LSTM aggregator is not permutation-invariant, as it relies on a specified order of nodes for aggregation. We selected the GraphSAGE layer for our approach because it offers superior inductive performance and scalability.
Additionally, as shown in Table 2, the GraphSAGE-based model performed better in our study than other standard GNN architectures under the same experimental setting. The main goal of the GraphSAGE layer is to combine node features with graph structural features to update node representations. This involves aggregating features from neighbouring nodes along with the node’s own features. If there are r GraphSAGE layers [28], then the node update equation for node v at layer r is given by Equation (5):
x v ( r ) = σ W r AGG x u ( r 1 ) , u N ( v ) , x v ( r 1 )
In Equation (5), W r represents a learnable weight matrix, x v ( r ) denotes the updated features of node v at the r-th layer, σ signifies the non-linear activation, specifically Leaky ReLU, N ( v ) denotes the neighbours of node v, and x u ( r 1 ) denotes the node embedding of neighbouring node u at the ( r 1 ) -th layer.
The output of the first GraphSAGE layer was passed through a non-linear activation function called Leaky ReLU. The purpose of using Leaky ReLU was to enable the learning of more complex relationships and patterns in the data and to overcome the issue of vanishing gradients. We standardised the feature to have a zero mean and unit variance. As a result, some of the features became negative, but these negative values are still considered by the Leaky ReLU activation function; this is an additional advantage of using Leaky ReLU, unlike the standard ReLU activation function, which discards all negative values and outputs only positive values. The output of this Leaky ReLU was the aggregated embeddings of one-hop away neighbours, which were passed to the next GraphSAGE layer and ROI-aware pooling layer.

ROI-Aware Pooling

It is a multi-head weighted pooling mechanism where each head learns a soft ROI attention over nodes and aggregates node embeddings into a graph-level representation [26]. Consider the node feature matrix X R N × F , where F is the feature dimension and N is the number of nodes. W R N × H is a learnable weight matrix, where H denotes the number of attention heads. The attention score vector for a single head across all nodes is represented by each column of W. A softmax function is applied to produce a normalised attention score vector:
A = softmax ( W ) R N × H
The graph-level embedding for each attention head h is given by z h :
z h = i = 1 N A i h · x i
Lastly, the outputs from each attention head are concatenated to create the overall graph-level representation at layer l:
Z ( l ) = [ z 1 z 2 z H ] R H · F
The output of the first ROI-aware global pooling is given as Z ( 1 ) . The proposed model consists of four cascaded GraphSAGE layers along with Leaky ReLU activations. Similarly, the second, third, and fourth GraphSAGE layers’ outputs passed through subsequent ROI-aware global pooling, which produces graph-level embeddings Z ( 2 ) , Z ( 3 ) , and Z ( 4 ) containing aggregated features from the second-, third-, and fourth-hops-away neighbouring nodes. The graph-level outputs from each layer are
outputs = Z ( 1 ) , Z ( 2 ) , , Z ( 4 )
These ROI-aware pooling output embeddings Z ( 1 ) , Z ( 2 ) , …, Z ( r ) after each GraphSAGE layer are passed to the corresponding auxiliary regression head, which produces the auxiliary prediction at each layer, respectively: y ^ aux ( 1 ) , y ^ aux ( 2 ) , and so on, until y ^ aux ( r ) , which are shown in Figure 3. Using these auxiliary predictions, the auxiliary losses are calculated at each layer. After processing all layers, the model fuses the outputs from each ROI-aware global pooling together. A variety of approaches, including mean, max, sum, weighted sum, attention-based mechanisms, and concatenation, can be used to accomplish this fusion, and their performance is presented in Table 3. In the case of concatenation, it can be given by
Z fused = Z ( 1 ) Z ( 2 ) Z ( r )
Subsequently, this fused embedding was passed to the regressor layer to produce the predicted value of brain age. The regressor layer is a multi-layer perceptron (MLP) comprising a fully connected layer (FC), a Leaky ReLU layer, and another FC layer with a single neuron responsible for producing the predicted brain age. Since our proposed model incorporates GraphSAGE layers and a fusion mechanism, we refer to it as SAGEFusionNet. All the steps of the proposed model are summarised in Algorithm 1. As part of this study, baseline models including FCNN, GCN, GraphSAGE, GIN, and GAT were incorporated for comparison purposes.
Algorithm 1 SAGEFusionNet model algorithm
  • Required inputs: Graph G = ( V , E ) ( edge _ index ), initial node features X R N × F , number of GraphSAGE layers r, dropout rate ρ , regularisation hyperparameter α
  • Output: Auxiliary predicted value at i-th layer y ^ aux ( i ) , final predicted value y ^
  • Initialise auxiliary regression head Reghead aux ( i ) for prediction at i-th layer
  • Initialise lists: pooled _ outputs [   ] , aux _ losses [   ]
  • for  i = 1 to r do
  •      X LeakyReLU ( SAGEConv i ( X , edge _ index ) )
  •      X Dropout ( X , p = ρ )
  •      X pooled ROIAwareGP ( X , batch )
  •     Append X pooled to pooled _ outputs
  •      y ^ aux ( i ) Reghead aux ( i ) ( X pooled )
  •      L aux i MSE ( y ^ aux ( i ) , y )
  •     Append L aux i to aux _ losses
  • end for
  •   Z fused Concat ( pooled _ outputs )
  • y ^ Regressor ( Z fused )
  • L aux 1 r i = 1 r L aux i
  • L hybrid 1 n i = 1 n ( y ^ i y i ) 2 + α · L aux
  • return  y ^

2.5. Training and Testing

In our study, a careful approach was taken to train the model effectively. All models in this study were trained using Google Colab, (https://colab.research.google.com/, accessed on 1 May 2025) which offers a virtual machine (VM) environment equipped with 12.67 GB of RAM and 107.72 GB of available disk space. The experiments were conducted using Python 3.11, leveraging the PyTorch framework and the PyTorch Geometric library for implementing Graph Neural Network (GNN) models. Initially, the dataset was split into two parts: one for training the models and the other for testing their performance, with an 80:20 ratio, respectively. To ensure the reliability and generalisation of the models, a 5-fold cross-validation technique was employed, involving repeated training and validation with different data combinations. The models were trained using mini-batches of 16 samples to enhance computational efficiency. An adaptive learning rate strategy was implemented, starting with an initial learning rate of 0.01 to adjust the learning rate during training dynamically. To prevent overfitting, a dropout rate of 0.15 was applied. A hybrid loss function given in Equation (10) was employed during training to incorporate the auxiliary loss. Several values of the hyperparameter α were evaluated, and our model achieved optimal performance at α = 0.30 . To further optimise the learning process, an Adam optimiser was employed to accelerate convergence and stabilise the training dynamics.
  • Loss Function
The hybrid loss function used in our study is a combination of two loss functions, the Mean Squared Error (MSE) loss and an auxiliary loss function, to ensure robustness in capturing differences between predicted brain age and actual age. It is given by the following Equation (10):
L hybrid = 1 n i = 1 n ( y ^ i y i ) 2 L main + α · 1 r l = 1 r MSE ( y ^ aux ( l ) , y ) L aux
where n represents the number of subjects, y i denotes the actual age of the i-th subject, and y ^ i refers to the predicted brain age for the i-th subject. The variable r denotes the number of GraphSAGE layers in the model, α is a regularisation hyperparameter, and y ^ aux ( l ) denotes the predicted value at the l th layer. The importance of the auxiliary loss in the hybrid loss function is controlled by the hyperparameter α . The main loss supervises the final prediction derived from fused ROI-aware representations across all GNN layers, while each layer receives intermediate supervision from the auxiliary loss.

Performance Metrics

To evaluate the efficacy of the models, the mean absolute error (MAE) and the Pearson Correlation Coefficient (PCC) were employed as metrics to assess how well the predicted age of the brain aligns with the actual age of the subjects. Additionally, the Dirichlet energy given in Equation (11) was used to analyse the oversmoothing of the models:
  • Dirichlet energy:
    It is one of the metrics to measure the oversmoothing in deep GNN on graph-structured data [36,53]. The normalised version of the Dirichlet energy at the l th GNN layer is given in the following Equation (11):
    E ( x l ) = i V j N i x i l 1 + d i x j l 1 + d j 2 2
    where V is the set of all nodes and N i is the set of neighbours of node i. The terms x i l and x j l represent the features of nodes i and j at the l th layer, respectively. The terms d i and d j denote the degrees of nodes i and j, respectively. Finally, · 2 2 denotes the squared 2 -norm.

3. Results

The performance of the proposed SAGEFusionNet model with different numbers of GraphSAGE layers is shown in Table 4. The Pearson Correlation Coefficient (PCC) and mean absolute error (MAE) were computed when the number of layers varied from two to six. Four layers achieved the most promising results, with an MAE of 4.24 ± 0.38 and a PCC of 0.72 ± 0.03 . This finding shows that while deeper configurations (five and six layers) begin to experience a reduction in performance, moderate depth promotes optimal learning. Table 3 reports the impact of various feature fusion strategies on model performance. A number of techniques were investigated to aggregate the ROI-aware pooled representations across GNN layers, including mean, max, sum, weighted sum, attention, and concatenation. The concatenation approach performed better than the others. This finding suggests that concatenation, rather than averaging or summing the embeddings into a single vector, better aids prediction by maintaining the distinctiveness of each layer’s representation. Figure 4 shows the variance of Dirichlet energy across different depths of GNN layers for all other models in order to assess the oversmoothing behaviour across GNN models. Evidence of similar node embeddings can be seen in models like GCN, GAT, GIN, and GraphSAGE, where Dirichlet energy sharply declines with depth. On the other hand, SAGEFusionNet continuously sustains greater Dirichlet energy levels, indicating a more robust feature diversity preservation. This provides empirical evidence that our model uses ROI-aware pooling and auxiliary supervision at each layer to reduce oversmoothing.
The proposed framework is compared with different baselines, including FCNN, GCN, GraphSAGE, GAT, and GIN, in Table 2. SAGEFusionNet produced the most promising results out of all the other approaches, with an MAE of 4.24 ± 0.38 and PCC of 0.72 ± 0.03 . The significance of structural modelling can be seen by the FCNN model, which performed the worst, with MAE = 6.16 ± 0.99 , PCC = 0.54 ± 0.07 . SAGEFusionNet specifically solves the oversmoothing and less efficient gradient propagation in deeper layers that hinder most of the GNN models, despite their reasonable performance.
A spider plot of the MAE, PCC, and RMSE metrics over five cross-validation folds is displayed within Figure 5. SAGEFusionNet’s excellent generalisation capacity is demonstrated by the consistent performance across all folds. The model’s stability and absence of overfitting on particular data partitions are highlighted by this consistent behaviour. A scatter plot of the predicted versus actual brain age values for the holdout test set is shown in Figure 6. High correlation with the ground truth is confirmed by the predictions’ near alignment with the diagonal identity line. With an MAE of 4.17 and a PCC of 0.79, the test results demonstrate the model’s strong generalisation to unseen holdout datasets and its ability to accurately capture the structural patterns of the brain, corresponding to brain age. To confirm its efficacy, we further validated the proposed SAGEFusionNet model on an external dataset comprising 215 PD patient scans from the Parkinson’s Progression Markers Initiative (PPMI) database. According to preliminary results, predicted brain ages were higher than actual ages in 213 out of 215 individuals with Parkinson’s disease and lower in only 2, with a mean absolute error (MAE) of 13.36 years (95% CI: 12.51–14.28), indicating stable predictive performance. These results appear to align with previous investigations that reported that people with Parkinson’s disease (PD) experience accelerated brain ageing [5].

4. Discussion

The efficacy of the proposed SAGEFusionNet model in brain age prediction tasks is clearly demonstrated by the experimental findings. Numerous modifications to GNN architecture led to enhanced brain age prediction. The incorporation of ROI-aware global pooling at every GraphSAGE layer performs a very important function. The ROI-aware pooling focuses more on important brain ROIs for brain age prediction, in contrast to traditional global pooling. Second, extra supervision at each GraphSAGE layer also helps to reduce two basic issues with deep GNNs, oversmoothing and vanishing gradients, while accelerating convergence. As we go deeper, SAGEFusionNet sustains considerably higher Dirichlet energy levels throughout layers than baseline models, which is verified by the theoretical analysis and practical observations in Figure 4. This suggests improved discriminative power and feature variance preservation at the deeper layer.
Third, to combine embeddings from multiple GraphSAGE layers, the fusion of multi-layer ROI-aware embeddings was investigated, employing a variety of approaches, such as mean, max, sum, attention, and concatenation. Table 3 demonstrates how this fusion process captures discriminative meaningful features from various network depths. It is found that models that utilised attention or concatenation fusion approaches performed better than those that employed a simple aggregation approach. These results highlight the importance of maintaining multi-scale representations to improve generalisation and reduce the effect of oversmoothing in deep GNNs. Figure 5 provides consistent findings throughout all five folds, and Figure 6 shows a strong correlation between predicted and actual brain ages, further confirming the method’s efficacy and robustness. Lastly, we tested the model using the Parkinson’s disease (PD) dataset and found that the predicted brain ages for most patients were greater than their actual ages. Our initial findings show that the MAE for healthy controls was 4.24 ± 0.38 years, while for PD subjects it was 13.36 years. This reveals that the brain age gap in PD subjects is larger relative to their chronological age compared to healthy controls. This finding is clinically relevant because it supports the various theories that neurodegenerative diseases cause accelerated brain ageing. Structural changes associated with Parkinson’s disease are effectively captured by our model by encoding pairwise relationships between ROIs. The ability to track changes in brain structure implies that brain age estimation frameworks such as SAGEFusionNet could offer a new way to monitor the course of illness and evaluate the effectiveness of therapy. Finally, the combination of ROI-aware pooling, auxiliary regression heads, and multi-layer fusion produces a model that outperforms baseline GNNs in terms of performance, learning efficiency and stability. Together, these elements allow the model to move beyond some of the inherent drawbacks of deep GNNs. The current study shows that SAGEFusionNet can achieve excellent brain age prediction performance using only single-modality structural MRI data, reducing complexity. To the best of our knowledge, SAGEFusionNet is the first framework that encodes both global and local structural brain features acquired from grey matter and white matter volumes by utilising layer-wise ROI-aware pooling, auxiliary supervision, and multi-layer feature fusion.
Pina et al. [26] used ROI-aware pooling after the last GNN layer to learn important ROIs and transform node-level embeddings into a graph-level representation. While this method is effective, its output solely depends on the representation of the last GNN layer. Hierarchical graph pooling with structure learning [54] performs top-k node pooling at each layer, which preserves a subset of informative nodes and forms a smaller induced subgraph, while structure learning subsequently reconstructs the graph structure for this pooled subgraph. DiffPool [44] performs hierarchical graph pooling by learning a soft cluster assignment from node embeddings to coarsen the graph structure across layers, where the coarsened graph serves as input to the next GNN layer. Numerous studies have looked into feature fusion from various dimensions. For example, JK-Nets [55] fuses node-level embeddings from each GNN layer using concatenation or max-pooling, demonstrating the benefits of combining representations from multiple depths. DeepGCNs [45] incorporate residual/dense connections to pass earlier layer outputs into deeper layers while simultaneously fusing features from various depths. GCNII [40] advances this idea through initial residual connections and identity mapping, performing implicit fusion across node embeddings within each layer. However, these methods do not perform explicit graph-level feature fusion across depths. In contrast, our method applies ROI-aware pooling at each GraphSAGE layer to learn important ROIs for age prediction without altering the original graph topology. The resulting graph-level representations from different network depths are fused to capture multi-scale structural information across the network hierarchy. Additionally, this framework is supported by auxiliary supervision, enabling more expressive and hierarchical representation learning, making the model more task-adaptive.
The latest studies in brain network modelling have shown the benefits of combining multimodal data, including MRI, DTI, fMRI, and EEG, using graph-based learning methods like GNNs and Transformers. These methods have facilitated enhancement in both predictive accuracy and interpretability by capturing the association between structural and functional connectivity [56,57,58,59]. Furthermore, edge-centric models such as Edge-Boosted Graph Learning [60] provide a new perspective to conventional node-based techniques through constructing edge functional connectivity (eFC) from time series data and employing a co-embedding technique to jointly learn from node and edge features. While our current work focuses solely on structural MRI, we recognise the potential of expanding SAGEFusionNet to integrate additional data types, such as DTI for white matter tractography or fMRI for capturing functional coupling. SAGEFusionNet does not employ energy-based models to simulate brain state transitions; it uses Dirichlet energy as an analytical metric to quantify oversmoothing across network layers. This usage is conceptually aligned with energy-based frameworks where, for example, the brain selectively reallocates network energy across tasks and networks [61], energy landscape analysis has been used to characterise brain state dynamics and transitions in Alzheimer’s disease [62], and lattice field theory has been applied to describe neural activity in terms of potential and kinetic energy [63]. Similarly, physics-inspired GNNs have been used to solve combinatorial optimisation problems [64] and to estimate the haemodynamic variables of the human brain [65], underscoring the value of integrating physical priors into modelling. These models offer rigorous, biophysically grounded formulations. Integrating them could enhance interpretability and robustness. However, the age range of the training dataset (51 to 95 years) is a constraint of the current work that may limit the model’s potential to be applied to larger populations. In particular, SAGEFusionNet might underestimate brain age for people over 95 and overestimate it for people under 51. To improve the model’s resilience and suitability for a greater variety of clinical situations, future research will concentrate on validating it on bigger, more age-diverse cohorts. Additionally, including environmental and lifestyle factors through real-world testbeds such as towns, neighbourhoods, and clinics to consider their impact on brain ageing [66] may enhance the model’s ability to capture the wide contributors of brain ageing and improve generalisability.

5. Conclusions

In this work, we introduced a novel Graph Neural Network framework called SAGEFusionNet that uses T1-weighted structural MRI data to predict brain age. Key issues with deep GNNs, such as oversmoothing and vanishing gradients, are addressed in the proposed model by combining ROI-aware global pooling, layer-wise auxiliary supervision, and multi-layer feature fusion. Additionally, based on clinical evaluation of PD patient data from the PPMI dataset, SAGEFusionNet consistently predicts brain ages that are greater than chronological ages, indicating accelerated ageing trends in PD patients. This demonstrates the therapeutic significance of the model and its potential as a non-invasive biomarker for neurodegenerative illness. This approach may offer a novel way to assess treatment effectiveness by comparing predicted brain age to actual age and monitoring changes in the brain age gap following medication. It might also improve knowledge of disease progression and how effective treatments are. This approach for brain age prediction has the potential to be an effective instrument for neurodegenerative disease diagnosis and monitoring in medical contexts, improving patient care and treatment results. The absence of a control-matched group is one of the limitations in our study. Confounding effects could be introduced since the healthy controls utilised for training were not specifically matched to PD individuals on age, scanner type, or acquisition methodology. Furthermore, the significance of the observed brain age gap in the PD group was not evaluated using any formal statistical comparisons (e.g., effect size or p-value). These results should therefore be regarded as initial and exploratory in nature. To address this, future research will use statistical inference to robustly assess the significance and generalisability of the observed effects, as well as age- and scanner-matched controls.

Author Contributions

C.N.G. proposed this idea to S.K. The entire framework was developed by S.K. The paper was written by S.K., C.N.G., and S.H. All authors contributed during revisions. This study is one of S.K.’s Ph.D. objectives. All authors have read and agreed to the published version of the manuscript.

Funding

S.K. received funding through a doctoral scholarship awarded by the Ministry of Education (MoE), Government of India. C.N.G.’s time was funded by the Department of Science and Technology (DST), Government of India (Project Code: DST/INT/SWD/VR/P-14/2019), and he is associated with the Indian Institute of Technology, Guwahati, Assam, India (cngupta@iitg.ac.in).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data used in this study are publicly available from the Alzheimer’s Disease Neuroimaging Initiative (ADNI) and the Parkinson’s Progression Markers Initiative (PPMI) databases. Researchers can apply to access ADNI data at (https://adni.loni.usc.edu/, accessed on 11 January 2024) and PPMI data at (www.ppmi-info.org/access-data-specimens/download-data, accessed on 30 July 2023), RRID:SCR_006431. Access is subject to data use agreements. The code and datasets used in this study are available at: (https://github.com/NeuralLabIITGuwahati/SAGEFusionNet, accessed on 7 July 2025).

Acknowledgments

Data used in preparation of this article were obtained from the Alzheimer’s Disease Neuroimaging Initiative(ADNI) database (adni.loni.usc.edu). As such, the investigators within the ADNI contributed to the design and implementation of ADNI and/or provided data but did not participate in analysis or writing of this report. A complete listing of ADNI investigators can be found at: http://adni.loni.usc.edu/wp-content/uploads/how_to_apply/ADNI_Acknowledgement_List.pdf, accessed on 7 July 2025. Data collection and sharing for this project was funded by the Alzheimer’s Disease Neuroimaging Initiative (ADNI) (National Institutes of Health Grant U01 AG024904) and DOD ADNI (Department of Defense award number W81XWH-12-2-0012). The ADNI is funded by the National Institute on Aging, the National Institute of Biomedical Imaging and Bioengineering, and through generous contributions from the following: AbbVie, Alzheimer’s Association; Alzheimer’s Drug Discovery Foundation; Araclon Biotech; BioClinica, Inc.; Biogen; Bristol-Myers Squibb Company; CereSpir, Inc.; Cogstate; Eisai Inc.; Elan Pharmaceuticals, Inc.; Eli Lilly and Company; EuroImmun; F. Hoffmann-La Roche Ltd and its affiliated company Genentech, Inc.; Fujirebio; GE Healthcare; IXICO Ltd.; Janssen Alzheimer Immunotherapy Research & Development, LLC.; Johnson & Johnson Pharmaceutical Research & Development LLC.; Lumosity; Lundbeck; Merck & Co., Inc.; Meso Scale Diagnostics, LLC.; NeuroRx Research; Neurotrack Technologies; Novartis Pharmaceuticals Corporation; Pfizer Inc.; Piramal Imaging; Servier; Takeda Pharmaceutical Company; and Transition Therapeutics. The Canadian Institutes of Health Research are providing funds to support ADNI clinical sites in Canada. Private sector contributions are facilitated by the Foundation for the National Institutes of Health (www.fnih.org). The grantee organisation is the Northern California Institute for Research and Education, and the study is coordinated by the Alzheimer’s Therapeutic Research Institute at the University of Southern California. ADNI data are disseminated by the Laboratory for Neuro Imaging at the University of Southern California. Data used in the preparation of this article was obtained on [2023-07-30] from the Parkinson’s Progression Markers Initiative (PPMI) database (www.ppmi-info.org/access-data-specimens/download-data, accessed on 30 July 2023), RRID:SCR_006431. For up-to-date information on the study, visit www.ppmi-info.org. PPMI—a public–private partnership—is funded by the Michael J. Fox Foundation for Parkinson’s Research and funding partners, including [https://www.ppmi-info.org/sites/default/files/docs/PPMI%20Funding%20Partners.pdf, accessed on 7 July 2025].

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Fjell, A.M.; Walhovd, K.B. Structural brain changes in aging: Courses, causes and cognitive consequences. Rev. Neurosci. 2010, 21, 187–222. [Google Scholar] [CrossRef] [PubMed]
  2. Lee, J.; Kim, H.J. Normal aging induces changes in the brain and neurodegeneration progress: Review of the structural, biochemical, metabolic, cellular, and molecular changes. Front. Aging Neurosci. 2022, 14, 931536. [Google Scholar] [CrossRef] [PubMed]
  3. Abbott, A. Dementia: A problem for our age. Nature 2011, 475, S2–S4. [Google Scholar] [CrossRef] [PubMed]
  4. Gaser, C.; Franke, K.; Klöppel, S.; Koutsouleris, N.; Sauer, H.; Initiative, A.D.N. BrainAGE in mild cognitive impaired patients: Predicting the conversion to Alzheimer’s disease. PLoS ONE 2013, 8, e67346. [Google Scholar] [CrossRef] [PubMed]
  5. Chen, Y.S.; Kuo, C.Y.; Lu, C.H.; Wang, Y.W.; Chou, K.H.; Lin, W.C. Multiscale brain age prediction reveals region-specific accelerated brain aging in Parkinson’s disease. Neurobiol. Aging 2024, 140, 122–129. [Google Scholar] [CrossRef] [PubMed]
  6. Zhu, J.D.; Wu, Y.F.; Tsai, S.J.; Lin, C.P.; Yang, A.C. Investigating brain aging trajectory deviations in different brain regions of individuals with schizophrenia using multimodal magnetic resonance imaging and brain-age prediction: A multicenter study. Transl. Psychiatry 2023, 13, 82. [Google Scholar] [CrossRef] [PubMed]
  7. Cole, J.H.; Poudel, R.P.; Tsagkrasoulis, D.; Caan, M.W.; Steves, C.; Spector, T.D.; Montana, G. Predicting brain age with deep learning from raw imaging data results in a reliable and heritable biomarker. NeuroImage 2017, 163, 115–124. [Google Scholar] [CrossRef] [PubMed]
  8. Franke, K.; Gaser, C. Ten years of BrainAGE as a neuroimaging biomarker of brain aging: What insights have we gained? Front. Neurol. 2019, 10, 789. [Google Scholar] [CrossRef] [PubMed]
  9. Franke, K.; Gaser, C. Longitudinal changes in individual BrainAGE in healthy aging, mild cognitive impairment, and Alzheimer’s disease. GeroPsych 2012, 25, 235–245. [Google Scholar] [CrossRef]
  10. Franke, K.; Ziegler, G.; Klöppel, S.; Gaser, C.; Alzheimer’s Disease Neuroimaging Initiative. Estimating the age of healthy subjects from T1-weighted MRI scans using kernel methods: Exploring the influence of various parameters. Neuroimage 2010, 50, 883–892. [Google Scholar] [CrossRef] [PubMed]
  11. Kalia, L.V.; Lang, A.E. Parkinson’s disease. Lancet 2015, 386, 896–912. [Google Scholar] [CrossRef] [PubMed]
  12. Zeighami, Y.; Fereshtehnejad, S.M.; Dadar, M.; Collins, D.L.; Postuma, R.B.; Mišić, B.; Dagher, A. A clinical-anatomical signature of Parkinson’s disease identified with partial least squares and magnetic resonance imaging. Neuroimage 2019, 190, 69–78. [Google Scholar] [CrossRef] [PubMed]
  13. Beheshti, I.; Mishra, S.; Sone, D.; Khanna, P.; Matsuda, H. T1-weighted MRI-driven brain age estimation in Alzheimer’s disease and Parkinson’s disease. Aging Dis. 2019, 11, 618. [Google Scholar] [CrossRef] [PubMed]
  14. Eickhoff, C.R.; Hoffstaedter, F.; Caspers, J.; Reetz, K.; Mathys, C.; Dogan, I.; Amunts, K.; Schnitzler, A.; Eickhoff, S.B. Advanced brain ageing in Parkinson’s disease is related to disease duration and individual impairment. Brain Commun. 2021, 3, fcab191. [Google Scholar] [CrossRef] [PubMed]
  15. Ashburner, J.; Friston, K.J. Voxel-based morphometry—The methods. Neuroimage 2000, 11, 805–821. [Google Scholar] [CrossRef] [PubMed]
  16. Mao, C.; Zhang, Y.; Jiang, J.; Qin, R.; Ye, Q.; Zhu, X.; Wu, J. Magnetic Resonance Imaging Biomarkers of Punding in Parkinson’s Disease. Brain Sci. 2023, 13, 1423. [Google Scholar] [CrossRef] [PubMed]
  17. Dafflon, J.; Pinaya, W.H.; Turkheimer, F.; Cole, J.H.; Leech, R.; Harris, M.A.; Cox, S.R.; Whalley, H.C.; McIntosh, A.M.; Hellyer, P.J. An automated machine learning approach to predict brain age from cortical anatomical measures. Hum. Brain Mapp. 2020, 41, 3555–3566. [Google Scholar] [CrossRef] [PubMed]
  18. Pang, Y.; Cai, Y.; Xia, Z.; Gao, X. Predicting brain age using Tri-UNet and various MRI scale features. Sci. Rep. 2024, 14, 13742. [Google Scholar] [CrossRef] [PubMed]
  19. More, S.; Antonopoulos, G.; Hoffstaedter, F.; Caspers, J.; Eickhoff, S.B.; Patil, K.R.; Alzheimer’s Disease Neuroimaging Initiative. Brain-age prediction: A systematic comparison of machine learning workflows. NeuroImage 2023, 270, 119947. [Google Scholar] [CrossRef] [PubMed]
  20. Bullmore, E.; Sporns, O. Complex brain networks: Graph theoretical analysis of structural and functional systems. Nat. Rev. Neurosci. 2009, 10, 186–198. [Google Scholar] [CrossRef] [PubMed]
  21. Bronstein, M.M.; Bruna, J.; Cohen, T.; Veličković, P. Geometric deep learning: Grids, groups, graphs, geodesics, and gauges. arXiv 2021, arXiv:2104.13478. [Google Scholar] [CrossRef]
  22. Kipf, T.N.; Welling, M. Semi-supervised classification with graph convolutional networks. arXiv 2016, arXiv:1609.02907. [Google Scholar]
  23. Besson, P.; Parrish, T.; Katsaggelos, A.K.; Bandt, S.K. Geometric deep learning on brain shape predicts sex and age. Comput. Med. Imaging Graph. 2021, 91, 101939. [Google Scholar] [CrossRef] [PubMed]
  24. Kumar, S.; Gupta, C.N. Multi-Volumetric Feature-Based Brain Age Prediction Using sMRI and Graph Neural Networks. In Proceedings of the 2024 IEEE International Conference on Systems, Man, and Cybernetics (SMC), Kuching, Malaysia, 6–10 October 2024; pp. 3300–3304. [Google Scholar]
  25. Parisot, S.; Ktena, S.I.; Ferrante, E.; Lee, M.; Moreno, R.G.; Glocker, B.; Rueckert, D. Spectral graph convolutions for population-based disease prediction. In Proceedings of the Medical Image Computing and Computer Assisted Intervention-MICCAI 2017: 20th International Conference, Quebec City, QC, Canada, 11–13 September 2017; Proceedings, Part III 20. Springer: Cham, Switzerland, 2017; pp. 177–185. [Google Scholar]
  26. Pina, O.; Cumplido-Mayoral, I.; Cacciaglia, R.; González-de Echávarri, J.M.; Gispert, J.D.; Vilaplana, V. Structural networks for brain age prediction. In Proceedings of the International Conference on Medical Imaging with Deep Learning. PMLR, Zurich, Switzerland, 6–8 July 2022; pp. 944–960. [Google Scholar]
  27. Gama, F.; Isufi, E.; Leus, G.; Ribeiro, A. Graphs, convolutions, and neural networks: From graph filters to graph neural networks. IEEE Signal Process. Mag. 2020, 37, 128–138. [Google Scholar] [CrossRef]
  28. Hamilton, W.; Ying, Z.; Leskovec, J. Inductive representation learning on large graphs. Adv. Neural Inf. Process. Syst. 2017, 30, 1024–1034. [Google Scholar]
  29. Veličković, P.; Cucurull, G.; Casanova, A.; Romero, A.; Lio, P.; Bengio, Y. Graph attention networks. arXiv 2017, arXiv:1710.10903. [Google Scholar]
  30. Xu, K.; Hu, W.; Leskovec, J.; Jegelka, S. How powerful are graph neural networks? arXiv 2018, arXiv:1810.00826. [Google Scholar]
  31. Zhao, T.; Zhang, X.; Wang, S. Graphsmote: Imbalanced node classification on graphs with graph neural networks. In Proceedings of the 14th ACM International Conference on Web Search and Data Mining, Jerusalem, Israel, 8–12 March 2021; pp. 833–841. [Google Scholar]
  32. Sasaki, H.; Fujii, M.; Sakaji, H.; Masuyama, S. Enhancing risk analysis with GNN: Edge classification in risk causality from securities reports. Int. J. Inf. Manag. Data Insights 2024, 4, 100217. [Google Scholar]
  33. Errica, F.; Podda, M.; Bacciu, D.; Micheli, A. A fair comparison of graph neural networks for graph classification. arXiv 2019, arXiv:1912.09893. [Google Scholar]
  34. Zhang, M.; Chen, Y. Link prediction based on graph neural networks. Adv. Neural Inf. Process. Syst. 2018, 31, 5165–5175. [Google Scholar]
  35. Li, X.; Sun, L.; Ling, M.; Peng, Y. A survey of graph neural network based recommendation in social networks. Neurocomputing 2023, 549, 126441. [Google Scholar] [CrossRef]
  36. Rusch, T.K.; Bronstein, M.M.; Mishra, S. A survey on oversmoothing in graph neural networks. arXiv 2023, arXiv:2303.10993. [Google Scholar] [CrossRef]
  37. Li, Q.; Han, Z.; Wu, X.M. Deeper insights into graph convolutional networks for semi-supervised learning. In Proceedings of the AAAI Conference on Artificial Intelligence, New Orleans, LA, USA, 2–7 February 2018; Volume 32. [Google Scholar]
  38. Keriven, N. Not too little, not too much: A theoretical analysis of graph (over) smoothing. Adv. Neural Inf. Process. Syst. 2022, 35, 2268–2281. [Google Scholar]
  39. Huang, W.; Rong, Y.; Xu, T.; Sun, F.; Huang, J. Tackling over-smoothing for general graph convolutional networks. arXiv 2020, arXiv:2008.09864. [Google Scholar]
  40. Chen, M.; Wei, Z.; Huang, Z.; Ding, B.; Li, Y. Simple and deep graph convolutional networks. In Proceedings of the International Conference on Machine Learning, PMLR, Virtual, 13–18 July 2020; pp. 1725–1735. [Google Scholar]
  41. Chen, D.; Lin, Y.; Li, W.; Li, P.; Zhou, J.; Sun, X. Measuring and relieving the over-smoothing problem for graph neural networks from the topological view. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; Volume 34, pp. 3438–3445. [Google Scholar]
  42. Zhao, L.; Akoglu, L. Pairnorm: Tackling oversmoothing in gnns. arXiv 2019, arXiv:1909.12223. [Google Scholar]
  43. Huang, N.; Villar, S.; Priebe, C.E.; Zheng, D.; Huang, C.; Yang, L.; Braverman, V. From local to global: Spectral-inspired graph neural networks. arXiv 2022, arXiv:2209.12054. [Google Scholar] [CrossRef]
  44. Ying, Z.; You, J.; Morris, C.; Ren, X.; Hamilton, W.; Leskovec, J. Hierarchical graph representation learning with differentiable pooling. Adv. Neural Inf. Process. Syst. 2018, 31, 4800–4810. [Google Scholar]
  45. Li, G.; Muller, M.; Thabet, A.; Ghanem, B. DeepGCNs: Can GCNs go as deep as CNNs? In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 9267–9276. [Google Scholar]
  46. Rorden, C. From MRIcro to MRIcron: The Evolution of Neuroimaging Visualization Tools. Neuropsychologia 2025, 207, 109067. [Google Scholar] [CrossRef] [PubMed]
  47. Gaser, C.; Dahnke, R.; Thompson, P.M.; Kurth, F.; Luders, E.; The Alzheimer’s Disease Neuroimaging Initiative. CAT: A computational anatomy toolbox for the analysis of structural MRI data. Gigascience 2024, 13, giae049. [Google Scholar] [CrossRef] [PubMed]
  48. The MathWorks, Inc. MATLAB2022a, Version 9.12; The MathWorks, Inc.: Natick, MA, USA, 2022.
  49. Shattuck, D.W.; Mirza, M.; Adisetiyo, V.; Hojatkashani, C.; Salamon, G.; Narr, K.L.; Poldrack, R.A.; Bilder, R.M.; Toga, A.W. Construction of a 3D probabilistic atlas of human cortical structures. Neuroimage 2008, 39, 1064–1080. [Google Scholar] [CrossRef] [PubMed]
  50. Tibshirani, R. Regression shrinkage and selection via the lasso. J. R. Stat. Soc. Ser. B Stat. Methodol. 1996, 58, 267–288. [Google Scholar] [CrossRef]
  51. Stanković, L.; Sejdić, E. Vertex-Frequency Analysis of Graph Signals; Springer: Cham, Switzerland, 2019. [Google Scholar]
  52. Cui, H.; Dai, W.; Zhu, Y.; Kan, X.; Gu, A.A.C.; Lukemire, J.; Zhan, L.; He, L.; Guo, Y.; Yang, C. Braingb: A benchmark for brain network analysis with graph neural networks. IEEE Trans. Med. Imaging 2022, 42, 493–506. [Google Scholar] [CrossRef] [PubMed]
  53. Cai, C.; Wang, Y. A note on over-smoothing for graph neural networks. arXiv 2020, arXiv:2006.13318. [Google Scholar] [CrossRef]
  54. Zhang, Z.; Bu, J.; Ester, M.; Zhang, J.; Yao, C.; Yu, Z.; Wang, C. Hierarchical graph pooling with structure learning. arXiv 2019, arXiv:1911.05954. [Google Scholar] [CrossRef]
  55. Xu, K.; Li, C.; Tian, Y.; Sonobe, T.; Kawarabayashi, K.i.; Jegelka, S. Representation learning on graphs with jumping knowledge networks. In Proceedings of the International Conference on Machine Learning. PMLR, Stockholm, Sweden, 10–15 July 2018; pp. 5453–5462. [Google Scholar]
  56. Mohammadi, H.; Karwowski, W. Graph Neural Networks in Brain Connectivity Studies: Methods, Challenges, and Future Directions. Brain Sci. 2024, 15, 17. [Google Scholar] [CrossRef] [PubMed]
  57. Wein, S.; Malloni, W.M.; Tomé, A.M.; Frank, S.M.; Henze, G.I.; Wüst, S.; Greenlee, M.W.; Lang, E.W. A graph neural network framework for causal inference in brain networks. Sci. Rep. 2021, 11, 8061. [Google Scholar] [CrossRef] [PubMed]
  58. Cong, S.; Wang, H.; Zhou, Y.; Wang, Z.; Yao, X.; Yang, C. Comprehensive review of Transformer-based models in neuroscience, neurology, and psychiatry. Brain-X 2024, 2, e57. [Google Scholar] [CrossRef]
  59. Levakov, G.; Rosenthal, G.; Shelef, I.; Raviv, T.R.; Avidan, G. From a deep learning model back to the brain—Identifying regional predictors and their relation to aging. Hum. Brain Mapp. 2020, 41, 3235–3252. [Google Scholar] [CrossRef] [PubMed]
  60. Yang, D.; Abdelmegeed, M.; Modl, J.; Kim, M. Edge-boosted graph learning for functional brain connectivity analysis. In Proceedings of the 2025 IEEE 22nd International Symposium on Biomedical Imaging (ISBI), Houston, TX, USA, 14–17 April 2025; pp. 1–4. [Google Scholar]
  61. Saberi, M.; Rieck, J.R.; Golafshan, S.; Grady, C.L.; Misic, B.; Dunkley, B.T.; Khatibi, A. The brain selectively allocates energy to functional brain networks under cognitive control. Sci. Rep. 2024, 14, 32032. [Google Scholar] [CrossRef] [PubMed]
  62. Xing, L.; Guo, Z.; Long, Z. Energy landscape analysis of brain network dynamics in Alzheimer’s disease. Front. Aging Neurosci. 2024, 16, 1375091. [Google Scholar] [CrossRef] [PubMed]
  63. Bardella, G.; Franchini, S.; Pan, L.; Balzan, R.; Ramawat, S.; Brunamonti, E.; Pani, P.; Ferraina, S. Neural activity in quarks language: Lattice Field Theory for a network of real neurons. Entropy 2024, 26, 495. [Google Scholar] [CrossRef] [PubMed]
  64. Schuetz, M.J.; Brubaker, J.K.; Katzgraber, H.G. Combinatorial optimization with physics-inspired graph neural networks. Nat. Mach. Intell. 2022, 4, 367–377. [Google Scholar] [CrossRef]
  65. Sarabian, M.; Babaee, H.; Laksari, K. Physics-informed neural networks for brain hemodynamic predictions using medical imaging. IEEE Trans. Med. Imaging 2022, 41, 2285–2303. [Google Scholar] [CrossRef] [PubMed]
  66. Woods, T.; Palmarini, N.; Corner, L.; Barzilai, N.; Maier, A.B.; Sagner, M.; Bensz, J.; Strygin, A.; Yadala, N.; Kern, C.; et al. Cities, communities and clinics can be testbeds for human exposome and aging research. Nat. Med. 2025, 31, 1066–1068. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Graphical abstract of proposed framework for brain age prediction.
Figure 1. Graphical abstract of proposed framework for brain age prediction.
Brainsci 15 00752 g001
Figure 2. Anatomical brain network constructed using grey matter.
Figure 2. Anatomical brain network constructed using grey matter.
Brainsci 15 00752 g002
Figure 3. Architecture of SAGEFusionNet model.
Figure 3. Architecture of SAGEFusionNet model.
Brainsci 15 00752 g003
Figure 4. Oversmoothing analysis of various GNNs at different depths.
Figure 4. Oversmoothing analysis of various GNNs at different depths.
Brainsci 15 00752 g004
Figure 5. Spider plot of different matrices during 5-fold cross-validation.
Figure 5. Spider plot of different matrices during 5-fold cross-validation.
Brainsci 15 00752 g005
Figure 6. Scatter plot of actual age vs. predicted age.
Figure 6. Scatter plot of actual age vs. predicted age.
Brainsci 15 00752 g006
Table 1. Performance of SAGEFusionNet model on anatomical networks constructed using different values of the sparsity parameter λ . Bold values indicate the best performance.
Table 1. Performance of SAGEFusionNet model on anatomical networks constructed using different values of the sparsity parameter λ . Bold values indicate the best performance.
ModelSparsity ParameterMAEPCC
SAGEFusionNet0.02 4.51 ± 0.37 0.67 ± 0.05
0.03 4.61 ± 0.38 0.66 ± 0.06
0.04 4.44 ± 0.54 0.69 ± 0.05
0.05 4.48 ± 0.39 0.67 ± 0.07
0.06 4 . 24 ± 0 . 38 0 . 72 ± 0 . 03
0.07 4.29 ± 0.38 0.70 ± 0.04
0.08 4.36 ± 0.34 0.68 ± 0.06
Table 2. Performance comparison of proposed and baseline models. Bold values indicate the best performance.
Table 2. Performance comparison of proposed and baseline models. Bold values indicate the best performance.
ModelMAEPCC
FCNN 6.16 ± 0.99 0.54 ± 0.07
GCN 4.76 ± 0.38 0.62 ± 0.06
GraphSAGE 4.70 ± 0.40 0.63 ± 0.07
GAT 4.74 ± 0.34 0.62 ± 0.07
GIN 4.87 ± 0.35 0.59 ± 0.06
SAGEFusionNet 4 . 24 ± 0 . 38 0 . 72 ± 0 . 03
Table 3. Performance comparison of fusion methods. Bold values indicate the best performance.
Table 3. Performance comparison of fusion methods. Bold values indicate the best performance.
ModelFusion MethodMeanPCC
SAGEFusionNetMean 4.60 ± 0.45 0.65 ± 0.07
Max 4.52 ± 0.51 0.66 ± 0.06
Sum 4.60 ± 0.44 0.64 ± 0.07
Weighted Sum 4.73 ± 0.41 0.63 ± 0.07
Attention 4.29 ± 0.45 0.70 ± 0.05
Concatenation 4 . 24 ± 0 . 38 0 . 72 ± 0 . 03
Table 4. Performance of SAGEFusionNet with varying numbers of layers. Bold values indicate the best performance.
Table 4. Performance of SAGEFusionNet with varying numbers of layers. Bold values indicate the best performance.
ModelLayersMAEPCC
SAGEFusionNet2 4.34 ± 0.42 0.70 ± 0.05
3 4.44 ± 0.56 0.68 ± 0.06
4 4 . 24 ± 0 . 38 0 . 72 ± 0 . 03
5 4.40 ± 0.26 0.69 ± 0.04
6 4.56 ± 0.39 0.65 ± 0.06
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Kumar, S.; Hazarika, S.; Gupta, C.N. SAGEFusionNet: An Auxiliary Supervised Graph Neural Network for Brain Age Prediction as a Neurodegenerative Biomarker. Brain Sci. 2025, 15, 752. https://doi.org/10.3390/brainsci15070752

AMA Style

Kumar S, Hazarika S, Gupta CN. SAGEFusionNet: An Auxiliary Supervised Graph Neural Network for Brain Age Prediction as a Neurodegenerative Biomarker. Brain Sciences. 2025; 15(7):752. https://doi.org/10.3390/brainsci15070752

Chicago/Turabian Style

Kumar, Suraj, Suman Hazarika, and Cota Navin Gupta. 2025. "SAGEFusionNet: An Auxiliary Supervised Graph Neural Network for Brain Age Prediction as a Neurodegenerative Biomarker" Brain Sciences 15, no. 7: 752. https://doi.org/10.3390/brainsci15070752

APA Style

Kumar, S., Hazarika, S., & Gupta, C. N. (2025). SAGEFusionNet: An Auxiliary Supervised Graph Neural Network for Brain Age Prediction as a Neurodegenerative Biomarker. Brain Sciences, 15(7), 752. https://doi.org/10.3390/brainsci15070752

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop