Next Article in Journal
Automated Assessment of Ki-67 Labeling Index Using Cell-Level Detection and Classification in Whole-Slide Images
Previous Article in Journal
Association Between Serum Testosterone Levels and Coronary Artery Stenosis: A Cross-Sectional Study in Central European Population
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Glaucoma Classification Using a NFNet-Based Deep Learning Model with a Customized Hybrid Attention Mechanism

Lister Hill National Center for Biomedical Communications, National Library of Medicine, National Institutes of Health, Bethesda, MD 20892, USA
*
Author to whom correspondence should be addressed.
Diagnostics 2026, 16(5), 815; https://doi.org/10.3390/diagnostics16050815
Submission received: 30 January 2026 / Revised: 23 February 2026 / Accepted: 1 March 2026 / Published: 9 March 2026
(This article belongs to the Section Machine Learning and Artificial Intelligence in Diagnostics)

Abstract

Background/Objectives: Glaucoma is a leading cause of irreversible blindness worldwide, making accurate and efficient detection methods essential. One primary concern with glaucoma is that it often presents no early symptoms. Vision loss typically begins at the periphery and progresses unnoticed until it significantly affects central vision. Due to this gradual and usually silent progression, early detection through regular eye exams is vital for preventing permanent vision loss. Methods: In this study, we propose a hybrid attention mechanism that recalibrates feature maps from the feature extractor for glaucoma detection. We explored normalization-free ResNet (NF-ResNet) architectures to evaluate the proposed attention mechanism, specifically NF-ResNet-26, NF-ResNet-50, and NF-ResNet-101, in comparison to baseline state-of-the-art ResNet variants. Our approach was evaluated on three publicly available glaucoma datasets, LAG, EyePACS, and BrG, to differentiate between normal and glaucomatous from fundus images. Results: The experimental results demonstrate that our proposed hybrid attention module, combined with normalization-free architectures, significantly enhances performance compared to state-of-the-art ResNet variants. The proposed attention model based on the normalization-free ResNet-50 achieved an accuracy of 0.9394 on the LAG dataset, 0.9117 on the EyePACS dataset, and 0.9020 on the BrG dataset. When evaluated on the combined dataset, the model achieved an accuracy of 0.9193, sensitivity of 0.9182, and specificity of 0.9202. Conclusions: The results from these representative datasets for glaucoma detection highlight the exceptional performance of our attention module, establishing it as a highly competitive classification model in the field of glaucoma detection.

1. Introduction

Glaucoma is a condition that damages the eye’s optic nerve and can result in vision loss and even blindness. It damages the optic nerve that sends images to the brain. As the world’s population ages, glaucoma is becoming an increasing cause of blindness. According to the WHO [1], glaucoma is the second leading cause of blindness globally, after cataracts. However, glaucoma presents a more significant challenge since the blindness it causes is irreversible. It has been estimated that 60.5 million people were affected by glaucoma in 2010, with a projected number of 112 million by 2040 [2]. Many people with the disease are unaware of it, even in developed countries. This lack of awareness can lead to advanced disease and significant vision loss at the time of diagnosis. So, finding the disease early and acting is crucial for reducing vision loss due to glaucoma.
Glaucoma occurs when the optic nerve becomes damaged, leading to a gradual deterioration that results in blind spots in vision. This damage is often associated with increased intraocular pressure, caused by a buildup of aqueous humour that usually flows inside the eye. Under normal conditions, this fluid drains through the trabecular meshwork located at the junction of the iris and cornea. The cornea is vital for vision, allowing light to enter the eye. However, intraocular pressure can rise if the eye produces too much fluid or if the drainage system malfunctions. In individuals with glaucoma, elevated pressure or reduced blood flow to the optic nerve can cause nerve fibres to die. This leads to enlargement of the optic cup relative to the optic disc due to the loss of supporting tissue. While optic nerve cupping is present in healthy individuals and those with glaucoma, those with the condition tend to have a larger cup-to-disc ratio. A cup-to-disc ratio greater than 0.6 is generally considered suspicious for glaucoma.
Currently, glaucoma diagnosis and monitoring require a complete eye examination, additional testing, and gathering a slew of data, which can be challenging to interpret. Moreover, ophthalmologists use multiple manual methods to diagnose glaucoma, including gonioscopy, pachymetry, tonometry, and perimetry [3]. However, manual assessment methods are very time-consuming and largely depend on the availability of ophthalmologists. Furthermore, there is a significant overlap in the ocular features of normal subjects and patients with early glaucoma. For these reasons, there is interest in developing complementary techniques—such as artificial intelligence (AI) systems—to diagnose glaucoma effectively.
Artificial intelligence (AI) has revolutionized the detection and diagnosis of retinal diseases in recent years, mainly through advances in deep learning and computer vision. AI algorithms, such as convolutional neural networks (CNNs), have been trained on large datasets of retinal images to accurately identify conditions like glaucoma, diabetic retinopathy, age-related macular degeneration, and cataracts. These models can quickly analyze medical images like optical coherence tomography (OCT) scans and fundus photographs, detecting subtle abnormalities the human eye may overlook. Using AI-powered tools improves diagnostic accuracy and enables early detection, a crucial factor in preventing vision loss. Furthermore, AI systems have the potential to enhance access to eye care, particularly in remote or underserved areas, by providing real-time screening and decision support for healthcare professionals. This integration of AI into ophthalmology is setting a new standard for personalized and efficient patient care, potentially benefiting society at large.
Das et al. [4] proposed a novel cascaded attention module (CAM) consisting of a triplet channel attention block (TCAB) and a spatial attention block (SAB) on top of a backbone network for glaucoma classification. A cascaded attention module (CAM) mainly consists of two components, namely a triplet channel attention block (TCAB) and a spatial attention block (SAB), connected in a cascaded manner. The CAM builds feature relationships along channel, cross-channel, and spatial dimensions and facilitates extracting rich discriminative features from the lesion regions of fundus images. The study clearly shows an improvement in performance on Harvard Dataset V1 and LMG datasets by adding the CAM on top of the backbone network. Various backbone networks have been tested, and DenseNet-121 with CAM performed better than the other backbone networks; the model with DenseNet-121 achieved an accuracy of 85.34% on the Harvard dataset and 83.85% on the LMG dataset.
Latif et al. [5] proposed a two-phase optic disc localization and glaucoma diagnosis network (ODGNet). A shallow convolutional neural network was incorporated in the first phase to localize optic discs from fundus images. In the later phase, AlexNet, ResNet, and VGGNet were used to diagnose glaucoma. The models were trained using transfer learning with pre-trained weights on five retinal datasets: ORIGA, HRF, DRION-DB, DR_HAGIS, and RIM_ONE.
Fan et al. [6] used a vision transformer deep learning technique, data-efficient image transformer (DeiT), and ResNet-50 trained on fundus images from an ocular hypertension treatment study to detect primary open-angle glaucoma. This study comprehensively explored the generalizability and explainability of vision transformers to detect glaucoma using fundus photographs. The extensive experiments suggest that vision transformers (ViT) generalize well to the eyes of individuals of Chinese, Japanese, Spanish, African and European fundus photographs. Furthermore, ViT focused on localized features of the neuroretinal rim, often used in the clinical management of glaucoma.
Silva et al. [7] developed a tool for classifying diabetic retinopathy from retinal images. They utilized the pre-trained weights of a Normalizer-Free Neural Network (NFNet) as a feature extractor. Subsequently, they incorporated classifiers such as Naïve Bayes, k-Nearest Neighbours (kNN), Multi-Layer Perceptron (MLP), Ordered Probabilistic Forest (OPF), and Support Vector Machine (SVM) to classify the retinal images as either healthy or affected by diabetic retinopathy.
Yang et al. [8] proposed a Mobilenetv2 lightweight network for retinopathy classification, incorporating spatial and channel reconstruction convolution and convolutional block attention module mechanisms into the framework. Incorporating attention modules helped reduce the features by limiting the redundant features and enhanced the feature representation ability for classification improvement. This study established a retinal dataset from ophthalmic clinical images and classified them based on retinal B-Scan image features, which were obtained from the First Affiliated Hospital of Zhengzhou University. This dataset has five classes: AMD, DME, DRUSEN, MNV, and Normal.
Chakraborty et al. [9] presented an attention-aided DenseNet-121 for detecting glaucoma from fundus images. They introduced an amalgamation of two attention modules, the convolutional block attention module (CBAM) and the Channel recalibration module (CRM), on the extracted features from DenseNet-121. The CBMA helps highlight relevant spatial and channel features extracted by the feature extractor. The channel recalibration module further enriches the features by utilizing the edge information and the spatial dimension’s statistical features. RIM-ONE and ACRIMA datasets have been used in this study.
Das et al. [10] proposed AES-NET for effective multi-stage glaucoma classification using fundus images. The fundus images are passed into a backbone network, which can be any ImageNet pre-trained CNN. The feature maps from the last convolution layer of the backbone network are then passed to the spatial adapter module. The spatial adapter module allows the adjustment of its spatial representations per the target task’s requirements. After the spatial adapter module, the features are passed to an enhanced self-attention module to selectively extract more detailed-lesion features from fundus images. Harvard Dataverse V1 and LMG datasets have been used in the study.
Most of the studies discussed have used small-sized datasets with high-quality fundus images. Moreover, minimal studies have been done on normalization-free networks. In this study, we used three datasets of different qualities to train the models and for robust evaluation of the proposed attention modules along with normalization-free architectures.
In this paper, we:
  • Propose a hybrid attention module that combines spatial and channel attention to calibrate features for glaucoma classification from fundus images.
  • Explore normalization-free networks as feature extractors in combination with hybrid attention modules. Various hybrid attention modules were utilized alongside distinct versions of normalization-free ResNets (NF-ResNets) [11] models.
  • Analyze our hybrid attention module with NFNets across three datasets: LAG [12], BrG [13], and EyePacs [14].
  • Further evaluate the proposed attention modules using five-fold cross-validation on the combined LAG, BrG, and EyePACS datasets and conduct comparisons with state-of-the-art (SOTA) models.

2. Materials and Methods

2.1. Datasets

We used three publicly datasets for training and evaluating the models robustly. Samples of images of standard and glaucomatous are displayed in Figure 1.
LAG [12]: The subset of the LAG database has been used for experiments and comprises 4844 fundus images, including 1711 samples confirmed to have glaucoma and 3143 normal samples sourced from Beijing Tongren Hospital. Each image has been annotated by certified glaucoma specialists, who consider various factors, including morphological and functional analyses such as intraocular pressure, visual field loss, and manual assessment of the optic disc to annotate the fundus image. The training set includes 3064 samples, the validation set contains 802 samples, and the test dataset consists of 991 samples.
EyePACS-light (v2) [13]: This enhanced glaucoma dataset comprises a balanced selection of standardized fundus images from the Rotterdam EyePACS AIROGS dataset. It includes data collected from 500 diverse sites with varying ethnicities. The dataset is organized into training, validation, and testing folders, containing 4000 (~84%), 385 (~8%), and 385 (~8%) fundus images across each class, respectively.
BrG [14]: The fundus images in this dataset were collected from Hospital de Olhos (HO) in Minas Gerais (MG) and Policlinica de Unai, MG, Brazil. The dataset includes 2000 images, with 1000 representing glaucoma cases and 1000 normal cases. A group of 1000 volunteers had their eye fundus images captured for both left and right eyes. The images were captured using a smartphone attached to a panoptic ophthalmoscope, utilizing the iExaminer app installed on the smartphone.
In our initial experiments, the LAG and EyePACS-light datasets were originally divided into training, validation, and test sets, whereas the BrG dataset did not include predefined splits. For the BrG dataset, we performed a patient-level split, allocating 70% of the data for training and 30% for testing. From the training set, 20% was further used as a validation set.
To conduct a more robust model evaluation, we adopted a five-fold cross-validation strategy on the combined dataset. Each individual dataset was first divided into five folds. For each cross-validation iteration, four folds from each dataset were merged to form the training set, while the remaining folds were reserved for testing. The merged training data were then randomly split into 80% for training and 20% for validation, while preserving class balance. This approach enables a thorough assessment across multiple datasets while maintaining balanced class distributions in both the training and validation sets. The model was trained and evaluated separately in each of the five cross-validation folds.

2.2. Deep Learning Architecture

The proposed deep learning architecture comprises a backbone network, a hybrid attention module, and a classification head, as shown in Figure 2. The backbone network captures hierarchical features from raw input images. These extracted features provide a strong foundation for downstream tasks, allowing complex information to be represented in a compact and useful format. The hybrid attention module is designed to selectively enhance important information from the feature maps that the feature extractor provides. Subsequently, the recalibrated features are passed through a classification head, which includes a global average pooling layer and a fully connected layer to perform glaucoma classification.

2.2.1. Backbone

As illustrated in the first block in Figure 2, we provide the fundus image to pass as an input through the backbone network. The features are extracted from the last convolutional layer of the backbone network. This study considered ResNet [15] and NF-ResNet variants as backbone networks.
The choice of the ResNet architecture was driven by both its theoretical foundation and practical experimental considerations. NFNet’s normalization-free design was originally an extension of the residual learning approach, making ResNet a logical and structurally consistent backbone for this study. Using ResNet enables a fair comparison among standard residual models, normalization-free variants, and our hybrid attention mechanism, while reducing architectural confounds. This selection was primarily dictated by methodological reasoning rather than solely by ablation outcomes, with subsequent experiments confirming the effectiveness of our proposed modifications.
In these experiments, we explored networks with 26, 50, and 101 layers for extreme analysis of the models.

2.2.2. Hybrid Attention Module

We propose an attention module that helps to recalibrate the features extracted from the feature extractor to help capture rich semantic information from fundus images. The hybrid attention module comprises spatial and channel attention as shown in the second block in Figure 2. The hybrid attention module is placed between the feature extractor and the classification, designed to selectively enhance the critical information from feature maps at both channel and spatial levels.
Let X   the R   C   ×   H   ×   W represent the feature map from the feature extractor, where C represents channels, height H, and width W. Before passing the features to the attention modules, a pointwise convolution is applied to reduce the number of channels and is controlled by the reduction ratio (r). The reduction ratio controls the degree of channel compression while retaining important information and reducing the number of computations within the module.
X r e d u c e d = C o n v 1 x 1   ( X )
where X r e d u c e d     R C ×   H   ×   W   and C = C / r   where r is the reduction ratio. This down sample output in X r e d u c e d   is passed through spatial and channel attention modules.
Channel attention block: A channel attention [16] module enhances the prioritized channels, which boosts performance and suppresses channels with redundant information. Channel attention exploits the inter-channel relationship of features. As shown in Figure 3, the first step in channel attention is to apply average pooling and max pooling operations, generating two different spatial context descriptors, as shown in the equations below.
F a v g c = A v g P o o l ( X r e d u c e d )
F m a x c = M a x P o o l ( X r e d u c e d )
The descriptors are then forwarded to a shared multi-layer perceptron (MLP) with one hidden layer to produce an attention map M c   R c ×   1   ×   1 . After applying the shared network to each descriptor, we merge the outputs using elementwise summation.
M c F = σ ( M L P ( F a v g c ) + M L P ( F m a x c ) ) = σ (   W 1 ( W 0 ( F a v g c ) ) + W 1 ( W 0 ( F m a x c ) ) )
where σ denotes the sigmoid function, W 0     R c r 2 ×   c   , and W 1     R c ×   c r 2 where r 2 is the reduction ratio to reduce the parameters in the MLP. MLP weights W 0 and W 1 are shared by both descriptors and the ReLU function is followed by W 0 . The final attention map and the input features X r e d u c e d   are multiplied to create the recalibrated features
X C h a n n e l = X r e d u c e d M c F
where X c h a n n e l     R c ×   H   ×   W .
Spatial attention block: Spatial attention [16] is a mechanism that selectively focuses on specific spatial locations within feature maps, highlighting areas of importance while de-emphasizing others. This approach improves the model’s ability to capture spatial dependencies and allows it to dynamically allocate attention across the feature map’s spatial dimensions (height and width). As shown in Figure 4, the first step in this module is to aggregate the channel information using two pooling operations, generating two 2D maps F a v g s   R 1 ×   H   ×   W   and F m a x s   R 1 ×   H   ×   W . F a v g s and F m a x s denote the average-pooled and max-pooled features across the channel. They are then concatenated and convolved by a standard convolution layer, producing an attention map. Overall, the spatial attention map is computed as follows.
M s   F = σ   ( f 7 × 7 ( [ F a v g s ; F m a x ] s ) )
where σ denotes the sigmoid function and f 7 × 7 , representing a convolution operation with a filter size of 7 × 7. The attention map M s is used for element-wise multiplication of each channel in X r e d u c e d
X s p a t i a l = X r e d u c e d     M s
Output of Hybrid Attention Module: The final step in this hybrid attention module is to concatenate X s p a t i a l , X c h a n n e l along the channel dimension as represented in Equation (8).
X o u t p u t = C o n c a t ( X c h a n n e l , X s p a t i a l )
where X o u t p u t     R   2   ×   C   ×   H   ×   W .
Through this module, the attention head module adaptively enhances relevant features, leveraging both channel and spatial contexts while maintaining an efficient parameter count.

2.2.3. Classification Head

After the feature extraction and attention layer, the features are passed through the classification head, as shown in the third block in Figure 2. The classification head contains a Global Average Pooling (GAP) layer, which reduces spatial dimensions of X o u t p u t by taking the average across each feature map. The GAP operation can be described by
F G A P c = 1 H × W   i = 1 H j = 1 W X o u t p u t [ c , i , j ]
where H and W are the height and width of feature maps, and c denotes each channel. The pooled feature vector F G A P is then passed through a fully connected layer to classify into normal and glaucoma, as illustrated in Figure 2.

3. Results

This section presents the implementation details and results of the proposed framework, along with state-of-the-art models. We compare our proposed model’s performance with normalization-free ResNets and baseline ResNet models.

3.1. Training Setting

We trained the models over 100 epochs for optimal performance, incorporating a warm-up period of five epochs and maintaining consistent settings across all experiments. A cosine scheduler was employed alongside the Adam optimizer, using a momentum of 0.9 and a learning rate of 0.00005. We selected cross-entropy loss as the loss function for model training. The fundus images in the experiments were resized to 224 × 224 pixels during model training and we applied data augmentation techniques, such as random cropping, horizontal flipping, and colour jittering at random during training. The models were implemented in PyTorch (2.7.1) .

3.2. Test Results

This section presents the results of implementing the new proposed hybrid attention model between feature extractors and classification heads in normalization-free ResNets. We also compared the performance against the baseline ResNet and normalization-free ResNet models. Multiple performance metrics—accuracy (Acc), kappa score, sensitivity (Sen), specificity (Spec), and area under the ROC curve (AUC)—were considered to effectively assess the performance of the proposed model.
Table 1 includes the quantitative performance metrics of models trained and tested on the LAG dataset. The normalization variants of ResNet outperformed the baseline ResNet models. The models trained with the proposed attention module performed better than ResNet and NF_ResNets. NF_ResNets101 with a hybrid attention module with a reduction ratio r = 4 in the attention module achieved the best accuracy of 0.9394 compared to all the other models. The hybrid attention mechanism, along with normalization-free networks, improved the accuracy of the ResNet26 (r = 4) variant from 0.9081 to 0.9364, the ResNet50 variant with r = 4 from 0.9162 to 0.9323, and the ResNet101 variant with r = 4 from 0.9232 to 0.9394 on the LAG dataset. All other performance metrics including sensitivity, specificity, AUC and kappa also improved after adapting the proposed attention module.
Table 2 illustrates our training and evaluation of baseline and proposed models using the EyePACS dataset. Models utilizing the hybrid attention approach outperformed the baseline models across all three variants. Moreover, normalization-free ResNets demonstrated enhanced performance compared to the standard ResNet models. Furthermore, introducing a customized attention module between the feature extractor and the classification head improved performance across all three variants. All the normalization-free variants with a reduction ratio r = 4 achieved the same accuracy of 0.9130 on this dataset. NF_ResNet26 with r = 2 achieved the best sensitivity of 0.9403, Normalization-free variants of resnet50 (r = 2) and resnet101 (r = 2, 4) achieved a similar specificity of 0.9039. Coming to kappa score, all the normalization-free ResNet variants with the hybrid attention module with r = 4 achieved a kappa score 0.8260.
Table 3 shows the performance of the models trained on the BrG dataset. Similarly to results from other datasets, the normalization-free networks combined with the proposed hybrid attention module demonstrate improved performance compared to the baseline ResNet models. Specifically, the NF_ResNet101 model with an attention module with r = 2 achieved an accuracy of 0.8920, outperforming all other variants on the BrG dataset. The NF_ResNet50 with a ratio r = 2 recorded an accuracy of 0.8887, while the NF_ResNet26 with a hybrid module with r = 2 achieved 0.8821. The NF_ResNet101 variant with the hybrid attention module with r = 2 also surpassed other models with a sensitivity of 0.8933, an AUROC score of 0.9494, and a kappa of 0.7842. Additionally, the NF_ResNet50 with the hybrid attention module with r = 2 achieved the best specificity of 0.9139, and NF_ResNet101 with r = 2 achieved the best sensitivity of 0.8933.
Table 4 shows the performance of models trained on the combined datasets using five-fold cross-validation. The NF_ResNet with the proposed hybrid attention modules outperformed the regular state-of-the-art ResNet and NF_ResNet models. Specifically, NF_ResNet50 (r = 2) achieved the highest accuracy of 0.9193, outperforming all other models, including the baseline ResNet50 and NF_ResNet50 models. In addition, NF_ResNet101 (r = 2 and r = 4) achieved the same accuracy of 0.9164, surpassing the performance of the baseline NF_ResNet101 and ResNet101 models.
When comparing the performance reported in Table 4, the NF-ResNet attention models trained on the combined dataset demonstrated slightly lower accuracy than the models trained on each individual dataset, as shown in Table 1, Table 2 and Table 3. However, we expect that the NF-ResNet attention models trained on the combined dataset may provide more reliable and robust performance when applied to images from previously unseen datasets.
Table 5 presents a performance comparison between existing studies and our proposed models. Our models demonstrate competitive performance relative to other approaches, despite being trained from scratch without pretrained weights from ImageNet or other external libraries. Although some previous studies report higher performance on glaucoma datasets, many of these works employ different experimental protocols, including combined multi-dataset training, diverse preprocessing pipelines, and alternative data splits. These variations make direct numerical comparisons challenging.
In contrast, our study adopts a controlled experimental framework to evaluate the effect of the proposed hybrid attention mechanism on NFNet under consistent training conditions. The primary objective of this work is methodological enhancement rather than large-scale dataset optimization. Nevertheless, we have included a literature comparison table to better contextualize our results within the field.
Table 6 compares the performance of the proposed models with existing state-of-the-art (SOTA) models. To ensure a fair comparison, all models were trained from scratch and evaluated using five-fold cross-validation under the same experimental settings. Among the SOTA models, DenseNet121 achieved the highest accuracy of 0.9193. However, all six proposed models outperformed the SOTA models in overall performance.

4. Discussion

We monitored the variations in training loss, validation loss, and validation accuracy during the training process. Figure 5 illustrates the training curves of these three metrics over epochs for NF_ResNet50 (r = 2), trained on the first fold of the combined dataset under five-fold cross-validation. The training loss, validation loss, and validation accuracy converge with minimal fluctuation after approximately 80 epochs. Similar patterns were observed across the remaining four folds.
Figure 6 displays the confusion matrix for ResNet101 variants and the best-performing NF_ResNet101 variants with the proposed hybrid attention modules across all datasets, as reported in Table 1, Table 2 and Table 3. The matrices in the first row (a, b, and c) show the performance of NF-ResNet101, while those in the second row (d, e, and f) present the performance of NF-ResNet101 with the proposed hybrid attention modules. The first, second, and third columns display the results for the LAG, EyePACS, and BrG datasets, respectively. Comparing ResNet101 to the NF_ResNet101 with the attention module at r = 2 on the LAG dataset, there are only 42 false negatives, and 18 false positives compared to 49 and 23 on ResNet101.
The other confusion matrices for the EyePACS and BrG datasets clearly indicate a reduction in both false positives and negatives for the NF_ResNet101 with the proposed hybrid attention module, in contrast to the ResNet101 models that do not incorporate the hybrid attention module.
Figure 7 shows the ROC curves of the trained models on each dataset for Table 1, Table 2 and Table 3. Figure 7a displays the ROC curves for models trained and evaluated on the LAG dataset. Normalization-free ResNet26 with a hybrid attention module of reduction ratio r = 2 achieved AUC = 0.9838 compared to all the other models trained on the LAG dataset. Figure 7b shows the ROC curves for models trained on the EyePACS dataset. NF_ResNet101 with the attention module using a reduction ratio of r = 2 achieved AUC = 0.9693, which is the best performance on the EyePACS dataset. Figure 7c displays the ROC curves for models trained on the BrG dataset; the NF_ResNet26, along with the attention module having a reduction ratio of r = 2, achieved 0.9498, which outperformed other models on this dataset.
Figure 8 presents the ROC curves of the models trained and evaluated on the five folds of the combined dataset, as summarized in Table 4. Figure 8a shows the ROC curves for models trained and tested on the first fold, where NF_ResNet101 (r = 2) achieved the highest performance with an AUC of 0.9731. For folds 2–5, NF_ResNet50 (r = 2) achieved the highest AUC values of 0.9711, 0.9722, 0.9707, and 0.9751, respectively, demonstrating superior performance across these folds. Overall, the proposed attention-based models consistently achieved higher AUC values than the baseline models across all five folds.
We conducted an error analysis based on the results presented in Table 1, Table 2, Table 3 and Table 4. Among all the proposed attention-based models, comparatively lower performance was observed on the BrG dataset than on the other two datasets. This may be attributed to relatively lower image quality, as the images were captured using an ophthalmoscope attached to a smartphone. In addition, the fundus images in the EyePACS dataset typically cover the entire retinal field, whereas images in the LAG dataset are more focused on the optic disc region. Since the optic disc region contains key clinical indicators of glaucoma, such as the cup-to-disc ratio, the LAG images may provide more detailed structural information relevant to glaucoma classification. Consequently, the proposed attention-based models generated fewer classification errors on the LAG dataset than on the EyePACS dataset.
Classification errors were frequently observed in low-quality images across the three datasets. Furthermore, when the optic disc region appeared excessively bright compared to the surrounding retinal areas, some non-glaucoma images were misclassified as glaucoma. Several non-glaucoma images with peripheral retinal abnormalities outside the optic disc were also incorrectly labelled as glaucoma. We also observed that glaucoma images with low color intensity were sometimes misclassified as normal, whereas non-glaucoma images with uniformly high color intensity across the retina were occasionally predicted as glaucoma.
In Figure 9, we present the heatmaps produced by Grad-CAM++ [19] from several randomly selected retina images across all datasets. NF_ResNet101 equipped with the proposed hybrid attention module (r = 2) was utilized to generate the heatmaps. The heatmaps shown in the first, second, third columns correspond to regular ResNet101, NF-ResNet101, and proposed NF-ResNet with the hybrid attention module (r = 2), respectively.
For ResNet101 and NF_ResNet101, the heatmaps were generated from the last convolutional layer, whereas for the proposed NF_ResNet101 (r = 2), the heatmaps were derived from the concatenated outputs of the channel and spatial attention modules.
All normalization-free ResNet variants, together with the proposed hybrid attention module, surpassed all metrics compared to ResNet and normalization-free ResNets. The architectures employing this hybrid attention improve feature maps by utilizing channel and spatial attention mechanisms. This enables the model to concentrate on the most significant channels and spatial areas, thereby enhancing both the relevance of features and the model’s efficiency. Furthermore, attention mechanisms are typically lightweight, introducing only a minimal number of additional parameters relative to the original architecture. By emphasizing key features selectively, they optimize parameter usage and bolster the model’s representational capabilities. This focused approach mitigates the likelihood of overfitting irrelevant details or noise in the training data, thus boosting performance on test data.
Unlike other retinal diseases, such as age-related macular degeneration (AMD), cataracts, and diabetic retinopathy, where key diagnostic features can be identified, glaucoma cannot be diagnosed solely based on fundus images. Clinical diagnosis requires a comprehensive evaluation that includes multiple examinations: tonometry to measure intraocular pressure (IOP), visual field testing to detect peripheral vision loss, optical coherence tomography (OCT) to measure retinal nerve fibre layer (RNFL) thickness, ophthalmoscopy to assess the cup-to-disc ratio (CDR) and neuroretinal rim thickness, pachymetry to measure corneal thickness, and gonioscopy to evaluate the anterior chamber angle. Among these, longitudinal assessment using tonometry, OCT, and visual field testing is particularly important for confirming diagnosis and monitoring disease progression.
Fundus imaging primarily provides structural information regarding the optic nerve head, including CDR and rim thickness. Traditionally, a CDR greater than 0.6 has been used as a threshold to classify glaucomatous eyes. However, reliance on CDR alone has limitations, as some healthy individuals present a relatively large CDR and thin neuroretinal rims without functional impairment. Consequently, fundus image analysis alone is insufficient for glaucoma diagnosis in clinical practice.
Despite these limitations, the proposed deep learning models demonstrated strong performance using fundus images alone, achieving accuracies of 0.8920 on the BrG dataset, 0.9130 on the EyePACS dataset, and 0.9394 on the LAG dataset. Notably, the heatmap visualizations from the ResNet101 model with the proposed hybrid attention module revealed activation patterns concentrated on retinal blood vessels. In contrast, models without the attention mechanism (ResNets and NF_ResNet variants) primarily focused on the central optic disc region as shown in Figure 9. These findings suggest that the hybrid attention module may capture subtle structural cues beyond conventional CDR-based features, potentially including vascular alterations associated with glaucomatous damage.
Nevertheless, given the multifactorial nature of glaucoma, multimodal analysis integrating fundus photography, OCT imaging, and visual field data is likely necessary for robust clinical evaluation. Future work should focus on multimodal fusion approaches to better characterize the relationship between structural changes and functional impairment, and to further investigate the potential association between retinal vascular features and glaucomatous progression.
NF-ResNet26 hybrid attention (r = 4) achieved the best performance on the LAG and EyePACS datasets and the second-best overall performance, with only a marginally higher error rate than NF-ResNet26 hybrid attention (r = 2). However, in clinical setting, NF-ResNet50 hybrid attention (r = 2) would be considered the preferred model, as it achieved the best performance on the combined dataset. This model is expected to provide more reliable and robust performance when applied to input images from previously unseen datasets.

5. Conclusions

This study proposes a deep learning model for glaucoma classification from fundus images using a hybrid attention module, which significantly enhances the model’s performance over baseline models when integrated between the feature extractor and the classification head. To thoroughly assess the effectiveness of our proposed module for glaucoma detection in fundus images, we investigated various configurations of Residual Networks (ResNets), including both standard ResNets and normalization-free (NF) ResNets. By evaluating these different variants, we aimed to provide a comprehensive analysis of the performance improvements offered by the proposed hybrid attention module. Our evaluation consisted of an extensive series of experiments conducted on three distinct datasets: LAG, BrG, and EyePACS. The results from these experiments demonstrated that models utilizing normalization-free architectures, when enhanced with our hybrid attention module, consistently outperformed baseline models across all datasets examined. Notably, normalization-free ResNet models exhibited superior performance compared to state-of-the-art ResNet models across all variants tested. For future work, we plan to expand the dataset by adding new retinal images and including data on other retinal diseases.

Author Contributions

Conceptualization, S.A. and J.K.; methodology, S.A. and J.K.; software, S.A.; validation, S.A., L.T. and J.K.; formal analysis, S.A. and J.K.; investigation, S.A. and J.K.; data curation, S.A.; writing—original draft preparation, S.A.; writing—review and editing, S.A., L.T. and J.K.; supervision, J.K. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the Lister Hill National Center for Biomedical Communications of the National Library of Medicine (NLM), National Institutes of Health (NIH). The contributions of the NIH authors are considered Works of the United States Government. The findings and conclusions presented in this paper are those of the authors and do not necessarily reflect the views of the NIH or the U.S. Department of Health and Human Services.

Informed Consent Statement

This study involved secondary analysis of de-identified, open-source data; therefore, informed consent was not required.

Data Availability Statement

The original data presented in this study are openly available. The LAG dataset is available at GitHub (https://github.com/smilell/AG-CNN?tab=readme-ov-file, accessed on 8 January 2025). The EyePACS dataset is available via Kaggle (https://www.kaggle.com/datasets/deathtrooper/eyepacs-airogs-light, accessed on 9 January 2025). The BrG dataset is available via Kaggle (https://www.kaggle.com/datasets/clerimar/brasil-glaucoma-brg, accessed on 28 January 2025). The source code is available at: https://github.com/LHNCBC/AI-based-Clinical-Data-Analysis/tree/main/202601_MDPI_Diagnostics_Glaucoma_HA, accessed on 28 January 2025).

Acknowledgments

We would like to thank Violanda Grigorescu and Paul Lynch, NLM, for reviewing the manuscript.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
AIArtificial intelligence
CNNConvolutional Neural Network
NFNetNormalizer-Free Neural Network

References

  1. Kingman, S. Glaucoma is second leading cause of blindness globally. Bull. World Health Organ. 2004, 82, 887–888. [Google Scholar] [PubMed]
  2. Tham, Y.C.; Li, V.; Wong, T.Y.; Quigley, H.A.; Aung, T.; Cheng, C.Y. Global prevalence of glaucoma and projections of glaucoma burden through 2040: A systematic review and meta-analysis. Ophthalmology 2014, 121, 2081–2090. [Google Scholar] [CrossRef] [PubMed]
  3. Lima, A.; Maia, L.B.; dos Santos, P.T.C.; Braz Junior, G.; de Almeida, J.D.S.; de Paiva, A.C. Evolving Convolutional Neural Networks for Glaucoma Diagnosis. Braz. J. Health Rev. 2020, 3, 1873–1882. [Google Scholar] [CrossRef]
  4. Das, D.; Nayak, D.R.; Pachori, R.B. CA-Net: A Novel Cascaded Attention-Based Network for Multistage Glaucoma Classification Using Fundus Images. IEEE Trans. Instrum. Meas. 2023, 72, 1–10. [Google Scholar] [CrossRef]
  5. Latif, J.; Tu, S.; Xiao, C.; Ur Rehman, S.; Imran, A.; Latif, Y. ODGNet: A deep learning model for automated optic disc localization and glaucoma classification using fundus images. SN Appl. Sci. 2022, 4, 98. [Google Scholar] [CrossRef]
  6. Fan, R.; Alipour, K.; Bowd, C.; Christopher, M.; Brye, N.; Proudfoot, J.A.; Goldbaum, M.H.; Belghith, A.; Girkin, C.A.; Fazio, M.A.; et al. Detecting Glaucoma from Fundus Photographs Using Deep Learning without Convolutions: Transformer for Improved Generalization. Ophthalmol. Sci. 2023, 3, 100233. [Google Scholar] [CrossRef] [PubMed]
  7. Da Silva, M.C.; Silva, C.M.; Matos, A.G.; Da Silva, S.P.P.; Sarmento, R.M.; Rebouças Filho, P.P.; Nascimento, N.M.N.; de Santiago, R.V.C.; Benevides, C.A.; Henrique Cunha, C.C. A New Diabetic Retinopathy Classification Approach Based on Normalizer Free Network. In Proceedings—IEEE Symposium on Computer-Based Medical Systems; Institute of Electrical and Electronics Engineers Inc.: Piscataway, NJ, USA, 2024; pp. 164–169. [Google Scholar]
  8. Yang, B.; Zhang, Z.; Yang, P.; Zhai, Y.; Zhao, Z.; Zhang, L.; Zhang, R.; Geng, L.; Ouyang, Y.; Yang, K.; et al. MobilenetV2-RC: A lightweight network model for retinopathy classification in retinal OCT images. J. Phys. D Appl. Phys. 2024, 57, 505401. [Google Scholar] [CrossRef]
  9. Chakraborty, S.; Roy, A.; Pramanik, P.; Valenkova, D.; Sarkar, R. A Dual Attention-aided DenseNet-121 for Classification of Glaucoma from Fundus Images. In Proceedings of the 2024 13th Mediterranean Conference on Embedded Computing (MECO), Budva, Montenegro, 11–14 June 2024. [Google Scholar]
  10. Das, D.; Nayak, D.R.; Pachori, R.B. AES-Net: An adapter and enhanced self-attention guided network for multi-stage glaucoma classification using fundus images. Image Vis. Comput. 2024, 146, 105042. [Google Scholar] [CrossRef]
  11. Brock, A.; De, S.; Smith, S.L.; Simonyan, K. High-Performance Large-Scale Image Recognition Without Normalization. Proc. Mach. Learn. Res. 2021, 139, 1059–1071. [Google Scholar]
  12. Li, L.; Xu, M.; Wang, X.; Jiang, L.; Liu, H. Attention based glaucoma detection: A large-scale database and CNN model. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 16–20 June 2019. [Google Scholar]
  13. Bragança, C.P.; Torres, J.M.; Soares, C.P.D.A.; Macedo, L.O. Detection of Glaucoma on Fundus Images Using Deep Learning on a New Image Set Obtained with a Smartphone and Handheld Ophthalmoscope. Healthcare 2022, 10, 2345. [Google Scholar] [CrossRef] [PubMed]
  14. De Vente, C.; Vermeer, K.A.; Jaccard, N.; Wang, H.; Sun, H.; Khader, F.; Truhn, D.; Aimyshev, T.; Zhanibekuly, Y.; Le, T.D.; et al. AIROGS: Artificial Intelligence for Robust Glaucoma Screening Challenge. IEEE Trans. Med. Imaging 2024, 43, 542–557. [Google Scholar] [CrossRef] [PubMed]
  15. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016. [Google Scholar]
  16. Woo, S.; Park, J.; Lee, J.Y.; Kweon, I.S. CBAM: Convolutional block attention module. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 3–19. [Google Scholar]
  17. Singh, P.B.; Singh, P.; Dev, H.; Batra, D.; Chaurasia, B.K. HViTML: Hybrid vision transformer with machine learning-based classification model for glaucomatous eye. Multimed. Tools Appl. 2025, 84, 33609–33632. [Google Scholar] [CrossRef]
  18. Yu, F.; Wu, C.; Zhou, Y. DG2Net: A MLP-Based Dynamixing Gate and Depthwise Group Norm Network for Classification of Glaucoma. In International Conference on Pattern Recognition; Springer Nature: Cham, Switzerland, 2024. [Google Scholar]
  19. Chattopadhay, A.; Sarkar, A.; Howlader, P.; Balasubramanian, V.N. Grad-CAM++: Generalized gradient-based visual explanations for deep convolutional networks. In Proceedings of the 2018 IEEE Winter Conference on Applications of Computer Vision, WACV, Lake Tahoe, NV, USA, 12–15 March 2018. [Google Scholar]
Figure 1. Sample images of normal and glaucoma from LAG, BrG, and EyePACS datasets.
Figure 1. Sample images of normal and glaucoma from LAG, BrG, and EyePACS datasets.
Diagnostics 16 00815 g001
Figure 2. The overall architecture of the proposed framework for automated classification of glaucoma.
Figure 2. The overall architecture of the proposed framework for automated classification of glaucoma.
Diagnostics 16 00815 g002
Figure 3. The architecture of the channel attention module.
Figure 3. The architecture of the channel attention module.
Diagnostics 16 00815 g003
Figure 4. The architecture of the spatial attention module.
Figure 4. The architecture of the spatial attention module.
Diagnostics 16 00815 g004
Figure 5. Loss and accuracy curves for NF_ResNet50 (r = 2) during training on the first fold of the combined dataset.
Figure 5. Loss and accuracy curves for NF_ResNet50 (r = 2) during training on the first fold of the combined dataset.
Diagnostics 16 00815 g005
Figure 6. Confusion matrix of best-performing models on each dataset. (a) Performance of NF_ResNet101 on LAG dataset. (b) Performance of NF_ResNet101 on EyePACS dataset. (c) Performance of NF_Resnet101 on BrG dataset. (d) Performance of NF_ResNet101 with attention module of r = 2 on LAG dataset. (e) Performance of NF_ResNet101 with attention module of r = 4 on EyePACS dataset. (f) Performance of NF_Resnet101 with attention module of r = 2 on BrG dataset.
Figure 6. Confusion matrix of best-performing models on each dataset. (a) Performance of NF_ResNet101 on LAG dataset. (b) Performance of NF_ResNet101 on EyePACS dataset. (c) Performance of NF_Resnet101 on BrG dataset. (d) Performance of NF_ResNet101 with attention module of r = 2 on LAG dataset. (e) Performance of NF_ResNet101 with attention module of r = 4 on EyePACS dataset. (f) Performance of NF_Resnet101 with attention module of r = 2 on BrG dataset.
Diagnostics 16 00815 g006
Figure 7. ROC curves for models trained on (a) LAG dataset, (b) EyePACS dataset, and (c) BrG dataset.
Figure 7. ROC curves for models trained on (a) LAG dataset, (b) EyePACS dataset, and (c) BrG dataset.
Diagnostics 16 00815 g007
Figure 8. ROC curves for all models trained on the combined dataset: (a) Fold 1, (b) Fold 2, (c) Fold 3, (d) Fold 4, and (e) Fold 5.
Figure 8. ROC curves for all models trained on the combined dataset: (a) Fold 1, (b) Fold 2, (c) Fold 3, (d) Fold 4, and (e) Fold 5.
Diagnostics 16 00815 g008aDiagnostics 16 00815 g008b
Figure 9. Heatmaps for the ResNet101 variants on various datasets. The heatmaps displayed in the first column belong to the regular ResNet101 models. The heatmaps in the second column belong to their respective normalization-free variants of the respective ResNet101 models. The heatmaps in the third column display the heatmaps from NF_ResNet101 with a hybrid attention module of r = 2. (a) LAG Dataset, (b) EyePACS Dataset, (c) BrG Dataset.
Figure 9. Heatmaps for the ResNet101 variants on various datasets. The heatmaps displayed in the first column belong to the regular ResNet101 models. The heatmaps in the second column belong to their respective normalization-free variants of the respective ResNet101 models. The heatmaps in the third column display the heatmaps from NF_ResNet101 with a hybrid attention module of r = 2. (a) LAG Dataset, (b) EyePACS Dataset, (c) BrG Dataset.
Diagnostics 16 00815 g009
Table 1. Performance of models trained and evaluated on LAG dataset.
Table 1. Performance of models trained and evaluated on LAG dataset.
ModelHybrid
Attention (r)
AccuracySensitivitySpecificityF1AUCKappa
 ResNet26 Variants
ResNet26No0.90810.87610.91800.92150.96640.7572
NF_ResNet26No0.91920.88890.92860.93030.97710.7850
NF_ResNet26Yes (2)0.93430.96150.92590.96900.98380.8299
NF_ResNet26Yes (4)0.93640.91030.94440.94450.98150.8290
 ResNet50 Variants
ResNet50No0.91620.89740.9220.93400.97530.7792
NF_ResNet50No0.92320.88460.93520.92890.97430.7940
NF_ResNet50Yes (2)0.92530.96150.91400.96720.98080.8087
NF_ResNet50Yes (4)0.93230.91030.93920.94370.98240.8192
 ResNet101 variants
ResNet101No0.92320.92310.92330.94830.97990.7992
NF_ResNet101No0.92730.90170.93520.93840.98000.8060
NF_ResNet101Yes (2)0.93740.92740.94050.95320.98170.8334
NF_ResNet101Yes (4)0.93940.92310.94440.95150.98370.8379
Table 2. Performance of models trained and evaluated on the EyePACS dataset.
Table 2. Performance of models trained and evaluated on the EyePACS dataset.
ModelHybrid
Attention (r)
AccuracySensitivitySpecificityF1AUCKappa
 ResNet26 Variants
ResNet26No0.88050.90130.85970.88290.95060.7610
NF_ResNet26No0.87660.87530.87790.87640.93950.7532
NF_ResNet26Yes (2)0.88960.91600.86230.89200.96110.7792
NF_ResNet26Yes (4)0.91300.94030.88570.91530.96560.8260
 ResNet50 Variants
ResNet50No0.89480.88830.90130.89410.95630.7896
NF_ResNet50No0.89740.87530.91950.89510.95400.7948
NF_ResNet50Yes (2)0.90650.90910.90390.90670.96890.8130
NF_ResNet50Yes (4)0.91300.93510.89090.91490.96600.8260
 ResNet101 variants
ResNet101No0.87920.88830.87010.88030.95460.7584
NF_ResNet101No0.90000.90910.89090.90090.95240.8000
NF_ResNet101Yes (2)0.90390.90390.90390.90390.96520.8078
NF_ResNet101Yes (4)0.91300.92210.90390.91380.96930.8260
Table 3. Performance of models trained and evaluated on BrG dataset.
Table 3. Performance of models trained and evaluated on BrG dataset.
ModelHybrid
Attention (r)
AccuracySensitivitySpecificityF1AUCKappa
 ResNet26 variants
ResNet26No0.83550.85000.82120.83840.90030.6711
NF_ResNet26No0.86210.89670.82780.86730.93570.7243
NF_ResNet26Yes (2)0.88210.87330.89070.88130.94980.7641
NF_ResNet26Yes (4)0.87710.86000.89400.87520.94150.7541
 ResNet50 variants
ResNet50No0.84880.86000.83770.85100.92760.6977
NF_ResNet50No0.88540.90330.86750.88780.93590.7708
NF_ResNet50Yes (2)0.88870.86330.91390.88600.94840.7774
NF_ResNet50Yes (4)0.88700.87670.89740.88620.94260.7741
 ResNet101 variants
ResNet101No0.87210.85670.88740.87040.93950.7442
NF_ResNet101No0.87870.86670.89070.87750.94710.7575
NF_ResNet101Yes (2)0.89200.89330.89070.89250.94940.7841
NF_ResNet101Yes (4)0.88540.87670.89400.88470.94830.7707
Table 4. Performance of models trained and evaluated on combined LAG, EyePACS, and BrG datasets based on five-fold cross validation.
Table 4. Performance of models trained and evaluated on combined LAG, EyePACS, and BrG datasets based on five-fold cross validation.
ModelHybrid
Attention (r)
AccuracySensitivitySpecificityF1AUCKappa
 ResNet26 variants
ResNet26No0.89760.89560.89940.88870.96050.7940
NF_ResNet26No0.90020.88930.90940.89050.96120.7989
NF_ResNet26Yes (2)0.91540.91530.91550.90800.96980.8297
NF_ResNet26Yes (4)0.91100.91000.91180.90320.96840.8208
 ResNet50 variants
ResNet50No0.90030.88690.91160.89030.96010.7990
NF_ResNet50No0.90720.89960.91350.89840.96530.8129
 NF_ResNet50 Yes (2)0.91930.91820.92020.91220.97230.8375
NF_ResNet50Yes (4)0.91550.91870.91280.90850.97090.8300
 ResNet101 variants
ResNet101No0.90190.88800.91350.89190.96230.8020
NF_ResNet101No0.91040.90590.91420.90220.96730.8195
NF_ResNet101Yes (2)0.91640.91580.91690.90900.97090.8317
NF_ResNet101Yes (4)0.91640.91430.91810.90890.97080.8316
Table 5. Performance comparison with other deep learning methods.
Table 5. Performance comparison with other deep learning methods.
ModelDatasetPretrained WeightAccuracySensitivitySpecificityF1AUC
Ensemble using 5 CNNs [13]BrGYes0.90500.85000.96000.89900.9650
ResNet50 [13]BrGYes0.88100.95300.81000.88900.9560
ResNet101 [13]BrGYes0.88000.91000.85000.88300.9490
DenseNet121 [9]LAGYes0.9381--0.93049-
HViTML [17]LAGYes0.9300----
DeiT [6]LAGYes----0.8800
DG2Net [18]EyePACS-
AIROGS-
light-V2
Yes0.9180  0.91830.9190
MaXViT [18]EyePACS-AIROGS-light-V2Yes0.9325  0.93240.9325
Proposed Hybrid Attention based on NF_ResNet
NF_ResNet101
HA (r = 4)
LAGNo0.93940.92310.94440.95150.9837
NF_ResNet50 HA (r = 2)EyePACSNo0.91300.93510.89090.91490.9660
NF_ResNet101 HA (r = 2)BrGNo0.89200.89330.89070.89250.9494
NF_ResNet50 HA (r = 2)LAG, BrG, EyPACSNo0.91930.91820.92020.91220.9723
Table 6. Performance of models trained and evaluated on combined LAG, EyePACS, and BrG datasets using five-fold cross validation, trained from scratch without pretrained weights.
Table 6. Performance of models trained and evaluated on combined LAG, EyePACS, and BrG datasets using five-fold cross validation, trained from scratch without pretrained weights.
ModelHybrid
Attention (r)
AccuracySensitivitySpecificityF1AUCKappa
ConvNext_Small-0.81460.80040.82640.79750.89790.6265
ConvNext_Base-0.81630.79870.83100.79870.89890.6297
DenseNet121-0.90980.91070.90900.90210.96800.8185
EfficientNetB0-0.80790.79130.82180.78990.89410.6130
EfficientNetB4-0.83980.82920.84860.82520.92120.6773
EfficientNetB7-0.82480.81550.83250.81080.90070.6477
ViT_Small-0.72200.68690.75130.69280.80020.4389
ViT_Base-0.73730.71270.75790.71180.81310.4705
 Proposed Hybrid Attention based on NF_ResNet
NF_ResNet26Yes (2)0.91540.91530.91550.90800.96980.8297
NF_ResNet26Yes (4)0.91100.91000.91180.90320.96840.8208
NF_ResNet50Yes (2)0.91930.91820.92020.91220.97230.8375
NF_ResNet50Yes (4)0.91550.91870.91280.90850.97090.8300
NF_ResNet101Yes (2)0.91640.91580.91690.90900.97090.8317
NF_ResNet101Yes (4)0.91640.91430.91810.90890.97080.8316
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Angara, S.; Tran, L.; Kim, J. Glaucoma Classification Using a NFNet-Based Deep Learning Model with a Customized Hybrid Attention Mechanism. Diagnostics 2026, 16, 815. https://doi.org/10.3390/diagnostics16050815

AMA Style

Angara S, Tran L, Kim J. Glaucoma Classification Using a NFNet-Based Deep Learning Model with a Customized Hybrid Attention Mechanism. Diagnostics. 2026; 16(5):815. https://doi.org/10.3390/diagnostics16050815

Chicago/Turabian Style

Angara, Sandeep, Loc Tran, and Jongwoo Kim. 2026. "Glaucoma Classification Using a NFNet-Based Deep Learning Model with a Customized Hybrid Attention Mechanism" Diagnostics 16, no. 5: 815. https://doi.org/10.3390/diagnostics16050815

APA Style

Angara, S., Tran, L., & Kim, J. (2026). Glaucoma Classification Using a NFNet-Based Deep Learning Model with a Customized Hybrid Attention Mechanism. Diagnostics, 16(5), 815. https://doi.org/10.3390/diagnostics16050815

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop