Next Article in Journal
Anthropometric Values in Spanish Elite Soccer: Differences between Divisions and Playing Positions
Previous Article in Journal
Dynamic Response of Slender Vertical Structures Subjected to Thunderstorm Outflows
Previous Article in Special Issue
Biomac3D: 2D-to-3D Human Pose Analysis Model for Tele-Rehabilitation Based on Pareto Optimized Deep-Learning Architecture
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Artificial Intelligence-Driven Eye Disease Classification Model

by
Abdul Rahaman Wahab Sait
Department of Documents and Archive, Center of Documents and Administrative Communication, King Faisal University, P.O. Box 400, Hofuf 31982, Al-Ahsa, Saudi Arabia
Appl. Sci. 2023, 13(20), 11437; https://doi.org/10.3390/app132011437
Submission received: 18 September 2023 / Revised: 12 October 2023 / Accepted: 17 October 2023 / Published: 18 October 2023
(This article belongs to the Special Issue Deep Neural Networks for Smart Healthcare Systems)

Abstract

:
Eye diseases can result in various challenges and visual impairments. These diseases can affect an individual’s quality of life and general health and well-being. The symptoms of eye diseases vary widely depending on the nature and severity of the disease. Early diagnosis can protect individuals from visual impairment. Artificial intelligence (AI)-based eye disease classification (EDC) assists physicians in providing effective patient services. However, the complexities of the fundus image affect the classifier’s performance. There is a demand for a practical EDC for identifying eye diseases in the earlier stages. Thus, the author intends to build an EDC model using the deep learning (DL) technique. Denoising autoencoders are used to remove the noises and artifacts from the fundus images. The single-shot detection (SSD) approach generates the key features. The whale optimization algorithm (WOA) with Levy Flight and Wavelet search strategy is followed for selecting the features. In addition, the Adam optimizer (AO) is applied to fine-tune the ShuffleNet V2 model to classify the fundus images. Two benchmark datasets, ocular disease intelligent recognition (ODIR) and EDC datasets, are utilized for performance evaluation. The proposed EDC model achieved accuracy and Kappa values of 99.1 and 96.4, and 99.4 and 96.5, in the ODIR and EDC datasets, respectively. It outperformed the recent EDC models. The findings highlight the significance of the proposed EDC model in classifying eye diseases using complex fundus images. Healthcare centers can implement the proposed model to improve their standards and serve a more significant number of patients. In the future, the proposed model can be extended to identify a comprehensive range of eye diseases.

1. Introduction

The eye is one of the most important organs for routine activities [1,2,3]. Eye diseases can cause permanent damage to the retina [4]. Eye disorders can result in visual impairment or blindness, making it difficult for individuals to carry out activities, including reading, driving, recognizing faces, and navigating the environment [4]. Thus, visual impairment can affect an individual’s quality of life. On a global scale, a minimum of 2.2 billion individuals are experiencing some form of visual impairment [4]. At least one billion of these cases could be prevented, or have failed to receive adequate treatment. The global cost of vision impairment is estimated at USD 411 billion annually, representing a significant burden on global economies [4]. Across the globe, only 36% of patients with vision impairment owing to refractive error and 17% of individuals with cataracts received adequate treatments [4].
Children with severe visual impairment may experience motor, verbal, emotional, social, and cognitive development delays [5,6,7,8]. Low academic achievement is a common problem for school-aged children who are visually impaired. Traditional treatment options such as those for dry eye, conjunctivitis, and blepharitis could not provide a solution for visual impairment. Additionally, these treatments can cause discomfort and pain [9]. Treating these disorders aims to reduce symptoms and prevent disease progression. Vision rehabilitation has demonstrated significant efficacy in enhancing functionality among those experiencing permanent vision loss from various eye disorders, including diabetic retinopathy (DR), glaucoma, trauma-related effects, and age-related macular degeneration [10].
Medical imaging encompasses diverse technologies to provide visual representations of the internal structures of the human body [11]. It is essential for non-invasive assessments of anatomical structures and physiological processes [12]. Healthcare experts diagnose, monitor, and treat diseases using these images. AI-based tools and technologies enable healthcare centers to treat ocular disorders effectively [13]. These technologies can analyze medical imagery, including retinal images and optical coherence tomography (OCT) scans, to identify signs of eye disorders, including diabetic retinopathy, glaucoma, and age-related macular degeneration in early stages [14].
Ophthalmologists can employ artificial intelligence (AI) to review patient images and provide consultations remotely [15,16,17,18]. Additionally, the AI application enables the remote monitoring of patients’ eye problems, eliminating the requirement for frequent in-person checkups [19]. AI algorithms can categorize patients into different risk groups by evaluating their ocular conditions’ severity [20]. This information assists healthcare centers in allocating resources to provide prompt care for patients at a higher risk. AI applications can assess various types of patient data, including medical history, genetic information, and medical imaging, to provide tailored treatment recommendations [21]. This practice guarantees that interventions are customized to address each patient’s unique requirements and attributes.
Recently, Image-based Convolutional Neural Network (CNN) models have been used to classify eye diseases using retinal, fundus, and OCT scans [22]. The initial stage is the collection of an extensive dataset of medical images encompassing multiple ocular pathologies. Medical imaging machinery or datasets facilitate the acquisition of visual representations of ocular diseases. The CNN model is trained using a labeled dataset with a supervised learning methodology [23]. During training, the model can discern patterns, characteristics, and textures within the images that serve as biomarkers for various ocular ailments [24]. After optimization, the model undergoes testing on an independent dataset to assess its performance in real-time [25].
Researchers and developers are focusing on improving the efficiency of the CNN models to overcome the existing challenges. A substantial number of parameters and high-performance computing resources are required to train an eye disease classification (EDC) model. Furthermore, the model’s performance is based on the dataset. Data imbalances may reduce the performance of the EDC. In order to address the shortcomings, a practical hyper-parameter tuning approach is required. The existing knowledge gap in the literature has motivated the authors to build a lightweight CNN model for classifying eye diseases with less computational power. Thus, the authors intend to build a lightweight EDC using the fundus images. The novelty of the study can be summarized as follows.
i.
A feature extraction technique for extracting crucial patterns from fundus images.
The existing feature extraction techniques demand substantial computational resources. These techniques may not capture the semantic meaning of the features. The proposed single-shot detection (SSD)-based feature extraction technique uses multiple feature maps at different scales to detect meaningful features with limited computational resources. It improves the performance of the proposed EDC.
ii.
A feature selection technique for generating compelling features.
The more recent feature selection techniques are limited to specific problems or datasets. They may not identify the key features. The proposed whale optimization algorithm (WOA)-based feature selection technique uses the Levy Flight and Wavelet search strategy to find the features. It assists the proposed model in identifying eye diseases with optimal accuracy.
iii.
An optimized CNN multi-class EDC model to classify ocular diseases with limited computational resources.
The remaining part of the study is organized as follows: Section 2 presents the recent techniques for detecting eye diseases. The research methodology of feature generation and multi-class image classification is described in Section 3. The experimental results are highlighted in Section 4. Section 5 shows the comparative analysis outcomes and outlines the study’s contribution. Finally, Section 6 presents the effects of the study in classifying the fundus images. In addition, it highlights the future aspects of the proposed study.

2. Literature Review

The conventional classification of eye illnesses depends on clinical observations, symptom evaluation, and diagnostic testing by ophthalmologists and optometrists [25]. The procedure commences with a thorough clinical assessment of the patient’s ocular condition. During this assessment, medical professionals specializing in ophthalmology or optometry observe and analyze observable indications and manifestations of visual disorders [26]. Tonometry is a well-known tool for detecting glaucoma by measuring intraocular pressure [26]. Ophthalmoscopy examines the retina, optic nerve, and blood vessels. Fundus photography offers high-resolution retinal images [27]. The utilization of these images has immense value in the diagnosis and continual monitoring of retinal disorders. The method of Fluorescein Angiography entails the intravenous administration of a fluorescent dye, which helps to emphasize the blood vessels within the retina [27]. This technique diagnoses retinal vascular disorders such as DR and retinal vein occlusion. OCT is a non-invasive imaging modality that offers high-resolution cross-sectional images of the retina and optic nerve [28]. Electroretinography is a technique used to assess the electrical responses of the retina in response to light stimulation [29]. This diagnostic tool is utilized to identify retinal conditions, specifically hereditary retinal diseases.
Image classifiers based on deep learning (DL) techniques are valuable in classifying eye diseases [30]. DL models are employed to extract relevant data from medical images, including retinal and OCT scans, and subsequently categorize these images. These features may include blood vessels, lesions, or retinal structural abnormalities in eye disease categorization. DL models can identify disorders such as DR, age-related macular degeneration, glaucoma, and cataracts [31]. These models can be especially advantageous in cases wherein a patient presents with many concurrent ocular conditions. Consistency in diagnosis and categorization can be achieved using DL models [32].
The existing studies have proposed an ocular disease dataset and classification techniques. Li et al. (2021) [33] presented a dataset of 10,000 fundus images. They evaluated the state-of-the-art DL techniques using the dataset. They emphasized the significance of an effective DL technique in classifying the fundus images. Sundaram et al. (2019) [34] proposed a model to extract the blood vessels from the images. They followed the hybrid segmentation approach for the classification. He et al. (2021) [35] proposed a multi-label disease classification model. They applied a dense correlation network for classifying the ocular diseases. In their study [36], Gour and Khanna (2021) proposed a multi-class image classification method. They followed two transfer learning-based models to identify abnormal patterns. Junayed et al. (2021) [37] employed the hyper-parameter-optimized CNN model for detecting cataracts. Xu et al. (2021) [38] proposed a DL-based model to detect glaucoma. They trained the model using a small set of training samples. Luo et al. (2021) [39] proposed an ophthalmic disease detection model using a novel mixture loss function. Shamsan et al. (2023) [40] proposed a hybrid CNN model for eye disease classification (EDC). They used principal component analysis for feature selection.
Researchers increasingly confront significant gaps and challenges in developing DL-based EDC. There is a requirement for large-scale, multicenter clinical trials to evaluate the efficacy of deep learning models in real-world healthcare settings, regardless of their promising results in research and small-scale clinical settings. Models trained on data from a specific population may exhibit limited generalizability when applied to different ethnic, age, or gender cohorts. DL models can interpret images rapidly. Nevertheless, computational resources, medical device integration, and regulatory approvals limit their real-time diagnosis. Finding strategies to apply findings from academic studies in real-world settings is crucial. DL models demand consistent benchmark datasets for a wide range of eye disorders. This would allow for more reliable comparisons and assessments of unique DL models. Limited research has been conducted evaluating the cost-effectiveness of DL algorithms for categorizing eye diseases in contrast to conventional diagnostic approaches. Thus, the authors focus on developing cost-effective EDC using DL techniques.

3. Materials and Methods

The study intends to build a lightweight DL model for identifying eye diseases in the earlier stages. The fundus images may contain noises and black spaces. Thus, image pre-processing is required to improve the standard of the images. A high-quality image can support the proposed model in identifying the key indicators of eye diseases. In addition, a dedicated feature extraction technique offers a set of features. Using these features, the proposed model can generate an outcome with limited computational resources. It comprises image pre-processing, feature extraction and selection, and image classification techniques.
SSD uses low-level features to capture minor details and high-level features to represent abstract information [41]. By combining these two methods, the proposed EDC model may be trained to recognize eye diseases with a wide range of characteristics. The feature extraction procedure of SSD incorporates location information, which is highly significant [33]. This information can be beneficial in precisely localizing crucial features of the fundus images. Assessing high-dimensional data can involve significant computing costs and may result in overfitting. WOA plays an essential role in identifying the subset of features with the highest clarity level [42]. Noise reduction and the elimination of extraneous information facilitates the concentration of the machine learning model on the crucial elements of the data, consequently potentially enhancing accuracy and improving generalization capabilities. WOA-based feature selection may contribute to the proposed EDC model’s optimal accuracy. ShuffleNet V2 demonstrates notable efficiency in both computational and memory usage [43]. The model attains exceptional performance in image classification with reduced computational resources. ShuffleNet V2 is well-suited for distributed processing or deployment on hardware with many processing units, as its architecture encourages efficient parallelism. It facilitates the deployment of proposed EDC on mobile devices and other low-powered endpoints. Thus, the authors employ the SSD technique, WOA, and ShuffleNet V2 for classifying the fundus images. Figure 1 presents the proposed model.

3.1. Image Acquisition

The authors obtained the fundus images from two benchmark datasets. The ocular disease intelligent recognition (ODIR) dataset offers left and right eye images of 5000 patients. It is available in the Kaggle repository [44]. Shanggong Medical Technology Co., Ltd., China, gathered the images from various hospitals across China. The images were captured using Canon, Zeiss, and Kowa cameras. Thus, the image resolutions are not unique. The patients were classified into eight classes: normal, DR, glaucoma, cataracts, age-related macular degeneration, hypertension, pathological myopia, and other diseases/abnormalities.
Furthermore, the EDC dataset was used to train the proposed model. It is publicly available in the repository [45]. The images were collected from multiple repositories. The dataset provides four categories: normal, glaucoma, DR, and cataract. The proposed model classifies the images into four classifications based on the image categories of the EDC dataset. The ground truth information is available for the ODIR and EDC datasets. Table 1 highlights the properties of the datasets. In addition, Figure 2 shows the sample images of the ODIR and EDC datasets.

3.2. Pre-Processing and Transformation

In order to pre-process the images, the authors applied contrast-limited adaptive histogram equalization (CLAHE), denoising autoencoders, and image transformations. The contrast value of each image differs based on the camera resolution. The CLAHE function improves the image contrast locally to overcome the different lighting conditions. In addition, it adapts to the characteristics of the multiple regions to cover a broader range of fundus images. Let I i be the fundus images, and CLAHE be the function. Equation (1) expresses the computational model of image contrast enhancement.
I i = C L A H E I i ,   i = 1 , , n
where n is the number of images.
Furthermore, the authors employed denoising autoencoders to remove the noises and artifacts. Denoising autoencoders are based on an AI technique that learns the image representations. The encoder converts the image into a lower dimension, whereas the decoder reconstructs the high-quality, noiseless images from the lower dimension. Equations (2) and (3) show the encoding and decoding processes.
I L i = E n c o d e   ( I i ) ,   i = 1 , , , n
I C i = D e c o d e ( I i ) ,   i = 1 , , , n
where I L i is the low-dimensional image, I C i is the high-quality image, and n is the number of images.
The images were resized into 512 × 512 pixels to support the feature extraction. To overcome the data imbalance, the images were transformed into different angles.

3.3. Feature Extraction

The feature extraction process in CNN entails the utilization of convolutional filters to scan the input data and produce feature sets with specific patterns and characteristics. These feature sets retain the input’s hierarchical information, allowing the proposed model to recognize and learn complicated patterns and structures essential for EDC. SSD [46] is an efficient object-identification technique that uses a single forward pass of a neural network to detect and localize objects in the fundus image. It utilizes a deep CNN as its core architecture. Features are extracted from the input image at different scales and resolutions using the detection heads. Feature maps are derived from different network layers, wherein each layer captures distinct levels of detail and abstraction. SSD introduces anchor boxes at feature map locations. The anchor boxes are predetermined concerning their dimensions and aspect ratios. Based on the feature map location, the detection head predicts two blocks of information for a single anchor box. Figure 3 outlines the process of feature extraction.
Moreover, a class score is used to identify the key feature. SSD integrates the feature maps at different ranges to cover the maximum size and aspect ratio. The confidence score is employed to ensure the existence of a crucial pattern. Equation (4) shows the feature set generation using the SSD technique.
F S = S S D I i , i = 1 , , , n
where F S is the feature set, SSD is the function for extracting the features, and n is the total number of images.

3.4. Feature Selection

The nature-inspired WOA is widely applied for addressing optimization problems. The concept is derived from the social behaviors shown by humpback whales, particularly their bubble-net hunting strategy [47]. The primary purpose of WOA is to identify the most favorable solution or collection of alternatives that either decrease or maximize a specified objective function. The authors applied WOA to select the key features to determine diseases using fundus images.
The algorithm integrates aspects of exploration and exploitation, emulating the collective hunting behavior of humpback whales. During exploration, whales move forward randomly in a specific search space. This action aims to promote the investigation of unexplored areas within the search domain. During the exploitation phase, whales modify their locations by considering the positions of other whales closer to the best solution. This phase aims to direct the algorithm’s convergence toward the best solution. The repositioning of whales is determined by considering their prior locations, the positions of other whales, and their respective fitness values. The algorithm concludes its execution upon reaching the maximum number of iterations or satisfying predetermined stopping criteria. The whale’s response determines the ideal solution to the optimization issue.
The algorithm effectively combines the investigation of unknown images with exploitation to navigate complicated optimization landscapes and swiftly find the optimum set of features. The method’s performance is contingent upon both the parameter configurations and the inherent characteristics of the optimization issue. The potential solutions (whales) population is randomly initialized within the search space. The search agent position is updated related to the best solution. To achieve a better solution, the authors introduced the Levy Flight and Wavelet [37] mutation strategy into exploration to improve the WOA performance. Initially, the WOA considers the best feature set as the target prey. Figure 4 highlights the feature selection process.
The search agents update their positions based on the best search agent. Equations (5) and (6) compute the position vector of the optimal feature sets.
E = C X t X ( t )
X t + 1 = X * t B . E
where t represents the current iteration, B , C , a n d E are the co-efficient vectors, X * is the position vector of optimum feature sets, “.” is the element-wise multiplication operator, and X is the position vector.
In the exploitation phase, the feature sets are generated using Equation (7).
X t + 1 = X t B . E , i f   P < 0.5 E . e b l . C o s 2 π l + X t , i f   P 0.5
where E indicates the distance of the ith whale to the prey, b is the constant in the exploitation phase, X t + 1 is the position vector at t + 1 time, and l and P are random numbers in the range [−1,1].
Equations (8) and (9) show the mathematical model for the exploration phase.
E = C . X r a n d X
X t + 1 = X r a n d t B . E
where X r a n d is the random position vector.
The random positions of the whales are computed using the Levy Flight and Wavelet search strategies. Equation (10) calculates the random position of the agent to obtain the best feature set.
X r a n d t + 1 = L e v y   X t , d i m + σ u b X t , σ > 0   L e v y   X t , d i m + σ X t l b , σ 0  
where L e v y ( ) is the Levy Flight function using X t , dim is the dimension of the feature sets, ub and lb are the upper and lower bounds, and σ is the wavelet function. Equation (11) shows the mathematical model of the wavelet function.
σ = 1 a e ( φ a ) 2 / 2 C o s φ a
where φ is the pseudo-random number in the range [−2.5a–2.5a] and a is the scale parameter of the wavelet function.

3.5. Eye Disease Classification

The ShuffleNetV2 model has superior model inference efficiency and lower processing needs, and achieves comparable performance in picture classification and other computer vision applications [43]. The architecture enables fast deep learning inference, focusing on resource-constrained devices, including mobile phones and handheld devices. Thus, the authors applied the ShuffleNet V2 model for classifying the fundus images.
The ShuffleNet model aims to enhance the balance between model accuracy and computing efficiency. It utilizes channel shuffling and pointwise group convolutions to decrease computational complexity. It presents a unique architectural component known as the shuffling unit. The utilization of this block simplifies the process of channel shuffling. It facilitates the interchange of information across different groups of channels within a feature map. Figure 5 elucidates disease classification using the ShuffleNet V2 model.
Shuffle units are employed to construct network topologies of greater complexity while preserving computational efficiency. Channel shuffling plays a pivotal role as a significant innovation inside the ShuffleNet V2 architecture. It enables the division of channels into various groups and the subsequent interleaving or shuffling of channels within groups. Channel shuffling improves the network’s representation capacity and ability to learn varied features without raising parameters. ShuffleNet V2 utilizes pointwise group convolutions, a technique that entails the application of 1 × 1 convolutions to distinct groups of channels inside a feature map. Group convolutions reduce computational complexity compared to conventional convolutions while maintaining the same level of expressive capability. The inverted residual blocks are used to enhance the network’s depth and effectively capture hierarchical information. Inverted residuals incorporate depth-wise separable and pointwise convolutions with expansion and projection layers.
Equation (12) highlights the image classification model using the ShuffleNet V2 model.
E D C = S h u f f l e N e t   V 2 R e L u + C L + F C + R e L u + F C + s o f t m a x ( F s )
where ReLu is the rectified linear unit, CL is the convolutional layer, FC is the fully connected layer, and softmax is the function for multi-class classification.

3.6. Hyperparameter Optimization

The AO algorithm dynamically adjusts the learning rate of the proposed EDC throughout the training process. It enables the system to achieve convergence faster and with greater reliability. In addition, it automatically modifies the learning rate based on the development progress. The automated modification minimizes the necessity for the manual fine-tuning of learning rates. ShuffleNet V2 includes a width multiplier (α) as a hyper-parameter, enabling the manipulation of channel quantities within each layer. A smaller α leads to a narrower network with lower computing costs. It is structured hierarchically, with distinct stages comprising shuffle units and convolutional layers. The authors applied AO to alter the number of settings and shuffling units to improve the performance of the proposed model. Equation (13) outlines the mathematical form of the fine-tuned proposed EDC.
E D C = A O ( S h u f f l e N e t   V 2 h y p e r p a r a m e t e r s

3.7. Evaluation Metrics

In order to evaluate the performance of the proposed EDC, the author follows the benchmark metrics. Accuracy is a measure of correct predictions made by the proposed EDC. In the context of fundus image analysis, precision and recall are crucial metrics to measure the influence of false positives (FP) and false negatives (FN) in the outcome. The F1-score is used to find the relationship between precision and recall. Sensitivity computes the model’s ability to identify true positives (TP), whereas specificity measures the EDC’s ability to find true negatives (TN). Equations (14)–(19) show the expressions for computing accuracy, precision, recall, F1-score, specificity, and sensitivity. These metrics are used to determine the model’s effectiveness in classifying eye diseases. In addition, Equation (20) highlights the mathematical form of Cohen’s Kappa. This metric evaluates the multi-class classification capability of EDC.
A c c u r a c y = T P + T N T P + T N + F P + F N
P r e c i s i o n = T P T P + F P
R e c a l l = T P T P + F N
F 1 S c o r e = 2 × P r e c i s i o n × R e c a l l P r e c i s i o n + R e c a l l
S e n s i t i v i t y = T P F N + T P
S p e c i f i c i t y = T N T N + F P
The confidence interval (CI) and standard deviation (SD) are applied to calculate the model’s uncertainties. In addition, the prediction errors are computed using mean average deviation (MAD) and root mean square error (RMSE). Equations (20) and (21) outline the mathematical form of MAD and RMSE,
M A D = P μ N
where P is the data point, μ is the mean, and N is the total number of images.
R M S E = 1 N i = 1 N K K ^ 2
where K is the measure of the predicted class, K ^ is the predicted class, and N is the number of images.
Furthermore, the author employed learning rate, FLOPs, number of parameters, and testing time for measuring the computational strategies of EDC to classify the fundus images.

4. Results

The authors implemented the proposed EDC mode in Python 3.8.3, Windows 10 Professional, 16 GB RAM, Intel core i7, NVIDIA GeForce RTX 3050. In order to train and test the EDC model, the authors divided the dataset into train (70%) and test (30%) sets. The source code of the ShuffleNet V2 and SSD was extracted from the repositories [38,39]. The authors accordingly executed the EDC model with batch sizes of 30 and 43 and epochs of 220 and 360 for the ODIR and EDC datasets. AO was used to identify the optimal parameters of the ShuffleNet V2 model in classifying the eye disease images. Table 2 presents the computational cost of the existing and proposed EDC models. Hyperparameter optimization supported the proposed model in classifying the fundus images with lower learning rates, parameters, and floating point operations (FLOPs). The proposed model maintained the trade-off between the learning rate and the training time. The AO algorithm streamlined the parameters of the ShuffleNet V2 model to obtain an exceptional outcome.
Table 3 shows the findings of the proposed models’ performance evaluation in the ODIR dataset. SSD extracted the appropriate features and supported the proposed model in overcoming the complexities in classifying the fundus images. The WOA optimization algorithm with Levy Flight and Wavelet mutation assisted the proposed EDC in detecting the diseases effectively. In addition, the optimized ShuffleNet V2 model-based multi-class classifier identified each class with optimal accuracy.
Table 4 highlights the outcome of the performance analysis of the proposed EDC model in the EDC dataset. The suggested image processing techniques produced high-quality images. The EDC dataset contains unique images that differ in size and pixel rates. However, the image pre-processing technique overcame the shortcomings and removed the noises and artifacts from the images. It supported the classifier in obtaining better results for each class. Figure 6 presents the findings of the performance analysis.
Table 5 reveals the findings of EDC generalization on the ODIR dataset. The proposed EDC outperformed the existing state-of-the-art techniques. The higher values of precision, recall, and F1-score highlight the significance of the proposed EDC model. The proposed model efficiently handled the true positives, true negatives, false positives, and false negatives. Figure 7 shows the outcomes of comparative analysis for the ODIR dataset.
Likewise, Table 6 shows the outcome of the EDC generalization using the EDC dataset. The proposed EDC model obtained a superior outcome and addressed the limitations of the existing EDC. The proposed model achieved a higher Kappa value compared to the current models. The findings suggest the efficiency of the proposed multi-class classification model. Figure 8 highlights the results of the generalization of EDC on the EDC dataset.
Table 7 reveals the MAD and RMSE values of the proposed EDC model. The lower value of MAD represents the importance of the suggested EDC in classifying the real-time images. It is evident that the proposed model predicted multi-class images with high accuracy and lower error rates. In addition, the RMSE value shows that the proposed EDC model can handle the outliers and produce an optimal outcome.
Lastly, Table 8 highlights the CI and SD of the EDC. The outcome highlights the significance of the proposed model in dealing with uncertainties. It indicates the reliability of the proposed model’s decisions. In addition, it shows that the model can deliver a helpful outcome in the real-time environment.

5. Discussion

The authors proposed a DL-based EDC for classifying the fundus images in this study. CLAHE and denoising autoencoders were employed to improve the image quality. The feature extraction was performed using the SSD technique. An improved WOA was employed to select the key features. The ShuffleNet V2 model was fine-tuned in order to classify the fundus images.
In medical image processing pipelines, image pre-processing is a crucial step. The utilization of image enhancement techniques enhances the visual quality of medical images. Additionally, it plays an essential role in improving the accuracy and dependability of subsequent analysis, diagnosis, and clinical decision-making processes. The author applied the CLAHE to cover the image’s maximum range. The denoising autoencoders technique helped the study to produce high-quality images using encoders and decoders. The proposed data augmentation technique increased the diversity of training data in order to improve the performance of the proposed EDC model. In feature extraction and selection, the SSD technique is used to extract features from complex images. The detection head technique has assisted the proposed model in extracting the key patterns of eye diseases. The minor spots were identified during this process. In addition, WOA, with the Levy Flight and Wavelet strategy, selected the critical features for classifying eye diseases.
Furthermore, the ShuffleNet V2 model was fine-tuned using the AO algorithm. The ShuffleNet V2 architecture contains fewer convolutional layers that help the proposed model to produce an effective outcome with limited resources. Moreover, the feature extraction and selection approaches assisted the ShuffleNet V2 model in generating the results with minimal convolutional layers.
Table 2 shows the significance of the proposed EDC in classifying the fundus images. The proposed feature selection process minimized the features and assisted the proposed EDC in identifying eye diseases with fewer computational resources. Due to the hyper-parameter optimization and feature selection, the proposed EDC required a lower number of parameters and FLOPs with a learning rate of 1 × 10 4 . Table 3 and Table 4 outline the performance evaluation of the proposed EDC. The proposed EDC achieved an accuracy of 99.1 and 99.4 for ODIR and EDC datasets, respectively. Compared to the EDC dataset, the ODIR dataset comprises fewer images. However, the proposed EDC performed exceptionally on the EDC and ODIR datasets.
Table 5 and Table 6 show the findings of the comparative analysis. The proposed EDC outperformed the recent EDC models by achieving optimal accuracy and sensitivity. The proposed feature extraction and selection techniques played a significant role in improving the efficiency of the proposed model. Table 7 highlights the outcome of the loss function. The AO hyper-parameter optimization fine-tuned the shuffleNet V2 model’s parameters to overcome the challenges in classifying the fundus images with fewer FLOPs and parameters. Finally, Table 8 outlines the statistical significance of the proposed EDC. The proposed EDC generated outcomes with SD values of 0.0021 and 0.0091 for the ODIR and EDC datasets. The findings outline the capability of the proposed EDC to deliver a high-standard outcome.
He et al. [35] employed a spatial correlation technique for extracting features. The technique used pixel-wise comparison for feature extraction. A dense correlation network was used for the image classification. In contrast, the author used an effective image pre-processing technique to improve the image quality. Moreover, the proposed model achieved a superior accuracy compared to the He et al. model.
Junayed et al. [37] developed a deep neural network model for detecting cataracts from fundus images. They optimized the network using AO. Similarly, the proposed EDC model applied AO for tuning the hyper-parameters of the ShuffleNet V2 model. However, the proposed EDC model outperformed the Junayed et al. model.
Xu et al. [38] followed a hierarchical DL approach for glaucoma diagnosis. They employed cup-to-disk ratio vectors for the image classification. The larger number of vectors reduced the performance of the classifier. On the other hand, the author followed the feature extraction and selection approaches for generating the feature sets and improved the overall prediction accuracy of the proposed EDC model. Luo et al. [39] proposed DL-based ophthalmic disease detection. They used focal loss and correntrophy-induced loss function to overcome the data imbalances of the complex datasets. In contrast, the author employed data augmentation and compelling image pre-processing approaches to address data bias.
Shamson et al. [40] employed principal component analysis for feature extraction. They used MobileNet and DenseNet 121 models to fuse the features to classify the images. On the other hand, the authors employed the lightweight version of the ShuffleNet model and obtained exceptional prediction accuracy.
The authors addressed the challenges in classifying the fundus images. The existing studies required substantial computational resources to obtain a reasonable outcome. In contrast, the proposed research achieved an exceptional outcome with limited resources. The generalization of the proposed method was performed on ODIR and EDC datasets, respectively. The findings reveal the significance of the proposed model in EDC. The proposed model can be implemented in remote places. In addition, it can be developed as a mobile application and assist clinicians in treating patients.
The proposed feature extraction and EDC play a significant role in the automation of diagnosing and evaluating different eye conditions. The feature extraction process discovers and extracts pertinent information from fundus images for classifying ocular diseases. It minimizes data dimensionality while maintaining essential data. The extracted feature enables the machine learning model to prioritize necessary data, consequently enhancing the manageability of the categorization operation. The suggested EDC detects eye diseases, including diabetic retinopathy, glaucoma, and cataracts. The proposed model can identify particular patterns or irregularities within the eye, such as detecting exudates or hemorrhages. In healthcare settings, these methods may improve the speed and precision of diagnosing and monitoring eye diseases.
The proposed model has a substantial impact on EDC. It was trained to recognize critical patterns, structures, and anomalies in retinal or fundus images to classify eye diseases. It can capture hierarchical and abstract characteristics by employing several layers of convolution, pooling, and non-linear activation functions. The basic features, including edges, textures, and shapes, were utilized to train the proposed model. The proposed classifier is data-driven and can learn from massive annotated image datasets. This ability enables the proposed model to classify novel data without additional training, which is crucial in medical imaging applications. The ShuffleNet model was pre-trained on large, general-purpose datasets. However, the authors fine-tuned the model with AO. The ShuffleNet architecture exhibits scalability and efficiently analyzes a substantial quantity of retinal data. This scalability is significant for healthcare providers and screening programs handling a massive volume of patients. The proposed model facilitates the continuous monitoring of ocular diseases. It can monitor the retinal images over time and notify physicians of any modifications that may indicate a demand for treatment. In addition, it can assist in driving forward research efforts in medicine.
While implementing the proposed model, the authors encountered a few limitations. Identifying small lesion patterns of ocular disorders may present difficulties for the proposed EDC due to image noise or other anatomical features. The datasets were highly imbalanced. In the ODIR dataset, the numbers of glaucoma and cataracts images were lower than those of the remaining diseases. However, the authors addressed the limitation using the effective data augmentation technique. The proposed EDC may face challenges in classifying multiple ocular disorders that may present in a single image. The quality of the fundus images was poor. Images were unique in size and clarity. Extensive image processing was required to normalize the images. The proposed model was generalized using two benchmark datasets. Implementing CNN models with minimal latency may be technically challenging. The proposed EDC may demand additional training in order to classify the real-time images.
In the future, the model’s overall performance can be improved using ensemble techniques, including boosting or bagging, oversampling the minority class, undersampling the majority class, or using loss functions involving the minority class to address class imbalance. In order to mitigate overfitting and enhance the generalization capabilities of the model, regularization methods such as dropout, L1/L2 regularization, or batch normalization can be employed. To build clinician trust, interpretable DL models can be used to justify the proposed model decisions.

6. Conclusions

In this study, the authors addressed the limitations of the existing EDC. A DL-based EDC was proposed to identify eye diseases from complex retinal images. The proposed image pre-processing approach produced high-quality images. The denoising autoencoders were employed to generate the images using encoder and decoder techniques. The authors applied the SSD approach for feature extraction and employed WOA with the Levy Flight and Wavelet mutation strategies to select critical features. In addition, an optimized ShuffleNet V2 model was employed for EDC. The author generalized the proposed EDC model on the ODIR and EDC datasets. The findings of generalization highlight that there is a significant improvement in the proposed model. The proposed model obtained optimal accuracy values of 99.1 and 99.4, and sensitivity of 98.9 and 98.7, for the ODIR and EDC datasets, respectively. It outperformed the recent models using the learning rate of 1 × 10 4 , parameters of 23.5 m and 27.4 m, FLOPs of 0.78 G and 0.68 G, and testing times of 1.08 s and 1.15 s for the ODIR and EDC datasets. The findings indicate that the proposed EDC can be implemented in healthcare centers with minimal hardware and software requirements. The experimental outcome suggests that the model can be deployed in healthcare centers across the globe. The proposed EDC has significant practical implications and real-world applications in ophthalmology and healthcare. It can potentially aid in the early detection of a range of ocular conditions, including DR, cataracts, and glaucoma. Screening programs for the general public can benefit from adopting the proposed EDC. The proposed model is beneficial in areas with limited expert eye care, as it enables the screening of huge populations at minimal cost. It can play a role in identifying potential drug candidates and assessing the effectiveness of new treatments for eye diseases. Implementing the proposed EDC can enhance research endeavors by facilitating the analysis of extensive datasets. However, the proposed model encountered a few limitations in classifying the fundus images. The highly imbalanced datasets caused challenges in pre-processing the fundus images. There is a demand for compelling image pre-processing methods to overcome image noises and artifacts. In the future, the authors will focus on improving the performance of the proposed EDC with additional eye diseases. The ensembling technique can be employed to improve the model’s performance.

Funding

This work was supported by the Deanship of Scientific Research, Vice Presidency for Graduate Studies and Scientific Research, King Faisal University, Saudi Arabia (Grant No. 4257).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

ODIR Dataset. Available online: https://www.kaggle.com/datasets/andrewmvd/ocular-disease-recognition-odir5k (accessed on 5 January 2023). EDC Dataset. Available online: https://www.kaggle.com/datasets/gunavenkatdoddi/eye-diseases-classification (accessed on 1 January 2023).

Conflicts of Interest

The author declares no conflict of interest.

References

  1. Chin, Y.H.; Ng, C.H.; Lee, M.H.; Koh, J.W.H.; Kiew, J.; Yang, S.P.; Sundar, G.; Khoo, C.M. Prevalence of thyroid eye disease in Graves’ disease: A meta-analysis and systematic review. Clin. Endocrinol. 2020, 93, 363–374. [Google Scholar] [CrossRef] [PubMed]
  2. Sarki, R.; Ahmed, K.; Wang, H.; Zhang, Y. Automated detection of mild and multi-class diabetic eye diseases using deep learning. Health Inf. Sci. Syst. 2020, 8, 32. [Google Scholar] [CrossRef] [PubMed]
  3. Al-Aswad, L.A.; Elgin, C.Y.; Patel, V.; Popplewell, D.; Gopal, K.; Gong, D.; Thomas, Z.; Joiner, D.; Chu, C.K.; Walters, S.; et al. Real-time mobile teleophthalmology for the detection of eye disease in minorities and low socioeconomics at-risk populations. Asia-Pac. J. Ophthalmol. 2021, 10, 461. [Google Scholar] [CrossRef] [PubMed]
  4. WHO Report. Available online: https://www.who.int/news-room/fact-sheets/detail/blindness-and-visual-impairment (accessed on 10 August 2023).
  5. Tong, Y.; Lu, W.; Yu, Y.; Shen, Y. Application of machine learning in ophthalmic imaging modalities. Eye Vis. 2020, 7, 1–15. [Google Scholar] [CrossRef] [PubMed]
  6. Balyen, L.; Peto, T. Promising artificial intelligence-machine learning-deep learning algorithms in ophthalmology. Asia-Pac. J. Ophthalmol. 2019, 8, 264–272. [Google Scholar]
  7. Alqudah, A.M. AOCT-NET: A convolutional network automated classification of multi-class retinal diseases using spectral-domain optical coherence tomography images. Med. Biol. Eng. Comput. 2020, 58, 41–53. [Google Scholar] [CrossRef]
  8. Li, T.; Bo, W.; Hu, C.; Kang, H.; Liu, H.; Wang, K.; Fu, H. Applications of deep learning in fundus images: A review. Med. Image Anal. 2021, 69, 101971. [Google Scholar] [CrossRef]
  9. Grassmann, F.; Mengelkamp, J.; Brandl, C.; Harsch, S.; Zimmermann, M.E.; Linkohr, B.; Peters, A.; Heid, I.M.; Palm, C.; Weber, B.H. A deep learning algorithm for prediction of age-related eye disease study severity scale for age-related macular degeneration from color fundus photography. Ophthalmology 2018, 9, 1410–1420. [Google Scholar] [CrossRef]
  10. Lin, W.C.; Chen, J.S.; Chiang, M.F.; Hribar, M.R. Applications of artificial intelligence to electronic health record data in ophthalmology. Transl. Vis. Sci. Technol. 2020, 9, 13. [Google Scholar] [CrossRef]
  11. An, G.; Omodaka, K.; Hashimoto, K.; Tsuda, S.; Shiga, Y.; Takada, N.; Kikawa, T.; Yokota, H.; Akiba, M.; Nakazawa, T. Glaucoma diagnosis with machine learning based on optical coherence tomography and color fundus images. J. Healthc. Eng. 2019, 2019, 4061313. [Google Scholar] [CrossRef]
  12. Ting, D.S.; Peng, L.; Varadarajan, A.V.; Keane, P.A.; Burlina, P.M.; Chiang, M.F.; Schmetterer, L.; Pasquale, L.R.; Bressler, N.M.; Webster, D.R.; et al. Deep learning in ophthalmology: The technical and clinical considerations. Prog. Retin. Eye Res. 2019, 72, 100759. [Google Scholar] [CrossRef] [PubMed]
  13. Soomro, T.A.; Afifi, A.J.; Zheng, L.; Soomro, S.; Gao, J.; Hellwich, O.; Paul, M. Deep learning models for retinal blood vessels segmentation: A review. IEEE Access 2019, 7, 71696–71717. [Google Scholar] [CrossRef]
  14. Sengupta, S.; Singh, A.; Leopold, H.A.; Gulati, T.; Lakshminarayanan, V. Ophthalmic diagnosis using deep learning with fundus images–A critical review. Artif. Intell. Med. 2020, 102, 101758. [Google Scholar] [CrossRef] [PubMed]
  15. Sreng, S.; Maneerat, N.; Hamamoto, K.; Win, K.Y. Deep learning for optic disc segmentation and glaucoma diagnosis on retinal images. Appl. Sci. 2020, 10, 4916. [Google Scholar] [CrossRef]
  16. González-Gonzalo, C.; Sánchez-Gutiérrez, V.; Hernández-Martínez, P.; Contreras, I.; Lechanteur, Y.T.; Domanian, A.; van Ginneken, B.; Sánchez, C.I. Evaluation of a deep learning system for the joint automated detection of diabetic retinopathy and age-related macular degeneration. Acta Ophthalmol. 2020, 98, 368–377. [Google Scholar] [CrossRef]
  17. Milea, D.; Najjar, R.P.; Jiang, Z.; Ting, D.; Vasseneix, C.; Xu, X.; Aghsaei Fard, M.; Fonseca, P.; Vanikieti, K.; Lagrèze, W.A.; et al. Artificial intelligence to detect papilledema from ocular fundus photographs. N. Engl. J. Med. 2020, 382, 1687–1695. [Google Scholar] [CrossRef]
  18. Sahlsten, J.; Jaskari, J.; Kivinen, J.; Turunen, L.; Jaanio, E.; Hietala, K.; Kaski, K. Deep learning fundus image analysis for diabetic retinopathy and macular edema grading. Sci. Rep. 2019, 9, 10750. [Google Scholar] [CrossRef]
  19. Wagner, S.K.; Fu, D.J.; Faes, L.; Liu, X.; Huemer, J.; Khalid, H.; Ferraz, D.; Korot, E.; Kelly, C.; Balaskas, K.; et al. Insights into systemic disease through retinal imaging-based oculomics. Transl. Vis. Sci. Technol. 2020, 9, 6. [Google Scholar] [CrossRef]
  20. Yan, Q.; Weeks, D.E.; Xin, H.; Swaroop, A.; Chew, E.Y.; Huang, H.; Ding, Y.; Chen, W. Deep-learning-based prediction of late age-related macular degeneration progression. Nat. Mach. Intell. 2020, 2, 141–150. [Google Scholar] [CrossRef]
  21. Thompson, A.C.; Jammal, A.A.; Medeiros, F.A. A review of deep learning for screening, diagnosis, and detection of glaucoma progression. Transl. Vis. Sci. Technol. 2020, 9, 42. [Google Scholar] [CrossRef]
  22. Juneja, M.; Singh, S.; Agarwal, N.; Bali, S.; Gupta, S.; Thakur, N.; Jindal, P. Automated detection of Glaucoma using deep learning convolution network (G-net). Multimed. Tools Appl. 2020, 79, 15531–15553. [Google Scholar] [CrossRef]
  23. Bellemo, V.; Lim, G.; Rim, T.H.; Tan, G.S.; Cheung, C.Y.; Sadda, S.; He, M.G.; Tufail, A.; Lee, M.L.; Hsu, W.; et al. Artificial intelligence screening for diabetic retinopathy: The real-world emerging application. Curr. Diabetes Rep. 2019, 19, 72. [Google Scholar] [CrossRef] [PubMed]
  24. Qiao, L.; Zhu, Y.; Zhou, H. Diabetic retinopathy detection using prognosis of microaneurysm and early diagnosis system for non-proliferative diabetic retinopathy based on deep learning algorithms. IEEE Access 2020, 8, 104292–104302. [Google Scholar] [CrossRef]
  25. Peng, Y.; Dharssi, S.; Chen, Q.; Keenan, T.D.; Agrón, E.; Wong, W.T.; Chew, E.Y.; Lu, Z. DeepSeeNet: A deep learning model for automated classification of patient-based age-related macular degeneration severity from color fundus photographs. Ophthalmology 2019, 126, 565–575. [Google Scholar] [CrossRef] [PubMed]
  26. Poplin, R.; Varadarajan, A.V.; Blumer, K.; Liu, Y.; McConnell, M.V.; Corrado, G.S.; Peng, L.; Webster, D.R. Prediction of cardiovascular risk factors from retinal fundus photographs via deep learning. Nat. Biomed. Eng. 2018, 2, 158–164. [Google Scholar] [CrossRef]
  27. Li, F.; Chen, H.; Liu, Z.; Zhang, X.; Wu, Z. Fully automated detection of retinal disorders by image-based deep learning. Graefe’s Arch. Clin. Exp. Ophthalmol. 2019, 257, 495–505. [Google Scholar] [CrossRef]
  28. Bajwa, M.N.; Malik, M.I.; Siddiqui, S.A.; Dengel, A.; Shafait, F.; Neumeier, W.; Ahmed, S. Two-stage framework for optic disc localization and glaucoma classification in retinal fundus images using deep learning. BMC Med. Inform. Decis. Mak. 2019, 19, 136. [Google Scholar]
  29. Kugelman, J.; Alonso-Caneiro, D.; Read, S.A.; Hamwood, J.; Vincent, S.J.; Chen, F.K.; Collins, M.J. Automatic choroidal segmentation in OCT images using supervised deep learning methods. Sci. Rep. 2019, 9, 13298. [Google Scholar] [CrossRef]
  30. Motozawa, N.; An, G.; Takagi, S.; Kitahata, S.; Mandai, M.; Hirami, Y.; Yokota, H.; Akiba, M.; Tsujikawa, A.; Takahashi, M.; et al. Optical coherence tomography-based deep-learning models for classifying normal and age-related macular degeneration and exudative and non-exudative age-related macular degeneration changes. Ophthalmol. Ther. 2019, 8, 527–539. [Google Scholar] [CrossRef]
  31. Gadekallu, T.R.; Khare, N.; Bhattacharya, S.; Singh, S.; Maddikunta, P.K.R.; Srivastava, G. Deep neural networks to predict diabetic retinopathy. J. Ambient. Intell. Humaniz. Comput. 2020, 2020, 5407–5420. [Google Scholar] [CrossRef]
  32. Ahn, J.M.; Kim, S.; Ahn, K.S.; Cho, S.H.; Lee, K.B.; Kim, U.S. A deep learning model for the detection of both advanced and early Glaucoma using fundus photography. PLoS ONE 2018, 13, e0207982. [Google Scholar] [CrossRef] [PubMed]
  33. Li, N.; Li, T.; Hu, C.; Wang, K.; Kang, H. A Benchmark of Ocular Disease Intelligent Recognition: One Shot for Multi-Disease Detection. In Benchmarking, Measuring, and Optimizing: Third Bench Council International Symposium, Bench 2020, Virtual Event, 15–16 November 2020; Revised Selected Papers 3; Springer International Publishing: Cham, Switzerland, 2021; pp. 177–193. [Google Scholar]
  34. Sundaram, R.; Ks, R.; Jayaraman, P. Extraction of blood vessels in fundus images of retina through hybrid segmentation approach. Mathematics 2019, 7, 169. [Google Scholar] [CrossRef]
  35. He, J.; Li, C.; Ye, J.; Qiao, Y.; Gu, L. Multi-label ocular disease classification with a dense correlation deep neural network. Biomed. Signal Process. Control 2021, 63, 102167. [Google Scholar] [CrossRef]
  36. Gour, N.; Khanna, P. Multi-class multi-label ophthalmological disease detection using transfer learning based convolutional neural network. Biomed. Signal Process. Control 2021, 66, 102329. [Google Scholar] [CrossRef]
  37. Junayed, M.S.; Islam, M.B.; Sadeghzadeh, A.; Rahman, S. CataractNet: An automated cataract detection system using deep learning for fundus images. IEEE Access 2021, 9, 128799–128808. [Google Scholar] [CrossRef]
  38. Xu, Y.; Hu, M.; Liu, H.; Yang, H.; Wang, H.; Lu, S.; Liang, T.; Li, X.; Xu, M.; Li, L. A hierarchical deep learning approach with transparency and interpretability based on small samples for glaucoma diagnosis. NPJ Digit. Med. 2021, 4, 48. [Google Scholar] [CrossRef]
  39. Luo, X.; Li, J.; Chen, M.; Yang, X.; Li, X. Ophthalmic Disease Detection via Deep Learning with a Novel Mixture Loss Function. IEEE J. Biomed. Health Inform. 2021, 25, 3332–3339. [Google Scholar] [CrossRef]
  40. Shamsan, A.; Senan, E.M.; Shatnawi, H.S.A. Automatic Classification of Colour Fundus Images for Prediction Eye Disease Types Based on Hybrid Features. Diagnostics 2023, 13, 1706. [Google Scholar] [CrossRef]
  41. Kumar, A.; Zhang, Z.J.; Lyu, H. Object detection in real time based on improved single shot multi-box detector algorithm. EURASIP J. Wirel. Commun. Netw. 2020, 2020, 204. [Google Scholar] [CrossRef]
  42. Mirjalili, S.; Lewis, A. The whale optimization algorithm. Adv. Eng. Softw. 2016, 95, 51–67. [Google Scholar] [CrossRef]
  43. ShuffleNet V2 Code. Available online: https://github.com/ZhuYun97/ShuffleNet-v2-Pytorch (accessed on 12 December 2022).
  44. ODIR Dataset. Available online: https://www.kaggle.com/datasets/andrewmvd/ocular-disease-recognition-odir5k (accessed on 5 January 2023).
  45. EDC Dataset. Available online: https://www.kaggle.com/datasets/gunavenkatdoddi/eye-diseases-classification (accessed on 1 January 2023).
  46. SSD Code. Available online: https://github.com/amdegroot/ssd.pytorch (accessed on 15 December 2022).
  47. Gao, Y.; Zhang, H.; Duan, Y.; Zhang, H. A novel hybrid PSO based on levy flight and wavelet mutation for global optimization. PLoS ONE 2023, 18, e0279572. [Google Scholar] [CrossRef] [PubMed]
Figure 1. The proposed EDC model.
Figure 1. The proposed EDC model.
Applsci 13 11437 g001
Figure 2. Sample images of the ODIR and EDC datasets.
Figure 2. Sample images of the ODIR and EDC datasets.
Applsci 13 11437 g002
Figure 3. Feature extraction using SSD.
Figure 3. Feature extraction using SSD.
Applsci 13 11437 g003
Figure 4. The proposed feature selection process.
Figure 4. The proposed feature selection process.
Applsci 13 11437 g004
Figure 5. ShuffleNet V2 model-based proposed EDC.
Figure 5. ShuffleNet V2 model-based proposed EDC.
Applsci 13 11437 g005
Figure 6. Performance analysis outcome.
Figure 6. Performance analysis outcome.
Applsci 13 11437 g006
Figure 7. Findings of comparative analysis—ODIR dataset [35,37,38,39,40].
Figure 7. Findings of comparative analysis—ODIR dataset [35,37,38,39,40].
Applsci 13 11437 g007
Figure 8. Findings of comparative analysis—EDC dataset [35,37,38,39,40].
Figure 8. Findings of comparative analysis—EDC dataset [35,37,38,39,40].
Applsci 13 11437 g008
Table 1. Dataset properties.
Table 1. Dataset properties.
ClassesODIREDC
Cataract2111038
DR11311098
Glaucoma2071007
Normal11351074
Table 2. Computational strategies.
Table 2. Computational strategies.
ODIREDC
Learning RateParameters
(in Millions (m))
FLOPs
(in Giga (G))
Learning RateParameters
(in Millions (m))
FLOPs
(in Giga (G))
Proposed EDC1 ×   10 4 23.50.781 ×   10 4 27.40.68
Luo et al., model [39]1 ×   10 3 26.71.051 ×   10 3 31.21.10
He et al., model [35]1 ×   10 3 25.40.851 ×   10 3 29.10.75
Junayed et al., model [37]1 ×   10 3 27.10.981 ×   10 3 33.40.91
Xu et al., model [38]1 ×   10 3 25.81.241 ×   10 3 28.50.82
Shamsan et al., model [40]1 ×   10 4 24.10.891 ×   10 4 28.10.72
Table 3. Proposed EDC generalization outcome—ODIR.
Table 3. Proposed EDC generalization outcome—ODIR.
Classes/MetricsAccuracyPrecisionRecallF1-ScoreKappaSensitivitySpecificity
Normal99.198.798.898.795.498.796.1
Glaucoma98.798.798.998.896.398.695.4
Cataracts99.399.199.499.296.799.196.1
DR99.499.198.999.097.299.297.6
Average99.198.999.098.996.498.996.3
Table 4. Proposed EDC generalization outcome—EDC.
Table 4. Proposed EDC generalization outcome—EDC.
Classes/MetricsAccuracyPrecisionRecallF1-ScoreKappaSensitivitySpecificity
Normal99.298.798.698.697.298.795.5
Glaucoma99.798.699.198.896.598.696.1
Cataracts99.598.398.898.595.798.397.5
DR99.399.298.698.996.899.196.1
Average99.498.798.798.796.598.796.3
Table 5. Findings of generalization applied to the ODIR dataset.
Table 5. Findings of generalization applied to the ODIR dataset.
Methods/MetricsAccuracyPrecisionRecallF1-ScoreKappaSensitivitySpecificity
Proposed EDC99.198.999.098.996.498.996.3
Luo et al., model [39] 96.594.595.194.894.294.592.4
He et al., model [35]97.896.195.895.993.196.193.1
Junayed et al., model [37]96.195.695.795.694.595.690.5
Xu et al., model [38]97.696.396.896.595.196.392.6
Shamsan et al., model [40]98.796.997.297.094.896.991.3
Table 6. Findings of generalization applied to the EDC dataset.
Table 6. Findings of generalization applied to the EDC dataset.
Methods/MetricsAccuracyPrecisionRecallF1-ScoreKappaSensitivitySpecificity
Proposed EDC99.498.798.798.796.598.796.3
Luo et al., model [39] 98.197.597.897.696.497.593.1
He et al., model [35]97.697.297.697.495.197.293.7
Junayed et al., model [37]98.498.598.298.396.398.593.5
Xu et al., model [38]97.798.398.898.596.598.393.8
Shamsan et al., model [40]99.399.198.798.996.199.192.5
Table 7. Findings of loss function analysis.
Table 7. Findings of loss function analysis.
ODIREDC
MADRMSETesting Time
(Seconds)
MADRMSETesting Time
(Seconds)
Proposed EDC0.2890.3411.080.2230.3421.15
Luo et al., model [39] 0.3140.4142.050.2490.3511.89
He et al., model [35]0.4120.3561.890.3150.3642.03
Junayed et al., model [37]0.3180.3482.120.2470.4021.76
Xu et al., model [38]0.2980.3421.570.2650.3711.85
Shamsan et al., model [40]0.2850.3241.320.2530.3571.17
Table 8. Findings of uncertainty analysis.
Table 8. Findings of uncertainty analysis.
ODIREDC
CISDCISD
Proposed EDC[97.8–98.3]0.0021[98.1–98.3]0.0019
Luo et al., model [39] [95.3–96.1]0.0023[96.7–97.1]0.0021
He et al., model [35][96.4–96.8]0.0027[97.4–97.8]0.0024
Junayed et al., model [37][97.1–97.6]0.0026[96.4–97.1]0.0019
Xu et al., model [38][96.9–97.4]0.0020[97.9–98.1]0.0017
Shamsan et al., model [40][97.1–97.5]0.0024[96.1–96.3]0.0022
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wahab Sait, A.R. Artificial Intelligence-Driven Eye Disease Classification Model. Appl. Sci. 2023, 13, 11437. https://doi.org/10.3390/app132011437

AMA Style

Wahab Sait AR. Artificial Intelligence-Driven Eye Disease Classification Model. Applied Sciences. 2023; 13(20):11437. https://doi.org/10.3390/app132011437

Chicago/Turabian Style

Wahab Sait, Abdul Rahaman. 2023. "Artificial Intelligence-Driven Eye Disease Classification Model" Applied Sciences 13, no. 20: 11437. https://doi.org/10.3390/app132011437

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop