Next Article in Journal
A Formal Model for Adaptive Free Choice in Complex Systems
Next Article in Special Issue
Amazon Employees Resources Access Data Extraction via Clonal Selection Algorithm and Logic Mining Approach
Previous Article in Journal
Optimization for Software Implementation of Fractional Calculus Numerical Methods in an Embedded System
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Machine Learning Based Automated Segmentation and Hybrid Feature Analysis for Diabetic Retinopathy Classification Using Fundus Image

1
Department of Computer Science & IT, The Islamia University of Bahawalpur, Bahawalpur 61300, Pakistan
2
Institute of Numerical Sciences, Kohat University of Sciences & Technology, Kohat 26000, Pakistan
3
Program in Applied Statistics, Department of Mathematics and Computer Science, Faculty of Science and Technology, Rajamangala University of Technology Thanyaburi (RMUTT), Thanyaburi, Pathumthani 12110, Thailand
4
Center of Excellence in Theoretical and Computational Science (TaCS-CoE) & KMUTT Fixed Point Research Laboratory, Room SCL 802 Fixed Point Laboratory, Science Laboratory Building, Departments of Mathematics, Faculty of Science, King Mongkut’s University of Technology Thonburi (KMUTT), 126 Pracha-Uthit Road, Bang Mod, Thrung Khru, Bangkok 10140, Thailand
5
Department of Medical Research, China Medical University Hospital, Taichung 40402, Taiwan
6
Department of Statistics, Mugla Sıtkı Koçman University, Mugla 48000, Turkey
7
Department of Statistics, Govt S.A Post Graduate College Dera Nawab Sahib, Bahawalpur 63351, Pakistan
8
Department of Mathematics, Université de Caen, LMNO, Campus II, Science 3, 14032 Caen, France
9
Department of Computer Science, Govt Degree College for Women Ahmadpur East, Bahawalpur 63350, Pakistan
10
Department of Mathematics, Abdul Wali Khan University Mardan, Mardan 23200, Pakistan
*
Authors to whom correspondence should be addressed.
Entropy 2020, 22(5), 567; https://doi.org/10.3390/e22050567
Submission received: 23 March 2020 / Revised: 23 April 2020 / Accepted: 23 April 2020 / Published: 19 May 2020
(This article belongs to the Special Issue Information-Theoretic Data Mining)

Abstract

:
The object of this study was to demonstrate the ability of machine learning (ML) methods for the segmentation and classification of diabetic retinopathy (DR). Two-dimensional (2D) retinal fundus (RF) images were used. The datasets of DR—that is, the mild, moderate, non-proliferative, proliferative, and normal human eye ones—were acquired from 500 patients at Bahawal Victoria Hospital (BVH), Bahawalpur, Pakistan. Five hundred RF datasets (sized 256 × 256) for each DR stage and a total of 2500 (500 × 5) datasets of the five DR stages were acquired. This research introduces the novel clustering-based automated region growing framework. For texture analysis, four types of features—histogram (H), wavelet (W), co-occurrence matrix (COM) and run-length matrix (RLM)—were extracted, and various ML classifiers were employed, achieving 77.67%, 80%, 89.87%, and 96.33% classification accuracies, respectively. To improve classification accuracy, a fused hybrid-feature dataset was generated by applying the data fusion approach. From each image, 245 pieces of hybrid feature data (H, W, COM, and RLM) were observed, while 13 optimized features were selected after applying four different feature selection techniques, namely Fisher, correlation-based feature selection, mutual information, and probability of error plus average correlation. Five ML classifiers named sequential minimal optimization (SMO), logistic (Lg), multi-layer perceptron (MLP), logistic model tree (LMT), and simple logistic (SLg) were deployed on selected optimized features (using 10-fold cross-validation), and they showed considerably high classification accuracies of 98.53%, 99%, 99.66%, 99.73%, and 99.73%, respectively.

1. Introduction

Diabetic retinopathy (DR) is one of the severe dysfunctions of the human eye [1]. If a person has diabetes for a long time, along with an uncontrolled blood sugar level, he/she is at a clearly high risk of having visual complications. DR takes years to progress, and a well-controlled blood sugar level can slow down DR progression [2]. Additionally, DR is one of the foremost causes of sightlessness, and its time taking progression makes it curable if diagnosed in early stage; however, in the case of failure in its early stage detection, it can damage the human eye’s retina, leading to uncurable sightlessness [3]. The normal retinal assessment of diabetic patients is useful for the early identification of DR, considerably reducing blindness occurrence. However, retinal screening is time-consuming and requires highly qualified ophthalmologists to scrutinize the fundus photographs probing the retinal lesions [2].
It has been shown that DR sometimes occurs due to characteristic changes in the retina, like diametric change in the blood vessels (which are made up of light-sensitive tissues) that lie near the optical nerve [4]. Light rays coming into the eyes are collected on the retina, and the transmission of vision signal to the brain is made through optical nerves where interpretation is done [4]. Macula (the small surface area at the midpoint of retina) is responsible for the recognition of face and read–write functions. The peripheral retina is a part of the retina surrounding macula that is responsible for effective communication between brain and eye [5].
Non-proliferative (NP) DR is a stage of DR where human blood vessels in the retina get weak, and very small dot sized hemorrhages are formed after the swelling of blood vessels; this swelling and hemorrhaging causes an edema in the retina which affects vision with mild symptoms [6]. If NP-DR is not treated effectively, it becomes proliferative (P) DR—an advanced stage of DR where a major component of human vision—the retina—gets oxygen deprived, leading to neovascularization, which results in the formation of spots or floaters; these are reasons for the loss of vision [6].
Dots, spots, and cobweb-like strings floating in the patient’s eye are the main symptoms of DR that are recognized by researchers [7]. Sometimes, a periodic change in sightedness from clear to blurry vision is observed with time. In some patients, dark/black areas are identified in the eye; these cause poor vision in low light or night, which leads to a complete loss of vision [7]. In some cases, vision loss is due to optic disc extraction, which is now becoming a serious complication of DR [8].
In our modern society, with the advancement of science and technology, humans have faced many diseases associated with age. In particular, the human eye can be divided into two classes: The first one is the dysfunctional eye due to illnesses like glaucoma, blepharitis, and conjunctivitis [9]. The other class of human eye diseases comprise the malfunctioning of the human eye due to other life-related diseases like hypertension, diabetes, and arteriosclerosis. Diabetes affects the human eye by damaging blood vessels in eye retina, causing symptoms from minor vision problems to complete vision loss—this is called DR [9]. The early diagnosis and treatment of DR help protect the eye form sever stages; these are done via retinal images called fundus images (FI), which are captured by using a medical image camera, and the disease is detected by a highly expert ophthalmologist [6]. For diseases falling in first-class mentioned above, assessment is done by the quantitative measurement of retinal deviation, but this class assessment is difficult in early stages because it is related to drusens and normally appears as a form of black/dark/yellow drops or cobweb-like strings in the eye [7]. In modern word color, retinal images (RIs) are used for the detection of drusens that accrue in the retina. Automated detection and segmentation have thus far prominently contributed to the provision of essential information for identifying the stages of DR [4].

1.1. Literature Review

Pires, R. et al. extracted information from retinal images. Neural networks were applied in this regard. This research extracted features that analyzed the patient and tried to find imperfections. The suggested technique attained a receiver–operating characteristic (ROC) area under the curve of 98.2%. The given technique was applied to the Messidor-2 dataset. Similar results obtained by using two-to-five-fold cross-validation protocols when applied to the Messidor-2 and DR2 datasets. Upon referral assessment, this method achieved the highest result, which was 99.0% when convolutional neural networks (CNNs) were used [10]. Zhang, W. et al. described an automatic DR identification system and a classification system. The system identified the occurrence and harshness of DR using fundus photographs. Additionally, an NN was applied. In this research, a high-class dataset was obtained from images. The identification model achieved the best result, with 97.5% sensitivity, 97.7% specificity, and 97.7% area under the curve. The classification model achieved 98.1% sensitivity and 98.9% specificity [11]. Harun, N. H. et al. classified fundus images. For this purpose, a multi-layer perceptron (MLP) ANN employed. The purposed system was trained for classification using Levenberg–Marquardt (LM) and Bayesian regularization (BR), and 19 extracted features were used in the NN to classify images. The MLP trained with the BR had a more improved classification rate than that with the LM. The MLP achieved a 72.11% rate for training and a 67.47% rate for testing with the BR [12]. Verbraak, F. D. et al. determined a system that automatically detected diabetic retinopathy. A total of 1616 people (having diabetes) were imaged. RIs were classified using hybrid deep learning enhanced devices and were also evaluated with a reference standard. The hybrid deep learning enhanced device achieved 99% sensitivity and 97.8% specificity, with respect to the reference standard for vision threatening (vt) DR, while the more than mild (mtm) DR device achieved 79.4% sensitivity and 93.8% specificity with respect to the reference standard. This proved that hybrid deep learning enhanced the device and achieved a high accuracy for the detection of vtDR and mtmDR [13]. Afrin, R. and Shill, P. C. proposed an automatic system for detecting retinal wounds using RIs, and they also categorized the phases of DR. The knowledge-based fuzzy classifier was applied to extracted features for classification, achieving 95.63% accuracy for the classification [14]. Parmar, R. et al. detected DR by using RIs. In this research, images were classified into five classes. An NN was applied in this model. This model achieved 85% accuracy [15]. Xu, K. et al. used an NN method for categorization of DR through the use of color FIs. The suggested method achieved 94.5% accuracy [16]. Sim, D. A. et al. proposed a system that used automatic real-time calculation for further analysis and medical appointment, if necessary. Moreover, the real-time calculation was combined with microelectronic medicinal records to enable more precise predictions for specific patients. In particular, a 7.1% accuracy of mentionable infection was achieved, 97.4–99.1% accuracy was achieved for sensitivity, and 98.3–99.3% accuracy was achieved for specificity [17]. Gulshan, V. et al., described the automatic recognition of DR and diabetic macular edema in RF images. This work reached 97.5% sensitivity and 93.4% specificity on the EyePACS-1 dataset, and it achieved 96.10% sensitivity and 93.90% specificity on the Messidor-2 dataset using deep learning algorithms [18]. Gargeya, R. and Leng, T. described strong analytical expertise for automatic DR screening. The suggested model attained an area under the curve (AUC) of 0.97 with 94% sensitivity and 98% specificity thought “five-fold cross-validation” on the local dataset. The Messidor-2 database attained a 0.94 AUC, and the E-Ophtha database attained a 0.95 AUC using a data-driven deep learning algorithm [19].

1.2. Contribution

The contribution of this study can be summarized as follows:
A novel segmentation technique, which is called clustering-based automated region growing segmentation (CARGS) and includes post optimal feature selection, is introduced. This technique includes four steps.
  • Firstly, the RF image is divided into four equal regions. The group of neighboring seeds is used for the formulation of an identifiable region. These seeds are in the shape of an irregular polygon with a variable radius from the center of an image in order to ensure the maximum chance of grouping seeds that belong to the same region. This allows for any possible size, dimension, and shape to be considered as regions of interest (ROIs). At the post-processing stage, the K-mean algorithm is employed on improved segmented regions.
  • After segmentation, hybrid-features data are extracted.
  • Four machine learning (ML) feature selection classifiers are employed for the post optimal hybrid feature selection.
  • The post-optimized hybrid features dataset is deployed for five ML classifiers, and efficient classification accuracy is acquired.

2. Materials and Methods

This study comprised a dataset that held five types of RF images; one was that of normal patient retinas and the other four that were DR patient stages named mild, moderate, non-proliferative, and proliferative, as shown in Figure 1. This dataset was collected via a fundus photography machine (Vision Star, 24.1 MegaPixel Nikon D5200 camera) available in Bahawal Victoria Hospital (BVH) Bahawalpur [20], Pakistan. RF images provide better structural changes in blood veins and tissues, and they also provide better contrast enhancement and sensitivity. For each type, a hundred RF images sized 256 × 256 and a total of 2500 (500 × 5) RF dataset images of normal and DR patients were acquired. All the images were manually examined by an expert ophthalmologist in light of different medical tests and biopsy reports. Finally, in the presence of a gold standard/ground truth RF image dataset, we proposed a novel CARGS approach.

2.1. Proposed Methodology

The proposed methodology is described in Figure 2. First of all, the proposed algorithm is described with all procedural steps.
Proposed Algorithm
Begin
Main {
Input ϵ Retinal fundus image dataset
For {
 Steps 1–10
  • Fundus image datasets ϵ five retinal types.
  • Image pre-processing.
  • Clustering-based automated region growing segmentation (CARGS).
  • Extract texture features ϵ histogram, co-occurrence matrix, run length matrix, and wavelet.
  • Generate fused hybrid-feature dataset.
  • Pre-optimization (Fisher (F), probability of error (POE) plus average correlation (AC), and mutual information (MI)) feature selection technique employed on fused hybrid-feature dataset.
  • Extract 30 pre-optimized fused hybrid-feature dataset.
  • Post-optimization (correlation-based feature selection—CFS) feature selection technique and employed pre-optimized fused hybrid-feature dataset.
  • Extract 13 post-optimized, fused hybrid-feature dataset.

End For
}
10.
ML classifiers are employed on post-optimized, fused hybrid-feature dataset.
Output = DR classification results
End main
}
Now, let us discuss the proposed methodology in detail. The first step consists of a collection of image datasets, which, in this step, are the RF images of a normal patient and four different types of DR patients, which, for this study, were collected from Bahawal Victoria Hospital Bahawalpur, Pakistan [20]. The second step is image pre-processing. In this step, digital color retinal images are converted into a gray level 8-bit image format. Secondly, noise removal is performed by using “Gaussian and Gabor filters”. Finally, data cleaning is done for the RF image datasets standardization. The third step is segmentation, which helps to remove the extra object and nominates the exact position and refines the texture of the lesion. The fourth step is hybrid feature extraction. In this step, four types of features—histograms, co-occurrence matrixes, run lengths matrix, and wavelet features—are extracted from a standardized RF image dataset. The fifth step is the formation of a fused hybrid-feature dataset using a data fusion technique, which is a very powerful technique for merging multiple features to produce an accurate classification compared to individual features. The sixth step is fused hybrid feature optimization. In this step, we select the best attribute of the extracted hybrid feature dataset. Firstly, 30 pre-optimized hybrid features are selected using ML (F, PA, and MI) feature selection techniques. Secondly, the (CFS) ML feature selection technique is deployed on a pre-optimized dataset, and 13 post-optimized fused hybrid features that are most useful for the classification of DR are selected. The last step is classification, where five ML classifier are employed (using 10-fold cross-validation) on the selected, post-optimized fused hybrid features datasets. Cross-validation, a standard evaluation approach, is a systematic way of obtaining repeated percentages. It follows the following scheme. Break a dataset into 10 parts (“folds”), then hold each part to test, and train the remaining nine at the same time; this returns an average of 10 evaluation results. By initializing, in “validated” cross-validation, it is ensured that there is, approximately, the correct proportion of class values in each class. After 10 times of cross-validation and the counting of the evaluation results, ML requires the learning algorithm to be used one last (11th) time in the final dataset to obtain the model it publishes. These ML classifiers are named sequential minimal optimization (SMO), logistic (Lg), multi-layer perceptron (MLP), logistic model tree (LMT), and simple logistic (SLg).

2.2. Image Preprocessing

The acquired (2D) color digital retinal fundus image dataset is converted into gray level (eight-bit) image format, and histogram equalization is used to normalize non-uniformities and improve contrast. The gray level image has a 256 gray-level (0–255), and, in a histogram, the vertical axis rests on a numeral of pixel in the image and the horizontal axis extends from 0 to 255 [21]. The “probability density function” (p.d.f.) of pixel intensity level 𝓇 i is shown in Equation (1):
P 𝓇 ( 𝓇 i ) = K i / K
where 0   𝓇 i 1 , K i is the number of pixels with intensity 𝓇 i , K represent the total number of pixels and i = 0–255.
During the acquisition of RF image data, freckled noise is detected due to the environmental conditions of the image sensor. A noise removal process is adopted to solve this problem. In this process, two image processing filters are used. Firstly, a Gaussian filter [22] is implemented to overcome this issue and enhance the contrast. This filter is expressed below.
Let, Y be a random variable following the Gaussian distribution with a zero mean and standard deviation (S.D) = λ , i.e., Y   ~ N ( 0 , λ 2 ) . In the unidimensional case, the p.d.f. of Y is shown in Equation (2):
f ( y ) = [ 1 λ .   2 π ]   e   y 2 2 λ 2
Let, Z be a random variable following the Gaussian distribution with zero mean and S.D = λ , i.e., Z   ~   N ( 0 , λ 2 ) . For the multidimensional case, the joint p.d.f. of (Y, Z) is denoted by ψ ( y , z ) , and for the two-dimensional case, it is defined in Equation (3):
ψ ( y , z ) = [ 1 2 λ 2   π   ]   e   (   y 2 + z 2 ) 2   λ 2
Secondly, the Gabor filter [23] is implemented to enhance and detect edges that are based on the Gaussian kernel function. Mathematically, the Gabor filter is expressed in Equation (4):
g ( y , z , λ , θ , η , σ , γ ) = e ( y 2 ´ + γ 2 z ´ 2 2 σ 2 )   e ( i   ( 2 π y ´ λ + η ) )
In this equation, λ denotes the wavelength of the sinusoidal factor, θ denotes the angle of the normal to the parallel stripes of a Gabor function, η is the point offset, σ is the sigma standard deviation of the Gaussian envelope, and γ is the spatial aspect ratio that specifies the ellipticity of the support of the Gabor filter.
y ´ =   ( y   c o s θ + z   s i n θ   )
z ´ =   ( y   s i n θ + z   c o s θ   )
Here, (y, z) is the image pixel, the scale is a location parameter, θ positioning angle, θ = ( k π ) n , (k = 0 to n − 1), where n represents the angles (k = 0 to 7) and 5 location parameters (scale = 1 to 5). By using these techniques, the noisy values are replaced with average values in the image, and a smooth and enhanced RF image dataset can be acquired.

2.3. Clustering-Based Automated Region Growing Segmentation (CARGS)

Segmentation is a procedure that helps to eliminate the extra object, refines the texture of the lesion, and recommends a precise position. ROI extraction has many automated and semi-automated methods. We know that automated ROI extraction is usually based on the idea of image segmentation, but there is no single method for ideal segmentation. On the other hand, there are semi-automated techniques based on expert opinion, but human-based extraction has some limitations. To solve this problem, CARGS is employed on DR images, which helps to examine the qualitative nature of data and useful information. This idea uses a group of neighboring seeds for the formulation of an identifiable region, which is entirely different approach than using contagious and connected sets of seeds fixed dimensions of 2 × 2, 3 × 3, 4 × 4, 5 × 5, or 8 × 8. These seeds are in the shapes of irregular polygons with variable radii from the center of their image. This methodology of seed selection ensures the maximum chance of grouping seeds that belong to same region, as shown in Figure 3.
The new scheme avoids domain-specific problems, as fixed dimensions can be useless/not fit for all images, especially images with micro textures like bio medical images, where abnormalities are not observed in some regular shapes but can be seen when drawing irregular polygons with different dimensions. Thus, we can expect calculations to be more accurate because they cover the maximum amounts of interested/abnormal regions (in case of medical images) accordingly. The main advantage of this scheme over a scheme with regular polygons is the segmentation of a fewer possible and more specific homogeneous number of seeds with minimum noise, as shown in Figure 4.
By using the regular polygonal seed selection methodology for bio-medical, images like retinal fundus images slices are not connected, which results in extreme difficulty for useful segmentation [24]. However, an irregular polygonal seed selection scheme resolves this issue because it allows for any possible sizes, dimensions, and shapes to be considered as regions of interest (ROI). At the post processing stage, a K-mean algorithm [25,26] is employed on improved segmented regions, which results in the removal of noise for improved results.
Here, the RF image with an (i × j) resolution and RF image was divided into K number of clusters. The value of K is considered as a threshold point, the range is between 7 and 12. Let p (i × j) be the RF image seeds to cluster and KC be the center point. Then, the K-mean clustering-based segmentation algorithm works as follows:
  • First, set the K (threshold value) and center point.
  • Second, calculate Euclidean distance ED for each seed of an RF image, between the center and each seed using the Equation (7):
E D = | |   p ( i ,   j ) K C | |
  • The nearest cluster holds the seed based of distance.
  • When all the seed have been allocated, recalculate the new location of center using Equation (8):
K C = 1 C   j K C i K C p ( i ,   j )
  • Continue the process unless it meets the value of tolerance or error.
  • Reshape the cluster seeds into RF images.
It has been experimentally proven that batch and volume processing for region growing are possible without expert knowledge, although it is a difficult process. The watershed (topographical) segmentation technique [27] that is used for retinal fundus image is not good enough because of issues like over-segmentation and sensitivity to noise; as a result of the watershed technique, any process is degraded by background noise and over-segmentation, as shown in Figure 5.
On the other hand, the proposed scheme uses marker-based segmentation approach radii, shapes, and centers of the irregular polygon-defined foreground, which improves segmentation results. The K-mean segmentation results are disturbed when K > 12 (K = 12; maximum threshold value) on retinal fundus images with the proposed seed-based segmentation scheme, which can successfully produce more useful clustering information compared to information produced by applying previous schemes, as shown in Figure 4.

2.4. Feature Extraction

For this study, a hybrid-feature dataset of DR was acquired using RF images; that is, the first-order histogram, second-order co-occurrence matrix, wavelet, and run-length matrix features. These features were grouped as 11 second-order co-occurrence matrix features including five average texture values in all four dimensions (0°, 45°, 90°, and 135°) and (11 × 5 × 4) a total of 220 calculated features, 15 run-length matrix features, 6 histogram features, and 4 wavelet features. Thus, 245 features per each ROI were extracted, and the total calculated features vector space (FVS) was 1,837,500 (245 × 7500) for the acquired RF image dataset. All these features were acquired using MaZda software, version 4.6. To carry out this study, all the experiments were carried out on an Intel® Core i3 1.9 gigahertz (GHz) processor with 8 gigabytes (GB) of RAM and a 62-bit Windows 10 operating system.

2.4.1. Run-Length Matrix (RLM)

Galloway [28] introduced the gray level run-length matrix (GLRM), a section of gray scale or color level—also known as a range or length of run—that is a linear multitude of continuous pixels with the same color or gray level in a particular direction. That is, N g is the total number of gray levels in the image, N r is the number of run lengths in the image, N p is the number of voxels in the image, N is the number of runs in the image in a particular direction, P ( l , m ) is the run length matrix for an particular direction, p ( l , m ) = P ( l , m ) / N is the normalized run length matrix for particular direction, and u = l = 1 N g m = 1 N r i   P ( l , m ) . Then, the short run emphasis (SRE) is described in Equation (9):
S R E = l = 1 N g m = 1 N r P ( l , m ) m 2 N
The SRE measures the distribution of short runs. A high cost indicates a fine texture. Long run emphasis (LRE) measures the distribution of long runs. The high value indicates the coarse texture described in Equation (10):
L R E = l = 1 N g m = 1 N r P ( l , m ) m 2 N
Gray level non-uniformity (GLN) measures run distribution over gray values. The cost of the feature is lower when the runs are evenly distributed. In this way, a lower value indicates a higher similarity in intensity values described in Equation (11):
G L N = l = 1 N g [ m = 1 N r P ( l , m ) ] 2 N  
Gray level non-uniformity normalized (GLNN) is a normalized version of the GLN feature defined in Equation (12):
G L N N = l = 1 N g [ m = 1 N r P ( l , m ) ] 2 N 2
Run length non-uniformity (RLN) measures the distribution of runs over the run lengths. The cost of the feature decreases as the length of the run is evenly distributed, as defined in Equation (13):
R L N = m = 1 N r [ l = 1 N g P ( l , m ) ] 2 N  
Run length non-uniformity normalized (RLNN) is a normalized version of the RLN feature defined in Equation (14):
R L N N = m = 1 N r [ l = 1 N g P ( l , m ) ] 2 N 2
Run percentage (RP) measures the fraction of the number of realized runs and the maximum number of potential runs. Strongly linear or highly uniform ROI volumes produce a low run percentage, as shown in Equation (15):
R P = N N p  
Low gray level run emphasis (LGLRE) is described in Equation (16):
L G L R E = l = 1 N g m = 1 N r   P ( l , m ) / l 2 N
High gray level run emphasis (HGLRE) is a grey level (GL) analogue to long run emphasis. The feature emphasizes high grey levels, as described in Equation (17):
H G L R E = l = 1 N g m = 1 N r   P ( l , m ) l 2 N
Short run low gray level emphasis (SRLGLE) emphasizes runs in the upper left quadrant of the GLRLM, where short run lengths (SRLs) and low GLs are located [29]. It is described in Equation (18):
S R L G L E = l = 1 N g m = 1 N r   P ( l , m ) / l 2 m 2 N
Short run high gray level emphasis (SRHGLE) emphasizes runs in the lower left quadrant of the GLRLM, where SRLs and high GLs are located. It is given in Equation (19):
S R H G L E = l = 1 N g m = 1 N r   P ( l , m ) l 2 / m 2 N
Long run low gray level emphasis (LRLGLE) emphasizes runs in the upper right quadrant of the GLRLM, where long run lengths (LRLs) and low GLs are located. It is defined in Equation (20):
L R L G L E = l = 1 N g m = 1 N r   P ( l , m ) m 2 / l 2 N
Long run high gray level emphasis (LRHGLE) emphasizes runs in the lower right quadrant of the GLRLM, where LRLs and high GLs are located. It is presented in Equation (21):
L R H G L E = l = 1 N g m = 1 N r   p ( l , m ) l 2 m 2 N
Grey level variance (GLV) measures the variance in runs for the GL. It is given in Equation (22):
G L V = l = 1 N g m = 1 N r   p ( l , m ) ( l u ) 2
Finally, run length variance (RLV) measures the variance in runs for run lengths. It is presented in Equation (23)
R L V = l = 1 N g m = 1 N r   p ( l , m ) ( m u ) 2

2.4.2. Histogram Features (H)

Histogram features are used by selecting the object with respect to rows and column [30]. This binary object is used as a mask of the original image for feature extraction. Histogram features are calculated by the intensity of the individual pixels, which are the part of objects. These features are based on the histogram. They are also called first order histogram or statistical features. The first order histogram probability P(h) is described in Equation (24):
P ( h ) = K ( h ) N
It is clear that N represents the total numbers of pixels in the image and K(h) shows the complete instances of the gray scale value of h. The mean is average of the values; it describes the bright mean and dark mean in an image. The mean is defined as follows in Equation (25):
i ¯ = h = 0 p 1 h P ( h ) = m n k   ( m , n ) k
Here, q represents the grayscale values that range from 0 to 255. The consequent values of m (rows) and n (columns) show the pixel. Standard deviation (SD) describes the contrast of image. It is presented in Equation (26):
σ h = h = 0 P 1 ( h h ¯ ) 2 P ( h )
The skewness measures a degree of asymmetry in comparison to a central value. It is defined as follows in Equation (27), and negative skewness is defined in Equation (28):
S k e w n e s s = 1 σ h 3 h = 0 P 1 ( h h ¯ ) 3   P ( h )
S k e w n e s s = h ¯ m o d e σ h
The gray level distribution is called energy. It is described in Equation (29):
E n e r g y = h = 0 P 1 [ P ( h ) ] 2  
Entropy is described as randomness in the image data. It is defined in Equation (30):
E n t r o p y = i = 0 q 1 q   ( i ) l o g 2 [ q ( i ) ]

2.4.3. Co-Occurrence Matrix (COM)

COM (co-occurrence matrix) features are also called the second order statistical features. These features are obtained on the distance and the angle between pixels, based on gray level co-occurrence matrix (GLCOM) [31]. For this study, 11 second order texture features were calculated in four dimensions—0°, 45°, 90°, and 135°—up to a 5-pixel distance. If Pn is the normalized type of the gray level ( g ), the co-occurrence matrices of four directions like 0°, 45°, 90°, and 135° are described in Equation (31):
P n = 1 N c ( [ P ( a , b ) ] b = 0 , 1 , 2 , , N g 1 a = 0 , 1 , 2 , , N g 1 )
Here, the normalization constant N c can be expressed as in Equation (32):
N c = a = 0 N g 1 b = 0 N g 1 P ( a , b )
First of all, energy is defined in Equation (33). Energy is calculated in distribution between gray level values:
g E = a = 0 N g 1 b = 0 N g 1 { P n ( a , b ) } 2
The contrast is a difference moment of a matrix and it is a measure of local changing in the image. It is described in Equation (34):
g c = n = 0 N g 1 n 2 { a = 0 N g 1 b = 0 N g 1 P n ( a , b ) }
Here, | a b | = n . Correlation describes the pixel similarity at particular pixel distance. It is described in Equation (35):
g c o r = a = 0 N g 1 b = 0 N g 1 ( r c ) P n ( a , b ) μ H μ V σ H σ V
where μ H , μ V , σ H , and σ V are the means and standard deviations of the marginal probability matrices MH and MV, respectively and normalized gray level denoted by Ng, which are shown as M H = [ M H ( a ) ] b = 0 , 1 , 2 , , N g 1 and M V = [ M V ( b ) ] a = 0 , 1 , 2 , , N g 1 . Moreover, M H ( a ) = b = 0 N g 1 M n ( a , b ) , a= 0, 1, 2,…, Ng – 1. While, a remains constant in complete row, the column sum of Mn. Similarly, M V ( b ) = a = 0 N g 1 M n ( a , b ) , b = 0, 1, 2,…, Ng − 1. Additionally, b is constant in whole column, which means that is the row sum of Mn.
Additionally, M H + V = [ M H + V ( k ) ] k = 0 , 1.2 , 3 , , 2 N g 2 , M H + V ( k ) = a = 0 N g 1 b = 0 N g 1 M n ( a , b ) , where k = a + b, M H V = [ M H V ( k ) ] k = 0 , 1.2 , 3 , , N g 1 , M H V ( k ) = a = 0 N g 1 b = 0 N g 1 M n ( a , b ) , and k = |a − b|.
This feature gives a measure of variability in specific regions of an image. If gray values have low differences in an image, then variance is low. This is described in Equation (36):
g v a r = a = 0 N g 1 b = 0 N g 1 ( a μ ) 2 M n ( a , b )
The local homogeneity of the image is called the inverse difference, and it is presented in Equation (37):
g I D M = a = 0 N g 1 b = 0 N g 1 1 1 + k 2 M n ( a , b )
Difference variance is defined in Equation (38):
g D V = a = 0 N g 1 ( a μ ) 2 M x y ( v )  
Sum average is presented in Equation (39):
g S A = a = 0 2 N g 2 a M x + y ( a )
Sum variance is described in Equation (40):
g S V = v = 0 2 N g 2 ( a μ ) 2 M x + y ( a )
Entropy measures the total content of the image. It is described in Equation (41):
g E N = a = 0 N g 1 b = 0 N g 1 M n ( a , b ) l o g ( M n ( a , b ) )
Sum entropy is presented in Equation (42):
g S E N = a = 0 2 N g 2 M x + y ( a ) l o g { M x + y ( a ) }
Finally, difference entropy is defined in Equation (43):
g D E N = a = 0 N g 1 M x y ( a ) l o g { M x y ( s ) }

2.4.4. Wavelet-Based Features (W)

Generally, discrete wavelet transforms (DWT) features are the linear transformation which implemented on a data vector whose value is an integer power of two, changing it numerically different data vectors in the same quantitative values. It means data is segmented into different frequency resolutions. After this DWT value is calculated by using successive flow by using a factor of two subsampling which is high pass filters (H) and low pass filters (L). For this study the Harr wavelet feature are calculated for DR dataset. This dataset is segmented into four sub bands (HH, HL, LH, and LL) at each scale. Sub band XLL is used to compute the DWT features. For DR dataset, the Harr wavelet features computed only if output sub bands have at least 8 by 8 dimension. Furthermore, the Equation (44) described the DWT features calculated on a given ROIs [32]:
E n e r g y S u b b a n d ,   s c a l e = y , z R i g i o n O f I n t r e s t ( d y ,   z s u b b a n d ) 2 k
where the constraint d y , z is the resultant matrix component. Σ is performed for each pixel (y, z) situated in the region of interest definition, and K is the total number of pixels in ROIs.

2.5. Feature Selection

Feature selection is the most important part of the ML process. The main goal of this process is to select the most worthwhile features and remove the most worthless features in a dataset. This process deals with the elimination of missing values and highly correlated, low variance, and recursive features, and it also deals with univariate feature selection by using ML. Usually, an acquired dataset is a combination of large numbers of features that are difficult to handle. It is essential to reduce the dimensionality of this feature vector space, which can efficiently distinguish and classify the different classes. These techniques can be implemented to achieve the most discriminating features and to get cost-effective classification results [33].
In this study, it was observed that all the extracted features were not equally worthwhile for DR classification. Since the acquired dataset had a large FVS (precisely 1,837,500), it was very difficult to deal with. To solve this problem, three ML-based feature selection techniques, namely F, MI, and probability of error (POE) plus average correlation (AC), were used to select the optimize features for the acquired DR dataset. This approach was the combination set of above mention ML techniques (F, MI, and PA) [34] that selected 30 pre-optimized features out of 245 features using the MaZda software [35]. Mathematically, the F technique can be represented as:
F =   N 2 V 2 = [ 1 / ( 1 i = 1 i   ρ i 2 ) ] i = 1 i .   y = 1 i ρ i     ρ y | L i L y | 2 i = 1 i .   ρ i M i 2
where N is between-class variance, V is within-class variance, ρ i is probability of feature i, and M i and L i are the variance and mean value of feature i in the given class, respectively. In addition, POE and AC are defined as:
P A ( f y ) = M i s s c l a s s i f i e d S a m p l e s T o t a l S a m p l e s
f 2 = f y   :   M i n y [   P A ( f y ) + | C o r r e l a t e   ( f 1 ,   f y ) | ]
f k = f y   :   M i n y [   P A ( f y ) + 1 N 1   | C o r r e l a t e   ( f 1 ,   f y ) | ]
Additionally, MI is represented as:
M I ( I ,   J ) =   I   J .     P ( I . J ) L o g 2 P   P ( I . J ) P ( I ) . P ( J )
In this study, it was observed that the above feature selection (F, MI, and PA) techniques had some limitations. They selected exactly 30 features, but only a few of them were worthwhile. Thus, one more feature selection technique, namely CFS, was used on pre-an optimized dataset [36]. CFS can extract the most projecting features in a dataset. CFS uses the information theory based on the entropy presented in Equation (50):
H ( Y ) =   P ( y i ) l o g 2   P ( y i )
The entropy of Z after observing values of another variable W is defined in Equation (51):
H ( Y | X )   =   P ( x j )   P ( y i x j ) l o g 2   P ( y i )
Here, P ( y i ) the primary probabilities for all values of Y and P ( y i / x j ) is the secondary probabilities of Y when values of X are given. The value found when the entropy of Y is reduced shows more information about Y given by X, and this is called additional information; see Equation (52):
I A ( Y | X ) = T ( Y ) T ( Y | X )
Here, the X feature is more correlated to the Y feature than to the Z feature if the following inequality holds:
I A ( Y | X ) > I A ( Z | X )
We have to quantify symmetrical uncertainty (SYU), which expresses the correlation among features. It is described in Equation (54):
S Y U ( Y , X ) = 2 { I A ( Y | X ) | T ( Y ) + T ( X ) }
Entropy-based quantification needs normal features, but it is implemented to properly calculate the correlations between continuous and discretized features. CFS was deployed on the pre-optimized dataset that selected 13 post-optimized features for further processing. The post-optimized features are shown in Table 1.
Finally, the 1,837,500 (245 × 7500) fused hybrid-features vector space was reduced to 97,500 (13 × 7500). CFS-based post-optimized datasets were generated for each type of extracted feature for DR and these post-optimized, fused hybrid-feature datasets were deployed to different machine learning classifiers. Figure 6 shows the three-dimensional (3D) representation of the post-optimized features. MDF1, MDF2, and MDF3 are three different dimensions (like x, y, z) of most discriminant features.

2.6. Classification

Five ML classifiers—SMO, Lg, MLP, LMT, and SLg—were employed on RF image datasets. First, it was observed that by employing these classifiers on the histogram, wavelet, and co-occurrence matrix feature datasets, unfortunately, a very low accuracy was found, with values of 73.73%, 79.93% and 89.93%, respectively. Similarly, when the same classifiers with the same strategies were deployed on the run-length matrix features, the overall accuracy of 95.93% was observed, which was slightly better than that of the other feature’s datasets. Finally, to obtain a more promising accuracy, we generated a post-optimized, fused hybrid-feature dataset, and very impressive results were observed for the deployed classifiers. Here, the SLg classifiers performed the best among the implemented classifiers, though it mostly performed well for noisy, big, and complex data [35]. The SLg classifier is theoretically expressed in Equation (55):
l o g ( P 1 P ) = w 0 + w 1 x 1 + w 2 x 2 + + w k x k
where:
p = 1 1 + b ( w 0 + w 1 x 1 + w 2 x 2 + + w k x k )
and x1 … xk are the explanatory variables with parameters w0, w1,…, wk.

3. Results and Discussion

For this study, five ML classifiers named SMO, Lg, MLP, LMT, and SLg were employed on selected post-optimized hybrid feature datasets (using 10-fold cross-validation) for the classification of DR. As discussed earlier, four types of texture features, namely H, W, COM, and RLM, were extracted using the RF image datasets. In the first step, the histogram features-based dataset was used for DR classification, and this did not give better results than the other employed classifiers, at less than or equal to 73.73%. In the second step, the same approach was employed on the wavelet features-based dataset, where the overall classification accuracy of the employed classifiers was less than or equal to 79.93%. Here, we observed some improvement in accuracy compared to the previous results, but it was not so inspiring. In the third step, the same approach was employed on the co-occurrence matrix features-based dataset, and it was observed that the overall classification accuracy of the employed classifiers was less than or equal to 89.93%. This was a very promising improvement in accuracy compared to the previous results, but this process will continue for further analysis. In the fourth step, the same approach was employed on run-length matrix features-based dataset, and the overall classification accuracy of the employed classifiers was less than or equal to 95.93%; we observed a very promising improvement in accuracy compared to the previous results. In this analysis, we observe that each texture feature carried its worth in the DR analysis case, and the most worthwhile feature was a run-length feature that gave the highest classification accuracy compared to others. To improve classification accuracy, a post-optimized, fused hybrid-feature dataset was generated by applying the data fusion approach, which is a very powerful technique for merging multiple features to produce an accurate classification compared to individual features. In the last step, the same approach was employed on a post-optimized, fused hybrid-features dataset, and the employed classifiers, SMO, Lg, MLP, LMT, and SLg showed very high classification accuracies of 98.53%, 99%, 99.66%, 99.73%, and 99.73%, respectively. The overall classification results of the histogram features-based dataset with the employed ML classifiers with other performance evaluating factors such as kappa statistics, true positive (TP), false positive (FP), ROC, mean absolute error (MAE), root mean squared error (RMSE), confusion matrix, time complexity (T), and overall accuracy (OA) are shown in Table 2.
It was observed that in the employed classifiers on histogram features, the MLP classifier showed a relatively better classification accuracy of 73.73% compared to other employed classifiers, as shown in Figure 7. To acquire some improvement in classification accuracy, we then employed the same classifiers on a wavelet features-based dataset, as shown in Table 3.
It was observed that the overall accuracies of MLP, LMT, Lg, SLg and SMO were 79.93%, 77.80%, 79.86%, 76.60%, and 75.80%, respectively, as shown in Figure 8.
To acquire some more improvement in classification accuracy, we then employed the same classifiers on a co-occurrence matrix (COM) features-based dataset, as shown in Table 4. It was observed that was the overall accuracies of LMT, MLP, SLg, SMO, and Lg were 89.93%, 89.86%, 87.86%, 86.80%, and 86.67%, respectively, as shown in Figure 9.
To acquire some more promising improvement in classification accuracy, we then employed the same classifiers on a run-length matrix (RLM) features-based dataset, as shown in Table 5.
It was observed that the overall accuracies of SLg, Lg, LMT, MLP, and SMO were 95.93%, 95.80%, 95.73%, 95.13%, and 93.67%, respectively, as shown in Figure 10.
Finally, the overall accuracy of the histogram, wavelet, COM, and RLM features were not so impressive; meanwhile, on a post-optimized, fused hybrid-features-based dataset, the same strategy with the same classifiers was employed, and very promising results were observed. The overall classification accuracies of the employed classifiers of SLg, LMT, MLP, Lg, and SMO were 99.73%, 99.73%, 99.67%, 99.13%, and 98.53% respectively, as shown in Figure 11. These results were very inspiring, and it was observed that SLg and LMT showed the same accuracy; however, when including the time factor, the SLg showed the best accuracy among all the implemented classifiers. The overall accuracy of SLg with others implemented ML classifiers on a post-optimized, fused hybrid-feature dataset with performance evaluating parameters is shown in Table 6.
Similarly, the confusion matrix (CM) of the post-optimized, fused hybrid-features dataset is shown in Table 7. The colored diagonal of the CM (Table 7) shows the classification accuracy in appropriate classes, while other instances show them in other classes. This table contains the information that is actual and predicted data for the SLg classifier. The SLg classifier showed a relatively better overall accuracy than any of the implemented classifiers.
The classification accuracy results of the five DR stage (both healthy and DR) RF datasets—that is, healthy, mild, moderate, non-proliferative, and proliferative—were 99.93%, 99.73%, 99.86%, 99.46%, and 99.66%, respectively. The graphical accuracy results are shown in Figure 12.
Finally, a comparative DR classification graph was made of the histogram, wavelet, co-occurrence matrix (COM), run-length matrix (RLM), and post-optimized, fused hybrid-feature datasets using the employed ML classifiers; this is shown in Figure 13. This graph shows the overall better accuracy (RED) for the DR classification that was found by using a post-optimized, fused hybrid-feature dataset compared to RLM (GREEN), COM (BLUE), wavelet (YELLOW), and histogram (SKY-BLUE) features-based datasets.
DR is caused by damage to the blood vessels of the light-sensitive tissue at the back of the eye (retina). Due to this damage, the fine texture of the retina is disturbed, and, due to this problem, texture features play an important role in DR classification. After reviewing the literature, we found different kinds of research work on different types of texture features. In this study, a comparative analysis was performed between all of these features and a fused hybrid feature dataset, which was the combination of all these features (as shown in Figure 13). The existing system has some limitations because it is a supervised, learning-based classifier. All testing was performed in a retina fundus image dataset that was acquired to meet all legal requirements of the ophthalmology department of the BVH, Bahawalpur, Pakistan [20], with the collaboration of the Department of Computer Science of the Islamic University from Bahawalpur (IUB) [37], Pakistan, to address the regional and local problem of identifying diabetic retinopathy stages. The proposed model was verified on the publicly available datasets of the diabetic retinopathy (DR) archives the High-Resolution Fundus (HRF) Image and Messidor-2 (M2) [38,39] databases. Fifty patients’ data of five stages of DR were acquired from the HRF and M2 datasets using a Canon CR-1 fundus camera with a field of view of 45° and different acquisition settings. Twenty retinal fundus images of each stage and total of 100 (20 × 5) datasets of five DR stages (where one is a normal patient retina and the other four are DR patient stages, namely mild, moderate, non-proliferative, and proliferative) were acquired from these public archives. The multi-institutional dataset differences between HRF, M2, and BVH were observed, and we tried our best to normalize the HRF and M2 datasets with respect to the BVH dataset. At first, we resized each slice of the HRF and M2 datasets as per the standard of BVH, and then we employed the proposed technique (CARGS) with the same classifiers shown in Table 8.
A very promising result was observed with a variation of 96.8333–98.8333% of classification accuracy, as shown in Figure 14.
Experimental results were observed to vary due to variations in the multi-institutional RF image datasets. We must support the development of a dataset of the global platform for medical patient data in which—despite differences in patient mode, region, demography, geography, and medical history—we can more accurately address these medical health issues. A comparison graph between the BVH and public RF image datasets is shown in Figure 15.
A comparison between the proposed methodology and the current state-of-the-art techniques in is Table 9.

4. Conclusions

This study focused on the classification of four DR stages (mild, moderate, non-proliferative, and proliferative), as well as the normal human retina, with the help of fused hybrid-feature analysis using fundus images by applying ML classifiers. The main goal was K-mean clustering-based segmentation, the selection of appropriately optimized hybrid-features, and the identification of suitable classifiers for well-organized classification. The variation of results was due to different modalities of texture analysis. Four types of features were extracted, namely histogram, wavelet, co-occurrence matrix, and run-length matrix. In the end, a fused hybrid-feature dataset, which is a combination of the above-mentioned extracted features, was generated. The fused hybrid-feature dataset was generated by using a data fusion approach. A large number of features can minimize the accuracy of classification and maximize total execution time, so the main focus was applied to the feature optimization process and used with a post-optimized hybrid-feature dataset. This study concluded that more accurate and precise result can be achieved by applying five ML classifiers—SLg, LMT, MLP, Lg, and SMO—on post-optimized hybrid-feature datasets. All the deployed classifiers showed promising results, but the SLg classifier results were outstandingly high. After employing the SLg classifier, it was observed that an overall accuracy of 99.73% was achieved. The accuracies obtained by SLg on the five DR classes of healthy, mild, moderate, non-proliferative, and proliferative were 99.93%, 99.73%, 99.86%, 99.46%, and 99.66%, respectively. The proposed model was verified on publicly available, high-resolution fundus, and Messidor-2 datasets. A very promising result was observed with a variation of 96.83–98.83% of classification accuracy.

Future Work

In the future, this study will assist in building up an image dynamic searching approach with deep learning algorithms to increase the momentum for classification. This technique can be further improved by using the 3D visualization of volumetric RF data. Among the planed future works, clinical applicability will be observed for this proposed model.

Author Contributions

A.A.: conceptualization, methodology, writing—original draft, and software. S.Q.: supervision. W.K.M.: project administration. W.K.: validation. P.K.: resources. S.N.: data curation. A.G.: investigation. F.J.: formal analysis and writing—review and editing. C.C.: validation and writing—review and editing. S.A.: visualization. M.S.: visualization. All authors have read and agreed to the published version of the manuscript.

Funding

Program in Applied Statistics, Department of Mathematics and Computer Science, Faculty of Science and Technology, Rajamangala University of Technology Thanyaburi (RMUTT), Thanyaburi, Pathumthani 12110, Thailand.

Acknowledgments

This work was supported by by Program in Applied Statistics, Department of Mathematics and Computer Science, Faculty of Science and Technology, Rajamangala University of Technology Thanyaburi (RMUTT), Thanyaburi, Pathumthani 12110, Thailand, who the authors would like to thank for supporting this research under the grant number 5892. Additionally, authors would like to thank the referees for their careful reading and for their comments, which significantly improved the paper.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Duh, E.J.; Sun, J.K.; Stitt, A.W. Diabetic retinopathy: Current understanding, mechanisms, and treatment strategies. Jci Insight 2017, 2, 93751. [Google Scholar] [CrossRef] [PubMed]
  2. Sohn, E.H.; van Dijk, H.W.; Jiao, C.; Kok, P.H.; Jeong, W.; Demirkaya, N.; Garmager, A.; Wit, F.; Kucukevcilioglu, M.; van Velthoven, M.E.; et al. Retinal neurodegeneration may precede microvascular changes characteristic of diabetic retinopathy in diabetes mellitus. Proc. Natl. Acad. Sci. USA 2016, 113, E2655–E2664. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  3. Lee, R.; Wong, T.Y.; Sabanayagam, C. Epidemiology of diabetic retinopathy, diabetic macular edema and related vision loss. Eye Vis. 2015, 2, 17. [Google Scholar] [CrossRef] [Green Version]
  4. Nirmala, S.R.; Nath, M.K.; Dandapat, S. Retinal image analysis: A review. Int. J. Comput. Commun. Technol. 2011, 2, 11–15. [Google Scholar]
  5. Pietsch, P. The Quest of Hologramic Mind; Houghton Mifflin Harcourt: Boston, MA, USA, 1981; ISBN 10: 0395294800. [Google Scholar]
  6. Nayak, J.; Bhat, P.S.; Acharya, R.; Lim, C.M.; Kagathi, M. Automated identification of diabetic retinopathy stages using digital fundus images. J. Med. Syst. 2008, 32, 107–115. [Google Scholar] [CrossRef]
  7. Jitpakdee, P.; Pakinee, A.; Bunyarit, U. A survey on hemorrhage detection in diabetic retinopathy retinal images. In Proceedings of the 2012 9th International Conference on Electrical Engineering/Electronics, Computer, Telecommunications and Information Technology, Phetchaburi, Thailand, 16–18 May 2012; pp. 1–4. [Google Scholar]
  8. Antal, B.; Hajdu, A. An ensemble-based system for microaneurysm detection and diabetic retinopathy grading. IEEE Trans. Biomed. Eng. 2012, 59, 1720–1726. [Google Scholar] [CrossRef] [Green Version]
  9. Kruthiventi, S.S.S.; Ayush, K.; Babu, R.V. Deepfix: A fully convolutional neural network for predicting human eye fixations. IEEE Trans. Image Process. 2017, 26, 4446–4456. [Google Scholar] [CrossRef] [Green Version]
  10. Pires, R.; Avila, S.; Wainer, J.; Valle, E.; Abramoff, M.D.; Rocha, A. A data-driven approach to referable diabetic retinopathy detection. Artif. Intell. Med. 2019, 96, 93–106. [Google Scholar] [CrossRef]
  11. Zhang, W.; Zhong, J.; Yang, S.; Gao, Z.; Hu, J.; Chen, Y.; Yi, Z. Automated identification and grading system of diabetic retinopathy using deep neural networks. Knowl. Based Syst. 2019, 175, 12–25. [Google Scholar] [CrossRef]
  12. Harun, N.H.; Yusof, Y.; Hassan, F.; Embong, Z. Classification of fundus images for diabetic retinopathy using artificial neural network. In Proceedings of the 2019 IEEE Jordan International Joint Conference on Electrical Engineering and Information Technology (JEEIT), Amman, Jordan, 9–11 April 2019; pp. 498–501. [Google Scholar]
  13. Verbraak, F.D.; Abramoff, M.D.; Bausch, G.C.F.; Klaver, C.; Nijpels, G.; Schlingemann, R.O.; van der Heijden, A.A. Diagnostic accuracy of a device for the automated detection of diabetic retinopathy in a primary care setting. Diabetes Care 2019, 42, 651–656. [Google Scholar] [CrossRef] [Green Version]
  14. Rubya, A.; Shill, P.C. Automatic lesions detection and classification of diabetic retinopathy using fuzzy logic. In Proceedings of the 2019 International Conference on Robotics, Electrical and Signal Processing Techniques (ICREST), Dhaka, Bangladesh, 10–12 January 2019; pp. 527–532. [Google Scholar]
  15. Ricky, P.; Lakshmanan, R.; Purushotham, S.; Soundrapandiyan, R. Detecting Diabetic Retinopathy from Retinal Images Using CUDA Deep Neural Network. Intell. Pervasive Comput. Syst. Smarter Healthc. 2019, 379–396. [Google Scholar] [CrossRef]
  16. Kele, X.; Feng, D.; Mi, H. Deep convolutional neural network-based early automated detection of diabetic retinopathy using fundus image. Molecules 2017, 22, 2054. [Google Scholar]
  17. Sim, D.A.; Keane, P.A.; Tufail, A.; Egan, C.A.; Aiello, L.P.; Silva, P.S. Automated retinal image analysis for diabetic retinopathy in telemedicine. Curr. Diabetes Rep. 2015, 15, 14. [Google Scholar] [CrossRef] [PubMed]
  18. Gulshan, V.; Peng, L.; Coram, M.; Stumpe, M.C.; Wu, D.; Narayanaswamy, A.; Venugopalan, S.; Widner, K.; Madams, T.; Cuadros, J.; et al. Development and validation of a deep learning algorithm for detection of diabetic retinopathy in retinal fundus photographs. JAMA 2016, 316, 2402–2410. [Google Scholar] [CrossRef]
  19. Rishab, G.; Leng, T. Automated identification of diabetic retinopathy using deep learning. Ophthalmology 2017, 124, 962–969. [Google Scholar]
  20. Quaid-e-Azam Medical College. Bahawal Victoria Hospital—Quaid-e-Azam Medical College. 2020. Available online: https://www.qamc.edu.pk/bahawalvictoriahospital/ (accessed on 22 February 2019).
  21. Zhang, X.; Cui, J.; Wang, W.; Lin, C. A study for texture feature extraction of high-resolution satellite images based on a direction measure and gray level co-occurrence matrix fusion algorithm. Sensors 2017, 17, 1474. [Google Scholar] [CrossRef] [Green Version]
  22. Wang, G.; Gao, Z.; Zhang, Y.; Ma, B. Adaptive maximum correntropy Gaussian filter based on variational Bayes. Sensors 2018, 18, 1960. [Google Scholar] [CrossRef] [Green Version]
  23. Meng, X.; Yin, Y.; Yang, G.; Xi, X. Retinal identification based on an improved circular gabor filter and scale invariant feature transform. Sensors 2013, 13, 9248–9266. [Google Scholar] [CrossRef] [Green Version]
  24. Guo, Y.; Ashour, A.S.; Smarandache, F. A novel skin lesion detection approach using neutrosophic clustering and adaptive region growing in dermoscopy images. Symmetry 2018, 10, 119. [Google Scholar] [CrossRef] [Green Version]
  25. Zhang, C.; Xiao, X.; Li, X.; Chen, Y.; Zhen, W.; Chang, J.; Zheng, C.; Liu, Z. White blood cell segmentation by color-space-based k-means clustering. Sensors 2014, 14, 16128–16147. [Google Scholar] [CrossRef]
  26. Chen, C.W.; Luo, J.; Parker, K.J. Image segmentation via adaptive K-mean clustering and knowledge-based morphological operations with biomedical applications. IEEE Trans. Image Process. 1998, 7, 1673–1683. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  27. Ng, H.P.; Ong, S.H.; Foong, K.W.C.; Goh, P.S.; Nowinski, W.L. Medical image segmentation using k-means clustering and improved watershed algorithm. In Proceedings of the 2006 IEEE Southwest Symposium on Image Analysis and Interpretation, Denver, CO, USA, 26–28 March 2006; pp. 61–65. [Google Scholar]
  28. Galloway, M. Texture analysis using gray level run lengths. Comput. Graph. Image Process 1975, 4, 172–179. [Google Scholar] [CrossRef]
  29. Dasarathy, B.V.; Holder, E.B. Image characterizations based on joint gray level—Run length distributions. Pattern Recognit. Lett. 1991, 12, 497–502. [Google Scholar] [CrossRef]
  30. Garali, I.; Adel, M.; Bourennane, S.; Guedj, E. Histogram-based features selection and volume of interest ranking for brain PET image classification. IEEE J. Transl. Eng. Health Med. 2018, 6, 1–12. [Google Scholar] [CrossRef]
  31. Beura, S.; Majhi, B.; Dash, R. Mammogram classification using two-dimensional discrete wavelet transform and gray-level co-occurrence matrix for detection of breast cancer. Neurocomputing 2015, 154, 1–14. [Google Scholar] [CrossRef]
  32. Kociołek, M.; Materka, A.; Strzelecki, M.; Szczypiński, P. Discrete wavelet transform-derived features for digital image texture analysis. In Proceedings of the International Conference on Signals and Electronic Systems, Lodz, Poland, 18–21 September 2001; pp. 99–104. [Google Scholar]
  33. Mateen, M.; Wen, J.; Song, S.; Huang, Z. Fundus image classification using VGG-19 architecture with PCA and SVD. Symmetry 2019, 11, 1. [Google Scholar] [CrossRef] [Green Version]
  34. Qadri, S.; Khan, D.M.; Qadri, S.F.; Razzaq, A.; Ahmad, N.; Jamil, M.; Shah, A.N.; Muhammad, S.S.; Saleem, K.; Awan, S.A. Multisource Data Fusion Framework for Land Use/Land Cover Classification Using Machine Vision. J. Sens. 2017, 2017, 1–8. [Google Scholar] [CrossRef] [Green Version]
  35. Strzelecki, M.; Szczypinski, P.; Materka, A.; Klepaczko, A. A software tool for automatic classification and segmentation of 2D/3D medical images. Nucl. Instrum. Methods Phys. Res. A 2013, 702, 137–140. [Google Scholar] [CrossRef]
  36. Hall, M.A. Correlation-Based Feature Selection for Machine Learning. Ph.D. Thesis, The University of Waikato, Hamilton, New Zealand, 1999; pp. 1–198. [Google Scholar]
  37. The Islamia University of Bahawalpur “IUB” Pakistan. 2020. Available online: https://www.iub.edu.pk/ (accessed on 10 November 2019).
  38. High-Resolution Fundus (hrf) Image Database. Available online: https://www5.cs.fau.de/research/data/fundus-images/ (accessed on 11 January 2019).
  39. MESSIDOR-2. Digital Retinal Images, LaTIM Laboratory, France. (MESSIDOR-2 Section). 2020. Available online: http://latim.univ-brest.fr/ (accessed on 15 December 2018).
Figure 1. Typical retinal fundus images.
Figure 1. Typical retinal fundus images.
Entropy 22 00567 g001
Figure 2. Clustering-based segmentation and hybrid-feature analysis for diabetic retinopathy (DR) classification framework.
Figure 2. Clustering-based segmentation and hybrid-feature analysis for diabetic retinopathy (DR) classification framework.
Entropy 22 00567 g002
Figure 3. Irregular polygonal seed selection proposed novel mechanism.
Figure 3. Irregular polygonal seed selection proposed novel mechanism.
Entropy 22 00567 g003
Figure 4. Irregular polygonal seed-based improved region growing outcome using K-mean clustering.
Figure 4. Irregular polygonal seed-based improved region growing outcome using K-mean clustering.
Entropy 22 00567 g004
Figure 5. DR segmentation based on the watershed technique.
Figure 5. DR segmentation based on the watershed technique.
Entropy 22 00567 g005
Figure 6. Data visualization in 3D vector space.
Figure 6. Data visualization in 3D vector space.
Entropy 22 00567 g006
Figure 7. The overall accuracy graph of the employed ML classifiers on a histogram features-based dataset.
Figure 7. The overall accuracy graph of the employed ML classifiers on a histogram features-based dataset.
Entropy 22 00567 g007
Figure 8. The overall accuracy graph of the employed ML classifiers on a wavelet features-based dataset.
Figure 8. The overall accuracy graph of the employed ML classifiers on a wavelet features-based dataset.
Entropy 22 00567 g008
Figure 9. The overall accuracy graph of the employed ML classifiers on a COM features-based dataset.
Figure 9. The overall accuracy graph of the employed ML classifiers on a COM features-based dataset.
Entropy 22 00567 g009
Figure 10. The overall accuracy graph of the employed ML classifiers on an RLM features-based dataset.
Figure 10. The overall accuracy graph of the employed ML classifiers on an RLM features-based dataset.
Entropy 22 00567 g010
Figure 11. The overall accuracy graph of the employed ML classifiers on a post-optimized fused dataset.
Figure 11. The overall accuracy graph of the employed ML classifiers on a post-optimized fused dataset.
Entropy 22 00567 g011
Figure 12. Classification accuracy graph of five DR stages using the SLg classifier on a post-optimized fused dataset.
Figure 12. Classification accuracy graph of five DR stages using the SLg classifier on a post-optimized fused dataset.
Entropy 22 00567 g012
Figure 13. Comparative analysis of DR stage classification accuracies among different type of feature modalities.
Figure 13. Comparative analysis of DR stage classification accuracies among different type of feature modalities.
Entropy 22 00567 g013
Figure 14. The overall accuracy graph of the employed ML classifiers on post-optimized, fused public dataset.
Figure 14. The overall accuracy graph of the employed ML classifiers on post-optimized, fused public dataset.
Entropy 22 00567 g014
Figure 15. A comparison accuracy graph between Bahawal Victoria Hospital (BVH) and public (RF image) dataset with the SLg classifier.
Figure 15. A comparison accuracy graph between Bahawal Victoria Hospital (BVH) and public (RF image) dataset with the SLg classifier.
Entropy 22 00567 g015
Table 1. Post-optimize feature selection table (Fisher (F), probability of error (POE) plus average correlation (AC), mutual information (MI), and correlation-based feature selection (CFS)).
Table 1. Post-optimize feature selection table (Fisher (F), probability of error (POE) plus average correlation (AC), mutual information (MI), and correlation-based feature selection (CFS)).
FeaturesRun LengthCo-OccurrenceWaveletHistogram
1. 45dgr_RLNonUni
2. 135dr_RLNonUni
3. Horzl_RLNonUni
4. Vertl_RLNonUni
5. 45dgr_GLevNonU
6. S(5,0)Entropy
7. S(5,5)Contrast
8. S(5,-5)Contrast
9. WavEnHL_s-4
10. WavEnLL_s-4
11. Variance
12. Skewness
13. Perc.01%
Table 2. The overall classification accuracy table of the employed machine learning (ML) classifiers on a histogram features-based dataset. ROC: receiver-operating characteristic; TP: true positive; FR: false positive; MAE: mean absolute errors; RMSE: root mean squared error; OA: overall accuracy; MLP: multi-layer perceptron; SLg: simple logistic; LMT: logistic model tree; Lg: logistic; and SMO: sequential minimal optimization.
Table 2. The overall classification accuracy table of the employed machine learning (ML) classifiers on a histogram features-based dataset. ROC: receiver-operating characteristic; TP: true positive; FR: false positive; MAE: mean absolute errors; RMSE: root mean squared error; OA: overall accuracy; MLP: multi-layer perceptron; SLg: simple logistic; LMT: logistic model tree; Lg: logistic; and SMO: sequential minimal optimization.
ClassifiersKappa StatisticsTP RateFP RateROCMAERMSETime (s)OA
MLP0.67170.7370.0660.9160.14850.27990.7873.73%
SLg0.66330.7310.0670.9210.16370.28130.5973.07%
LMT0.66250.7300.0680.9190.15340.28070.5273.00%
Lg0.66170.7290.0680.9230.15560.27890.3772.07%
SMO0.60750.6860.090.8780.26020.34650.2568.60%
Table 3. The overall classification accuracy table of the employed ML classifiers on a wavelet features-based dataset.
Table 3. The overall classification accuracy table of the employed ML classifiers on a wavelet features-based dataset.
ClassifiersKappa StatisticsTP RateFP RateROCMAERMSETime (s)OA
MLP0.74920.7990.0500.9430.08940.25591.0279.93%
LMT0.72250.7780.0560.9330.10720.25711.9877.80%
Lg0.71080.7690.0580.9440.12990.25810.0476.86%
SLg0.70750.7660.0590.9450.13620.25820.9876.60%
SMO0.69750.7580.0610.9140.25330.33670.0975.80%
Table 4. The overall classification accuracy table of the employed ML classifiers on co-occurrence matrix (COM) features-based dataset.
Table 4. The overall classification accuracy table of the employed ML classifiers on co-occurrence matrix (COM) features-based dataset.
ClassifiersKappa StatisticsTP RateFP RateROCMAERMSETime (s)OA
LMT0.87420.8990.0250.9820.0570.17550.9989.93%
MLP0.87330.8990.0250.8910.0460.18651.2189.86%
SLg0.84830.8790.0300.9810.0750.18860.4387.86%
SMO0.8350.8680.0330.9500.2470.32690.1286.80%
Lg0.85830.8870.0280.9780.0510.18960.0986.67%
Table 5. The overall classification accuracy table of the employed ML classifiers on a run-length matrix (RLM) features-based dataset.
Table 5. The overall classification accuracy table of the employed ML classifiers on a run-length matrix (RLM) features-based dataset.
ClassifiersKappa StatisticsTP RateFP RateROCMAERMSETime (s)OA
SLg0.94920.9590.0100.9960.03070.11670.6295.93%
Lg0.94750.9580.0110.9960.02470.11660.7295.80%
LMT0.94670.9570.0110.9940.02920.11860.8295.73%
MLP0.93920.9510.0120.9940.02920.12420.9895.13%
SMO0.92080.9370.0160.9780.2430.32090.0693.67%
Table 6. The overall classification accuracy table of the employed ML classifiers on a post-optimized, fused hybrid-features dataset.
Table 6. The overall classification accuracy table of the employed ML classifiers on a post-optimized, fused hybrid-features dataset.
ClassifiersKappa StatisticsTP RateFP RateROCMAERMSETime (s)OA
SLg0.99670.9970.0011.0000.00510.03670.3299.73%
LMT0.99670.9970.0011.0000.00480.03530.5899.73%
MLP0.99580.9970.0011.0000.00380.03230.3399.67%
Lg0.98920.9910.0020.9990.00360.05660.4299.13%
SMO0.98170.9850.0040.9960.24060.31720.2198.53%
Table 7. The confusion matrix of post-optimized, fused hybrid-feature dataset for the SLg classifier.
Table 7. The confusion matrix of post-optimized, fused hybrid-feature dataset for the SLg classifier.
Classified asHealthyMildModerateNon-ProliferativeProliferativeTotal
Healthy149900001500
Mild014960001500
Moderate101498001500
Non-Proliferative000149211500
Proliferative010114951500
Table 8. The overall classification accuracy table of the employed ML classifiers on a post-optimized, fused public dataset.
Table 8. The overall classification accuracy table of the employed ML classifiers on a post-optimized, fused public dataset.
ClassifiersKappa StatisticsTP RateFP RateROCMAERMSETime (s)OA
SLg0.97670.9880.0121.0000.01910.09140.3898.83%
LMT0.97330.9870.0130.9850.01650.11370.1898.67%
MLP0.96670.9830.0170.9980.02680.10890.2898.33%
Lg0.94330.9720.0280.9980.03130.14040.3997.17%
SMO0.93670.9680.0320.9970.03020.15420.1196.83%
Table 9. A comparison table among the proposed and current state-of-the-art techniques.
Table 9. A comparison table among the proposed and current state-of-the-art techniques.
Source/ReferenceMethodologyModalityAccuracy
Pires, R. et al. [10]Convolutional Neural NetworksRF Image99.0%
Zhang, W. et al. [11]Neural NetworksRF Image98.1%
Harun, N. H. et al. [12]MLP and Artificial Neural NetworkRF Image72.11%
Verbraak, F. D. et al. [13]Hybrid Features and Deep LearningRF Image93.8%
Afrin, R. and Shill, P. C. [14]Fused Feature and Fuzzy LogicRF Image95.63%
Parmar, R. et al. [15]Neural NetworksRF Image85%
Xu, K. et al. [16]Neural NetworksRF Image94.5%
Gulshan, V. et al. [18]Deep LearningRF Image97.5%
Gargeya R. [19]Data-Driven Deep Learning AlgorithmRF Image94%
Proposed MethodologyCARGS, Post-Optimized, Fused Hybrid-Features, and Simple LogisticBVH RF Image Dataset 99.73%
Proposed Methodology
(Validation)
CARGS, Post-Optimized, Fused Hybrid-Features, and Simple LogisticPublicly
Available
Dataset (RF-Image)
98.83%

Share and Cite

MDPI and ACS Style

Ali, A.; Qadri, S.; Khan Mashwani, W.; Kumam, W.; Kumam, P.; Naeem, S.; Goktas, A.; Jamal, F.; Chesneau, C.; Anam, S.; et al. Machine Learning Based Automated Segmentation and Hybrid Feature Analysis for Diabetic Retinopathy Classification Using Fundus Image. Entropy 2020, 22, 567. https://doi.org/10.3390/e22050567

AMA Style

Ali A, Qadri S, Khan Mashwani W, Kumam W, Kumam P, Naeem S, Goktas A, Jamal F, Chesneau C, Anam S, et al. Machine Learning Based Automated Segmentation and Hybrid Feature Analysis for Diabetic Retinopathy Classification Using Fundus Image. Entropy. 2020; 22(5):567. https://doi.org/10.3390/e22050567

Chicago/Turabian Style

Ali, Aqib, Salman Qadri, Wali Khan Mashwani, Wiyada Kumam, Poom Kumam, Samreen Naeem, Atila Goktas, Farrukh Jamal, Christophe Chesneau, Sania Anam, and et al. 2020. "Machine Learning Based Automated Segmentation and Hybrid Feature Analysis for Diabetic Retinopathy Classification Using Fundus Image" Entropy 22, no. 5: 567. https://doi.org/10.3390/e22050567

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop