Classiﬁcation of Alzheimer’s Disease Patients Using Texture Analysis and Machine Learning

: Alzheimer’s disease (AD) has been studied extensively to understand the nature of this complex disease and address the many research gaps concerning prognosis and diagnosis. Several studies based on structural and textural characteristics have already been conducted to aid in identifying AD patients. In this work, an image processing methodology was used to extract textural information and classify the patients into two groups: AD and Cognitively Normal (CN). The Gray Level Co-occurrence Matrix (GLCM) was employed since it is a strong foundation for texture classiﬁcation. Various textural parameters derived from the GLCM aided in deciphering the characteristics of a Magnetic Resonance Imaging (MRI) region of interest (ROI). Several commonly used image classiﬁcation algorithms were employed. MATLAB was used to successfully derive 20 features based on the GLCM of the MRI dataset. Based on the data analysis, 8 of the 20 features were determined as signiﬁcant elements. Ensemble (90.2%), Decision Trees (88.5%), and Support Vector Machine (SVM) (87.2%) were the best performing classiﬁers. It was observed in GLCM that as the distance (d) between pixels increased, the classiﬁcation accuracy decreased. The best result was observed for GLCM with d = 1 and direction ( d , d , − d ) with age and structural data.


Introduction
Alzheimer's Disease (AD) has been extensively studied to better understand this complex disease's nature and address the numerous unresolved challenges concerning prognosis and diagnosis. According to current projections, the number of people with dementia globally will rise to 75 million by 2030, up from 47 million in 2020.
The hippocampus is vital in developing, structuring, and storing new memories and the association of sensations and feelings with some of these memories. AD causes gradual tissue loss throughout the brain. Hippocampal atrophy is a significant biomarker used to assess and detect AD. The hippocampus undergoes multiple small-scale alterations before a significant change occurs. These changes are visible in clinical magnetic resonance imaging (MRI) at the current resolution. Previous research has demonstrated that texturebased classification produces better results than volume-based classification [1]. The hippocampus appears to demonstrate the most rapid loss of tissue early in the disease course. The increasing atrophy of the hippocampus is responsible for Alzheimer's-related short-term memory loss [2]. MRI is a highly effective technology for nonaggressive in vivo imaging of the human brain. A widely proven approach is used for robust cross-sectional and longitudinal segmentation of Alzheimer's Disease Neuroimaging.
Initiative Alzheimer's Disease Neuroimaging Initiative (ADNI) cohort brain MRI. Structural Magnetic Resonance Imaging (sMRI) detects subtle anatomical changes in the brain to detect AD [3]. MRI is favoured over other modalities for detecting structural changes in the brain caused by neurodegenerative diseases and its significance in AD diagnosis and prognosis [4]. In most sMRI investigations, large-scale tissue changes have been used as biomarkers based on derived features like atrophy to acquire a better diagnosis of AD. Subtle neurodegenerative changes can be noticed and analyzed using statistical properties acquired from texture analysis on an MRI based on a biomarker.
Medical imaging is obtaining images of body parts for diagnostic applications such as detecting or studying diseases. Every week, millions of imaging tasks are performed around the world. Medical imaging is quickly evolving because of advances in image processing methods such as image recognition, analysis, and enhancement. Image processing improves the proportion and quantity of recognized tissues.
Machine learning is a pattern recognition technique that can be used in medical imaging. It is a vital tool that can aid in medical diagnosis. Typically, machine learning begins with the machine learning algorithm system determining the image attributes that are thought to be important in producing the desired prediction or diagnosis. The machine learning algorithm system then determines the optimum solution of these image attributes for classification.
The contributions of this research work are: • To use texture analysis to examine the minor changes in the hippocampus caused by AD.

•
To extract features by implementing texture analysis technique on MRI. • To classify subjects into AD and cognitively normal (CN) groups by providing the information extracted as input for training machine learning models.
An extensive literature survey was performed, which helped delimit the research domain, and focus is directed to the crucial issues to be resolved. Gray Level Co-occurrence Matrix (GLCM) has been verified to be a standard statistical process of extricating textural characteristics from illustrations. The number of grey levels in the image are equivalent to the rows and columns of a GLCM. GLCM was used to acquire textural features. Angular Second Moment (ASM), Inverse Difference Moment, Entropy and Correlation were the second-order features calculated using the GLCM. High discrimination accuracy was achieved by offering these features for image approximation [5]. Xia, H. et al., considered the bilateral hippocampus and medial temporal lobe (MTL) for volumes and texture features in AD subject's MRI. Texture features were extracted using GLCM on 66 subjects; 29 of these were AD subjects. Texture features and corresponding volumes were analyzed for correlation between them. Volumes and texture features between the AD and CN group in both the hippocampus and MTL were found to be significantly different. The texture features matching to volumes of both positions were correlated. The CN group was further classified into the non-demented (EC) and young control (YC) group. Texture analysis and volumetry may aid in AD diagnosis, indicating no significant difference in CN groups. It implies that atrophy is observed to be distinct in AD and be an age and AD-associated change [6].
Mohanaiah, P. et al., describes how to use the GLCM to extract texture features for image motion approximation. The Xilinx FPGA was used to compute the four features: ASM, Correlation, Entropy, and Inverse Difference Moment. The findings confirm that the texture features have excellent discriminating precision, take minimal calculation time, and can thus be employed effectively for real-time pattern recognition applications [7]. Fuse, H. et al. investigated the efficacy of classifying CN and AD patients relying on brain texture features. As for shape features, a P-type Fourier descriptor was used, and the lateral ventricle except for the septum lucidum was studied. They had used SVM to conduct classification utilizing a collection of multiple characteristics as features. The classifier accuracy was 87.5%, which was higher than the accuracy reached using volume ratio to intracranial volume (81.5%), which is frequently recognized for traditional evaluation of degenerative characteristics. As per recent research, shape information may be more relevant in diagnosis than traditional volume ratio [1]. Telagarapu P. et al. examined AD in T1-weighted MRI combining Texture Features and the k-NN Classifier. Texture features were obtained via GLCM. With the sole purpose of evaluating the suggested approach, tests were undertaken on the MRI obtained from the OASIS dataset. The presented method has a typical testing accuracy of 74.73%, suggesting that it can differentiate AD and CN effectively [3]. Martinez-Murcia et al. intended to better understand the disease's development by evaluating neuropsychological test results based purely on MRI data. To accomplish this, they used a deep convolutional auto encoder to do a self-supervised segmentation of the MRI data. The dispersion of features in the z-layer was then employed in a neural regression to a complex linear relationship. Then to see if these imaging-derived indicators are associated with various neuropsychological tests. R 2 rates of more than 0.3 were achieved in predicting neuropsychological analysis results, with correlations greater than 0.5 in the case of variables significantly associated with neurodegeneration and cognitive states, such as the MMSE or the ADASS11 scores [8][9][10].

Alzheimer's Disease Neuroimaging Initiative
The ADNI database was used to collect data for this inquiry; as shown in Figure 1, collecting the necessary database was the first stage of this endeavour. ADNI began in 2003 as a public-private partnership directed by lead investigator Michael W. Weiner, MD. The key objective of ADNI would be to see if serial MRI, PET, other biological markers, and clinical and cognitive tests could be used to track the development of early AD. The ADNI research is a multisite longitudinal study aimed at finding clinical, imaging, genetic, and biochemical indicators for initial tracking and AD identification. More than a decade has passed since its beginnings; it has made significant contributions to AD research, enabling researchers worldwide to share data [11]. the lateral ventricle except for the septum lucidum was studied. They had used SVM to conduct classification utilizing a collection of multiple characteristics as features. The classifier accuracy was 87.5%, which was higher than the accuracy reached using volume ratio to intracranial volume (81.5%), which is frequently recognized for traditional evaluation of degenerative characteristics. As per recent research, shape information may be more relevant in diagnosis than traditional volume ratio [1]. Telagarapu P. et al. examined AD in T1-weighted MRI combining Texture Features and the k-NN Classifier. Texture features were obtained via GLCM. With the sole purpose of evaluating the suggested approach, tests were undertaken on the MRI obtained from the OASIS dataset. The presented method has a typical testing accuracy of 74.73%, suggesting that it can differentiate AD and CN effectively [3]. Martinez-Murcia et al. intended to better understand the disease's development by evaluating neuropsychological test results based purely on MRI data. To accomplish this, they used a deep convolutional auto encoder to do a self-supervised segmentation of the MRI data. The dispersion of features in the z-layer was then employed in a neural regression to a complex linear relationship. Then to see if these imaging-derived indicators are associated with various neuropsychological tests. R 2 rates of more than 0.3 were achieved in predicting neuropsychological analysis results, with correlations greater than 0.5 in the case of variables significantly associated with neurodegeneration and cognitive states, such as the MMSE or the ADASS11 scores [8][9][10].

Alzheimer's Disease Neuroimaging Initiative
The ADNI database was used to collect data for this inquiry; as shown in Figure 1, collecting the necessary database was the first stage of this endeavour. ADNI began in 2003 as a public-private partnership directed by lead investigator Michael W. Weiner, MD. The key objective of ADNI would be to see if serial MRI, PET, other biological markers, and clinical and cognitive tests could be used to track the development of early AD. The ADNI research is a multisite longitudinal study aimed at finding clinical, imaging, genetic, and biochemical indicators for initial tracking and AD identification. More than a decade has passed since its beginnings; it has made significant contributions to AD research, enabling researchers worldwide to share data [11].

MRI Data and Subjects
The Image and Data Archive (IDA) search returns raw, unprocessed images based on various selection parameters. Few criteria were used, including a study group in which AD and CN groups were chosen with T1 weighted MRI with slice thickness between 1 mm to 2.5 mm. Three-dimensional Magnetization-prepared Rapid Acquisition Gradient Echo (MPRAGE) MRI with a field strength of 3T is specified for the acquisition. When evaluating texture features extracted from a 1.5T MRI to texture features retrieved from a

MRI Data and Subjects
The Image and Data Archive (IDA) search returns raw, unprocessed images based on various selection parameters. Few criteria were used, including a study group in which AD and CN groups were chosen with T1 weighted MRI with slice thickness between 1 mm to 2.5 mm. Three-dimensional Magnetization-prepared Rapid Acquisition Gradient Echo (MPRAGE) MRI with a field strength of 3T is specified for the acquisition. When evaluating texture features extracted from a 1.5T MRI to texture features retrieved from a 3T MRI, it is feasible to identify even more changes in texture characteristics, which could lead to an earlier diagnosis [12,13]. Instead of the original acquisition format (Digital Imaging and Communications in Medicine DICOM), all the MRIs were downloaded in Nifti format [14]. Table 1 describes the Demographical data of the subjects in details The ADNI dataset included MRIs from 234 subjects, with 119 (50.9%) being AD and 115 (49.1%) being CN. The youngest participants were 55 and 59 years old in the AD and CN groups, respectively, while the oldest were 91 and 90.
The dataset was divided into 70% for training and 30% for testing. While training all the all the models, 5-fold cross-validation is used to avoid overfitting. In this technique, the data was split into k = 5 subsets, where one subset is treated as a validation set for the model and k−1, i.e., four subsets for training the model. The data samples were single-trained and then split for validation testing.

ROI Selection
volBrain is an MRI brain volumetry system available online. volBrain is a fully automated system that can deliver brain structure volumes without the need for human participation. volBrain is mainly composed of image processing tasks that strive to increase the performance of the input images and organize them in a predefined geometric and intensity space so that they may be segmented later [15]. Figure 2 shows the functional block scheme of volBrain. It comprises the HIPS pipeline devised to cope with high-resolution or standard T1w images (Spoiled Gradient Echo (SPGR) and MPRAGE at 1.5T and 3T, respectively) without any preprocessing. T1 weighted sequences are used substantially in all MRI treatments and are best described as the most "anatomical" of images, that closely resemble the looks of tissues macroscopically. HIPS considers the use of the Winterburn segmentation protocol, which is freely available. MOPAL, which performs quick segmentation, offers accurate findings in less than twenty minutes and provides a mask of Winterburn segmentation protocol with reports stating the volumes of all the subfields and the total volume of the structure. HIPS exceeds other state-of-the-art algorithms regarding segmentation accuracy, earning a DICE of 0.661 for the Winterburn dataset. CA1, CA2/CA3, CA4/DG, SR/SL/SM, and Subiculum are the five subfields in the Winterburn dataset, as shown in Figure 3. volBrain is one of the first few freely available platforms that offer hippocampus segmentation based on the European Alzheimer's Disease Consortium (EADC) protocol, which will become the standard definition for AD in the next few years [15,16]. In terms of volume and shape estimation, volBrain is deemed more reliable than Freesurfer and FIRST [16].

Hippocampus Segmentation
MRIcron is a cross-platform image viewer for the Nifti format. It can import several layers of images, produce volume renderings, and draw volumes of interest. It also incorporates Node Package Manager (NPM) for statistics and dcm2nii for converting DICOM images to Nifti format. We employed MRIcron to segment the data [17,18]. The complete brain scan is the data set obtained from ADNI. MRI data were processed by this program to obtain the ROI, which yielded only the hippocampus region with its original grey intensities preserved. We overlaid the Winterburn segmentation mask generated by volBrain on the original MRI and maintained the area beneath the mask, removing the pixels beyond the mask boundary as shown in Figure 4.

Texture Analysis
The texture is a crucial feature for recognizing ROI in an image. Haralick et al. [5] introduced GLCM as one of the first methods for texture feature extraction in 1973. For texture analysis, GLCM was used. The two pixels are considered at a time namely the reference and the neighbour pixel. The grey level distribution of pairs of pixels in a specific distance and image configuration was analyzed by GLCM, reflecting the texture or spatial

Hippocampus Segmentation
MRIcron is a cross-platform image viewer for the Nifti format. It can import several layers of images, produce volume renderings, and draw volumes of interest. It also incorporates Node Package Manager (NPM) for statistics and dcm2nii for converting DICOM images to Nifti format. We employed MRIcron to segment the data [17,18]. The complete brain scan is the data set obtained from ADNI. MRI data were processed by this program to obtain the ROI, which yielded only the hippocampus region with its original grey intensities preserved. We overlaid the Winterburn segmentation mask generated by volBrain on the original MRI and maintained the area beneath the mask, removing the pixels beyond the mask boundary as shown in Figure 4.

Texture Analysis
The texture is a crucial feature for recognizing ROI in an image. Haralick et al. [5] introduced GLCM as one of the first methods for texture feature extraction in 1973. For texture analysis, GLCM was used. The two pixels are considered at a time namely the reference and the neighbour pixel. The grey level distribution of pairs of pixels in a specific distance and image configuration was analyzed by GLCM, reflecting the texture or spatial

Hippocampus Segmentation
MRIcron is a cross-platform image viewer for the Nifti format. It can import several layers of images, produce volume renderings, and draw volumes of interest. It also incorporates Node Package Manager (NPM) for statistics and dcm2nii for converting DICOM images to Nifti format. We employed MRIcron to segment the data [17,18]. The complete brain scan is the data set obtained from ADNI. MRI data were processed by this program to obtain the ROI, which yielded only the hippocampus region with its original grey intensities preserved. We overlaid the Winterburn segmentation mask generated by volBrain on the original MRI and maintained the area beneath the mask, removing the pixels beyond the mask boundary as shown in Figure 4.

Hippocampus Segmentation
MRIcron is a cross-platform image viewer for the Nifti format. It can import several layers of images, produce volume renderings, and draw volumes of interest. It also incorporates Node Package Manager (NPM) for statistics and dcm2nii for converting DICOM images to Nifti format. We employed MRIcron to segment the data [17,18]. The complete brain scan is the data set obtained from ADNI. MRI data were processed by this program to obtain the ROI, which yielded only the hippocampus region with its original grey intensities preserved. We overlaid the Winterburn segmentation mask generated by volBrain on the original MRI and maintained the area beneath the mask, removing the pixels beyond the mask boundary as shown in Figure 4.

Texture Analysis
The texture is a crucial feature for recognizing ROI in an image. Haralick et al. [5] introduced GLCM as one of the first methods for texture feature extraction in 1973. For texture analysis, GLCM was used. The two pixels are considered at a time namely the reference and the neighbour pixel. The grey level distribution of pairs of pixels in a specific distance and image configuration was analyzed by GLCM, reflecting the texture or spatial

Texture Analysis
The texture is a crucial feature for recognizing ROI in an image. Haralick et al. [5] introduced GLCM as one of the first methods for texture feature extraction in 1973. For texture analysis, GLCM was used. The two pixels are considered at a time namely the reference and the neighbour pixel. The grey level distribution of pairs of pixels in a specific Appl. Syst. Innov. 2021, 4, 49 6 of 17 distance and image configuration was analyzed by GLCM, reflecting the texture or spatial arrangements of pixel intensities existing in ROI. The four directions of two-dimensional (2D) GLCM are 0 • , 45 • , 90 • , and 135 • , but for 3D GLCM, there are thirteen independent directions as described in Table 2. Table 2. Directions for 3D GLCM.

Sr. No. Directions
At short distances d, GLCM provides fine texture, whereas it provides coarse texture in longer distances. The roughness of texture is associated with spatial frequency. Coarse texture has a low spatial frequency, while fine texture has a high spatial frequency. Before computing the GLCM, we defined a certain spatial relationship between the reference and neighbouring pixels, as shown in Figure 5. arrangements of pixel intensities existing in ROI. The four directions of two-dimensional (2D) GLCM are 0°, 45°, 90°, and 135°, but for 3D GLCM, there are thirteen independent directions as described in Table 2. Table 2. Directions for 3D GLCM.

Sr. No. Directions
At short distances d, GLCM provides fine texture, whereas it provides coarse texture in longer distances. The roughness of texture is associated with spatial frequency. Coarse texture has a low spatial frequency, while fine texture has a high spatial frequency. Before computing the GLCM, we defined a certain spatial relationship between the reference and neighbouring pixels, as shown in Figure 5. The Figure 6 shows distances mentioned in terms of the X, Y and Z-axis [19,20]. The Figure 6 shows distances mentioned in terms of the X, Y and Z-axis [19,20]. A GLCM is generated by size (Range of Intensities x Range of Intensities). All intensities are initialized to 0 once a spatial relationship is defined. By transversely going through the image. Identifying the defined spatial relationship for every pair of intensities, and incrementing that cell in the matrix. The texture attributes can be calculated from the matrix to represent the textures in the image once the GLCM has been calculated. Haralick retrieved the following features from the GLCMs to characterize texture. A few features are shown in Table 3   A GLCM is generated by size (Range of Intensities x Range of Intensities). All intensities are initialized to 0 once a spatial relationship is defined. By transversely going through the image. Identifying the defined spatial relationship for every pair of intensities, and incrementing that cell in the matrix. The texture attributes can be calculated from the matrix to represent the textures in the image once the GLCM has been calculated. Haralick retrieved the following features from the GLCMs to characterize texture. A few features are shown in Table 3 [5,21,22].
µ-the GLCM mean (being an estimate of the intensity of all pixels in the relationships that contributed to the GLCM) σ-the variance of the intensities of all reference pixels in the relationships that contributed to the GLCM (symmetric) 4 Cluster Prominence Maximum Probability Inverse difference moment normalized

Support Vector Machine (SVM)
For clean and preprocessed data that is linearly separable, SVM creates a hyperplane between the support vectors called the decision boundary, as shown in Figure 7. Data points on either side of the decision boundary belong to different classes. The hyperplane is one dimension lower than the dimension of the dataset; e.g., if the dataset is three-dimensional, the hyperplane would be two-dimensional. SVM tries to maximize the margin from the closest points of each class using supervised learning [23,24].

Decision Trees
A decision tree is a supervised learning algorithm. It is used when there is a nonlinear or complex nonlinear relationship between independent and dependent variables. First, it divides the dataset into different nodes/leafs, as shown in Figure 8. Then, based on the number of target variables, these nodes are split further into different subnodes; a node that cannot be split into more subnodes is called the terminal node. Based on the dataset, different algorithms can be used for doing the split in the available variables [25].

Decision Trees
A decision tree is a supervised learning algorithm. It is used when there is a nonlinear or complex nonlinear relationship between independent and dependent variables. First, it divides the dataset into different nodes/leafs, as shown in Figure 8. Then, based on the number of target variables, these nodes are split further into different subnodes; a node that cannot be split into more subnodes is called the terminal node. Based on the dataset, different algorithms can be used for doing the split in the available variables [25].

Decision Trees
A decision tree is a supervised learning algorithm. It is used when there is a nonlinear or complex nonlinear relationship between independent and dependent variables. First, it divides the dataset into different nodes/leafs, as shown in Figure 8. Then, based on the number of target variables, these nodes are split further into different subnodes; a node that cannot be split into more subnodes is called the terminal node. Based on the dataset, different algorithms can be used for doing the split in the available variables [25].

Random Forest
One of the most common problems with decision trees is overfitting. If no limit is given to the algorithm, it might give 100% training accuracy by creating a different node for each observation. Random forest, also called Ensemble Bagged trees, reduces overfitting in decision trees by combining multiple trees and selecting the best performance. First, the dataset is divided into N samples, and then a common learning rule is used for each sample; the prediction generated from this step is then combined using voting or by taking the average as, shown in Figure 9 [26,27].

Random Forest
One of the most common problems with decision trees is overfitting. If no limit is given to the algorithm, it might give 100% training accuracy by creating a different node for each observation. Random forest, also called Ensemble Bagged trees, reduces overfitting in decision trees by combining multiple trees and selecting the best performance. First, the dataset is divided into N samples, and then a common learning rule is used for each sample; the prediction generated from this step is then combined using voting or by taking the average as, shown in Figure 9 [26,27].

Results
After obtaining the segmented hippocampus, 20 GLCM features were extracted with d = 1 and 2 for all the 13 possible directions, of which 8 (Autocorrelation, Energy, Entropy, Homogeneity, Maximum Probability, Correlation, and Information measure of correla-

Results
After obtaining the segmented hippocampus, 20 GLCM features were extracted with d = 1 and 2 for all the 13 possible directions, of which 8 (Autocorrelation, Energy, Entropy, Homogeneity, Maximum Probability, Correlation, and Information measure of correlations 1 and 2) were deemed to contribute the most to the classification by obtaining their feature score, and later analysis methods were shortlisted based on their GLCM feature score. After running GLCM on the dataset, 20 features were generated; however, the accuracy of the machine learning technique was relatively low. The lower accuracy was due to the dataset's more considerable variance; it had a few factors contributing to randomness; therefore, those specific attributes were removed. An F-type statistic test was employed, which analyses the variance of two dataset means to eliminate the implausible values. It provided the standard deviation value, which explained the variance between datasets, resulting in reduced accuracy. This finding of the F value of statistical attributes aided in eliminating features that were not viable. Age and structural data such as hippocampus volume were included in the dataset for comparison analysis, in addition to these 8 GLCM features.
The research implementation platforms used python libraries and MATLAB. The classification model was trained using three different machine learning algorithms. SVM, Decision Trees, and Ensemble Bagged Trees were the models used (another name for Random Forest classifier). The dataset was taken into consideration when selecting these models. The classification was performed without setting advanced optimization techniques. Since the training dataset was limited (234), a classifier that can perform better with fewer data had to be chosen. The dataset with a clear separation between its classes performed well with SVM. Figure 10 gives the notion that for any machine learning algorithm, if the training and testing data was linearly separable, then the algorithms performed better than other nonlinear dataset classifiers. When instances of both the classes could be easily separated either by a straight line or nonlinearly, SVM became a suitable option for such data points. The scatter plot of several input features indicated that SVM would better fit the data set than other classifiers. SVM used a hard margin because the data was linearly separable. The regularization parameter (lambda) served as a degree of importance that was given to missclassifications. SVM posed a quadratic optimization problem that looks for maximizing the margin between both classes and minimizing the amount of missclassifications. The reason for choosing SVM, Decision Trees, and random forests is because of their implementation simplicity and relatively better model accuracy. Logistic regression, artificial neural networks, and nonlinear SVM work better when the data is nonlinear [28][29][30].
The number of trees in the Random Forest depends on the number of rows in the data set. In the case of a random forest, hyper parameters include the number of decision trees in the forest and the number of features considered by each tree when splitting a node. (The parameters of a random forest are the variables and thresholds used to split each node learned during training) [31].
Decision trees, which are also supervised learning algorithms, are another algorithm utilized in this section. The key reason for utilizing this technique is that it eliminates the requirement for the normalization of datasets. The dataset used has different scales for some attributes, therefore using this technique eliminates normalization.
Random Forest, also known as Ensemble Bagged Tree, is the third and final algorithm. It is a Decision Trees extension. It selects the best model from a set of uncorrelated decision tree classifiers. It takes longer to train than the other two algorithms, but it produces higher accuracy than the Decision Trees algorithm.
As per the data points generated from our dataset, Decision Trees, SVM, and Random Forest classifiers were chosen as they are simpler and more accurate as compared to other machine learning classifiers [32]. The dataset has linear boundaries and hence was better suited for the selected classifiers.
Random Forest, also known as Ensemble Bagged Tree, is the third and final algorithm. It is a Decision Trees extension. It selects the best model from a set of uncorrelated decision tree classifiers. It takes longer to train than the other two algorithms, but it produces higher accuracy than the Decision Trees algorithm.
As per the data points generated from our dataset, Decision Trees, SVM, and Random Forest classifiers were chosen as they are simpler and more accurate as compared to other machine learning classifiers [32]. The dataset has linear boundaries and hence was better suited for the selected classifiers. The confusion Matrix for the SVM, Decision Trees and Ensemble are illustrated in Figure 11.  The accuracy of the Decision Trees, SVM, and Ensemble models are shown in Table  4. Only the volume and age were used as input features in these models. SVM outperformed the other two models in terms of accuracy. After observing the scatter plot of age The accuracy of the Decision Trees, SVM, and Ensemble models are shown in Table 4. Only the volume and age were used as input features in these models. SVM outperformed the other two models in terms of accuracy. After observing the scatter plot of age and volume in Figure 10a, volume was the factor due to which SVM had superior accuracy. Figure 12 shows the Receiver Operating Characteristics (ROC) curve for SVM, Decision Trees, and Ensemble techniques. and volume in Figure 10a, volume was the factor due to which SVM had superior accuracy. Figure 12 shows the Receiver Operating Characteristics (ROC) curve for SVM, Decision Trees, and Ensemble techniques.  From Figure 10b, the autocorrelation for AD patients was lower than compared to CN subjects. The uniformity in volumes of AD patients was higher than that of CN subjects, and the difference between neighbouring pixels was lower than that of CN subjects, From Figure 10b, the autocorrelation for AD patients was lower than compared to CN subjects. The uniformity in volumes of AD patients was higher than that of CN subjects, and the difference between neighbouring pixels was lower than that of CN subjects, as can be observed from Figure 10c. The randomness of intensity distribution was high for CN and was low for AD subjects. AD subjects had slightly more linear structures compared to CN. Figure 13 shows the comparative analysis of the volume data between AD and CN groups and it is seen that there is a noticeable distinction among the data samples. From Figure 10b, the autocorrelation for AD patients was lower than compared to CN subjects. The uniformity in volumes of AD patients was higher than that of CN subjects, and the difference between neighbouring pixels was lower than that of CN subjects, as can be observed from Figure 10c. The randomness of intensity distribution was high for CN and was low for AD subjects. AD subjects had slightly more linear structures compared to CN. Figure 13 shows the comparative analysis of the volume data between AD and CN groups and it is seen that there is a noticeable distinction among the data samples. In the next phase, classification models were trained only on the GLCM characteristics that had been identified. For textural features, Ensemble provided the highest accuracy (85%) at d = 1 and direction (d, d, 0), while Decision Trees was least accurate at d = 2 In the next phase, classification models were trained only on the GLCM characteristics that had been identified. For textural features, Ensemble provided the highest accuracy (85%) at d = 1 and direction (d, d, 0), while Decision Trees was least accurate at d = 2 in direction (d, d, d). One evident finding from the accuracy table is that as the GLCM distance parameter was increased, the model's accuracy decreased since average accuracy dropped from 82.60% to 82.23% for textural features. The right and left hippocampus were exceptionally small in the MRI, and as the distance parameter was increased, the quality of the GLCM matrix decreased due to its small dimensions; hence the distance was limited to d = 2. While the model was trained with eight GLCM features, i.e., Textural data, the results produced were unsatisfactory as it had less accuracy than the results produced with Textural and Structural Data. Table 5 represents the classification accuracy with Textural and structural data as input.
The texture features were merged with volume and age data. While the model was trained just on GLCM features, accuracy ranged from 79.1% to 85%. When only volume data was provided, the maximum accuracy was 87.2%. When the GLCM features dataset was paired with volume data, including the patient's age and hippocampal volume, the accuracy improved up to 90.2%. The Ensemble Bagged Trees algorithm performed the best with the highest accuracy; one factor for better performance was the additional variables provided for the dataset. By including the structural data, a 4.38% improvement was observed on average.
The combination of features produce improvement by 4.38% on the structural data. Hence, it was observed as a better fit for the dataset. Figure 14 shows average accuracy for top performing machine learning algorithms applied on the data set. Table 6 summarizes the classification accuracy based on texture analysis and machine learning techniques.

Conclusions
AD is one of the world's leading causes of death from neurological disorders. The number of people with Alzheimer's is steadily rising over the world. The neurologist's manual diagnosis approach for AD is time-consuming and may not always yield correct results. One of the most significant elements in using brain images to classify AD is feature extraction. It is revealed that all 20 features retrieved from the GLCM of the participants' MRI are not required. It is observed that the crucial eight features contribute the most to the classification of the individuals using the F-type statistics test. According to the results, the classification accuracy declines as the distance between neighbouring pixels increases. The comparisons demonstrate that using structural data, textural data, and age as input combined improves classification accuracy. The top-performing algorithms were Ensemble and Decision Trees, with a classification accuracy of 90.2% and 88.5%, respectively, and SVM, with a classification accuracy of 87.2%. The subject classification was on par with research articles that have used feature extraction using various image processing techniques [39].  Institutional Review Board Statement: Ethical review and approval were waived for this study, due to freely publicly available data used for experimentations.
Informed Consent Statement: Patient consent was waived due to freely publicly available data used for experimentations.

Data Availability Statement:
The data presented in this study are available on request from the corresponding authors.