Next Article in Journal
A Perspective Review on Green Nanotechnology in Agro-Ecosystems: Opportunities for Sustainable Agricultural Practices & Environmental Remediation
Next Article in Special Issue
EffiMob-Net: A Deep Learning-Based Hybrid Model for Detection and Identification of Tomato Diseases Using Leaf Images
Previous Article in Journal
Bioactive Compounds of Barbatimão (Stryphnodendron sp.) as Dietary Additive in Lamb Diets
Previous Article in Special Issue
Pest Localization Using YOLOv5 and Classification Based on Quantum Convolutional Network
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Hybrid Framework for Detection and Analysis of Leaf Blight Using Guava Leaves Imaging

1
Department of Computer Science, COMSATS University Islamabad, Wah Campus, Wah Cantt 47040, Pakistan
2
Department of Electrical/Electronics & Computer Engineering, Faculty of Engineering, University of Uyo, Uyo 520103, Nigeria
3
Industrial Engineering Department, College of Engineering, King Saud University, P.O. Box 800, Riyadh 11421, Saudi Arabia
4
Centre for Smart Systems, AI and Cybersecurity, Staffordshire University, Stoke-on-Trent ST4 2DE, UK
*
Authors to whom correspondence should be addressed.
Agriculture 2023, 13(3), 667; https://doi.org/10.3390/agriculture13030667
Submission received: 31 January 2023 / Revised: 23 February 2023 / Accepted: 6 March 2023 / Published: 13 March 2023
(This article belongs to the Special Issue Big Data Analytics and Machine Learning for Smart Agriculture)

Abstract

:
Fruit is an essential element of human life and a significant gain for the agriculture sector. Guava is a common fruit found in different countries. It is considered the fourth primary fruit in Pakistan. Several bacterial and fungal diseases found in guava fruit decrease production daily. Leaf Blight is a common disease found in guava fruit that affects the growth and production of fruit. Automatic detection of leaf blight disease in guava fruit can help avoid decreases in its production. In this research, we proposed a CNN-based deep model named SidNet. The proposed model contains thirty-three layers. We used a guava dataset for early recognition of leaf blight, which consists of two classes. Initially, the YCbCr color space was employed as a preprocessing step in detecting leaf blight. As the original dataset was small, data augmentation was performed. DarkNet-53, AlexNet, and the proposed SidNet were used for feature acquisition. The features were fused to get the best-desired results. Binary Gray Wolf Optimization (BGWO) was used on the fused features for feature selection. The optimized features were given to the variants of SVM and KNN classifiers for classification. The experiments were performed on 5- and 10-fold cross validation. The highest achievable outcomes were 98.9% with 5-fold and 99.2% with 10-fold cross validation, confirming the evidence that the identification of Leaf Blight is accurate, successful, and efficient.

1. Introduction

Food is the fundamental requirement for the existence of human beings, and it is the notable outcome of agricultural activities. Agriculture is assumed to be the backbone of economic development, as it exhibits the cultivation of multiple crops, fruits, and vegetables. There is a large difference between the cultivation and annual production of fruits because of inappropriate advancements in technology, lack of knowledge, and diseases that negatively affect the production [1]. Disease detection in plants is a challenging task and is essential to diagnose at early stages. Diseases are mostly diagnosed through leaves because they tend to highlight contaminated parts immediately. Guava is an important fruit in agriculture; therefore, its leaves are selected for the detection and recognition of diseases [2]. Guava is nutritionally beneficial, serving calcium and iron to the human body. It is cultivated in America, especially in Mexico, Thailand, South Africa, and many other countries. Many laboratories such as the Central Institute of Subtropical Horticulture (CISH) and different institutes are continuing to work on guava production in different areas of the world [3]. Several diseases, such as bacterial and fungal diseases, attack the guava fruit, which badly affects its production [4]. There are different techniques of ML applied for disease detection. Almost 177 types of diseases are found that damage leaves, causing leaf blight and leaf spots. Known diseases include brown roots, twig drying, bacterial wilt, anthracnose, ring rots, and many others [5].
Many researchers aim for innovations in disease detection. Disease detection relies on five major steps. Usually, the first step in image processing is image acquisition. After obtaining images, preprocessing incorporates multiple steps that result in better accuracy. After preprocessing, feature extraction is performed, where the features of the images are boosted for further computation and selection. The final stage is classification. A variety of models are presented using diverse methodologies such as convolution neural network (CNN), gradient descent (GD), and many others for classification purposes [6]. Convolution neural networks play an essential part in the extraction of features through hidden layers, as manual extraction is costly and time-consuming [7]. Plant pathologists need an automatic detection system to diagnose leaf blight in plants.
The main focus of the proposed methodology is the detection and classification of leaf blight. Leaf blight affects plants as a result of a pathogenic organism infecting leaves. Therefore, an automated system is needed to detect leaf blight disease. Research of diseases in guava fruit is a challenging task, as it seeks a variety of data regarding diseases in the relevant field [8]. The forecasted production of guava is 498.95 thousand tonnes in 61.37 acres in the year 2020–2021 and the production of guava is 499.68 thousand tonnes in 61.37 acres in the year 2021–2022. Evaluation of the researched tasks becomes critical with time due to the wide range of diseases, and a great deal of effort has already been applied towards the relevant field [9].
There are several limits and difficulties in detecting and classifying guava plant diseases in the existing literature. Some major problems per the literature are the poor contrast, variation in shape, texture, and size, and illumination problems found in disease images that make them difficult to recognize and classify.
This article presents a new methodology for the detection of leaf blight. The purpose is to classify the healthy and diseased images of guava leaves with improved accuracy. The significant contributions presented in this research are as follows:
  • A new deep CNN Net named SidNet is presented, which consists of 33 layers along with 35 connections. The pretraining of SidNet is performed on a plant imaging dataset. The features are extracted from the proposed SidNet, darknet53, and AlexNet, which are further fused using serial fusion. The deep features are also known as automatic features; they automatically solve the issues related to contrast, shape, texture, and illumination.
  • The features are sorted using an Entropy Algorithm, and for better feature selection, Binary Gray Wolf Optimization is used. The selected features are used to make a single feature vector for classification using an SVM and KNN Classifier to achieve the best performance and results.
  • Data Augmentation is performed, as the selected dataset is small; therefore, the images are flipped both horizontally and vertically to make the dataset large.
The paper consists of five sections, where Section 1 explains the introduction, motivation, contribution, and problem statements for leaf blight detection. Section 2 covers the recent existing work. Section 3 provides the details of the presented proposed framework and Section 4 describes the details of the experiments and outcomes. Lastly, Section 5 covers the conclusion.

2. Related Work

Diseases in fruit plants and leaves are a major cause of destruction and economic loss. Automated systems help greatly with the detection of diseases at early stages. While considering the field of detection of disease in plants, deep neural networks work perfectly to identify and classify diseases. These networks are mentored to conduct high-value results in detecting and classifying diseases, and to fulfill the demands of food deformation prevention.
There are different methods for collecting images under certain conditions. Images are captured by multiple appliances, such as cameras, sensors, mobile phones, and other devices. In this era, more datasets containing guava are publicly available on multiple forums, such as Kaggle, Mendeley, and many others [10]. Pre-processing of images is an important phase in image processing. The pre-processing phase entails multiple steps which help highlight the focused parts and remove irrelevant information from guava leaf images. In the real world, label noise on images is a matter of concern. Multiple techniques have attained the best results in denoising images, especially mixed noise, speckle noise, and salt and pepper noise. Low contrast and color distortion in guava leaf images make them blur. Scattering and light absorption also affect clear image visualization [11]. Images are preprocessed by using rotational filters such as horizontal and vertical flipping. Data augmentation techniques are used, such as applying rotations and zooming into images [12]. Color spacing techniques are extensively applied in image processing. RGB, CIELAB, and CMYK models are mostly used as color spacing techniques to give the best results.
Feature extraction is a process in which reatures are reduced from the raw dataset and new features for manageable processing are created. Texture analysis has a wide range of applications [13]. Pattern recognition requires feature extraction to solve problems in prediction, cluster discrimination, and representation of data in the best way [14,15]. Content-Based Image Retrieval (CBIR) converts high-level image visuals into feature vectors that contain some properties [16]. There are multiple techniques to extract the features from guava leaf images, such as handcrafted-based features, region-based features, deep CNN-based features, texture-based features, color-based features, morphological-based features, etc. Extraction of features is categorized into hand-crafted-based features and deep-based features.
The selection of features from plant leaf images is carried out after the extraction of hand engineered and deep-based features. The set of features is chosen while noisy, poor, and extra features are eliminated from the original set of features [17]. There are five main types of feature selection, which are (1) Linear Method, (2) Non-Linear Method, (3) Filter-Based Method, (4) Wrapper Method, (5) Embedded Method. Linear methods include PCA and LDA. PCA stands for Principal Component Analysis, which is used for data reduction [18]. LDA stands for Linear Discriminant Analysis and is used for the conversion of high dimension features into lower dimension features [19]. Non-linear methods include Entropy [20], Genetic Algorithm (GA) [21], Binary Gray Wolf [22], Slap Swarm [23], Atom Search [24] and many others. Filter methods includes mRMR [25], Missing Value Ratio [26], and many others. Wrapper methods include Jackstraw [27] and Boruta [28]. Finally, embedded methods include LASSO [29], Ridge [30], Elastic [31], and many others. Image fusion [32] helps greatly in improving classifier accuracy with less computational cost [33]. Different algorithms are proposed that use image fusion to get the best accuracy results.
Image classification is the last step in image processing [34]. Classification tends to dominate the feature vector to determine which object belongs to which class [35]. There are different types of techniques used for the classification of healthy and diseased images of plant leaves. Image classification is divided into three main categories, which are (1) Supervised Learning, (2) Unsupervised learning, and (3) Object-based image analysis. Supervised Learning is used to detect the new category of the object from training data [36]. Unsupervised Learning is a process in which an image is identified in an image collection without using labeled training data [37]. Object-based analysis involves the grouping of pixels on the basis of some similarities such as shape and neighborhood [38]. To get the most accurate results, the Plant Village dataset is used for testing and training purposes. A total of 80% of guava leaf images are used for testing while 20% of them are used for training purposes. The achievable accuracy is 97.22% using Alex-Net and Squeeze-Net after segmentation and classification [39]. Atila et al. [40] designed the Efficient-Net architecture, which is designed for classification purposes. Different architectures are applied using CNN for model training to get highly accurate results. The model training is performed on the dataset of 87,848 images. Images are preprocessed using different techniques such as downscaling and squaring methods, they are then classified, and an accuracy of 99.53% is achieved using AlexNet, VGG-16, and GoogLeNet [41]. Several algorithms are used for image classification. These are SVM [42], K-Nearest Neighbor (KNN) [43], Naïve Bayes [44], Shadow algorithms [45], Minimum Mean Distance (MMD) [46], Decision Trees [47], K-Means Clustering [48] and many others. The datasets are frequently classified by SVM. This involves supervised learning and comprises points that are in the sample space and different regions [49]. Segmentation is performed on the preprocessed data in three stages. In the first stage, the deep CNN is trained to learn the mapping from the space map. In the second stage, prediction-based labels are acquired. At the last stage, these acquired labeled images are sent to SVM for classification and achieve an accuracy of 86% [50]. In machine learning, KNN is a statistical classification algorithm. It gathers the objects selected by neighbors having the highest number of votes [51]. KNN is inspected for the detection of weeds from UAV images of the chili crop of Australia. In comparison with KNN, SVM and Random Forest (RF) are used. The achievable accuracies across RF, SVM, and KNN are 96%, 94%, and 63%, respectively [52]. KNN is also used for classifying facial expressions [53]. Additionally, KNN is used for the classification of grape leaves into healthy and unhealthy leaves. Texture-based and color-based features are extracted from grape leaf images and are classified by the KNN classifier, and an accuracy of 96.66% is achieved [54]. Table 1 depicts an overview of recent works related to plants diseases analysis

3. Materials and Methods

The proposed methodology consists of multiple phases of image processing. In the first step, as a preprocessing step, color spacing is performed on the images of the dataset. Images are converted from RGB color format to YCbCr color format. After getting preprocessed data, feature extraction is performed using two pretrained models and a newly proposed CNN deep model known as SidNet. This newly proposed model is based on 33 convolutional layers. This proposed CNN deep model is pretrained using the Plant-Village dataset, which consists of 38 classes. After pretraining, features are collected from the proposed CNN deep model. These extracted features are then fused with AlexNet and DarkNet-53 to find the best and most appropriate results. A deep CNN known as AlexNet was developed in 2012, which is an 8-layer deep model and consists of 5 convolution layers and 3 Fully Connected layers. After every convolution layer and fully connected layer, ReLU is applied. It contains a dropout layer, which is applied after the first and second fully connected layer. The input size in AlexNet is 227 × 227. In all layers, the activation function is ReLU. Softmax is designed as an activation function in the output layer. AlexNet is the simplest deep CNN model and it is used to get highly accurate results. DarkNet-53 is another used deep model, which works as the backbone for YoloV3 in object detection. It contains 25 layers for batch normalization and Leaky ReLU. The input size is 256 × 256 in DarkNet-53. In the second step, the feature sorting entropy algorithm is used for the selection of the best features. These selected core features are fused using serial-based fusion. After fusing the core features, the binary gray wolf optimization algorithm is used. Finally, these extracted fused features are provided to SVM and KNN classifiers for the classification of Leaf Blight for the best achievable results. The complete view of the proposed model is shown in Figure 1. It covers a complete view of the flow of the diagram of the designed structure.

3.1. Image Preprocessing

In image processing, color space conversion is an important task. RGB is used to store real-time images and videos because it allows for the sensitivity of color detection cells for the human visual system. YCbCr is beneficial for its low-resolution capability for the human visual system. Therefore, the conversion of RGB to YCbCr is mostly used in image processing. The general formula is given below:
YCbCr =   RGB   2 YCbCr   ( RGB )
This is the general formula that is used for the conversion of RGB to YCbCr and it represents 8 bits per sample pixel in RGB format. The white and black colors are represented on a scale from 0 to 255. Therefore, the components of YCbCr are obtained from the following equations.
Y = 16 + 65.738 / 256   R + 129.057 / 256   G + 25.064 / 256   B
C b = 128 + 37.945 / 256   R 74.494 / 256   G + 112.439 / 256   B
C r = 128 + 112.439 / 256   R 94.154 / 256   G + 18.285 / 256   B
where Y is used to represent the luma (luminance) component. C b represents chrominance blue and C r represents chrominance red. These numbers are the constant values that are used to adjust the value of Y. Figure 2 shows the conversion from RGB to Y C b C r .

3.2. Data Augmentation

For data augmentation, horizontal and vertical flipping are used. According to the mathematical model, the horizontal flipping of the images is presented as follows:
H F   ( x ,   y ) = H O   ( x ,   y )  
And the vertical flipping of the images is presented as follows:
H V   ( x , y ) = H O ( x ,   y )  
HF shows the flipping function and HO shows the original image function that is to be flipped. Hv shows the flipping function and HO shows the original image function that is to be flipped.

3.3. Feature Extraction

After preprocessing and augmentation, the next phase is feature extraction. In this phase, with the help of pretrained models, the most optimal features are extracted. According to our proposed methodology, a newly designed model named SidNet, and other pretrained models such as DarkNet-53 and AlexNet, are used for extracting the most optimal features.

3.4. Proposed SidNet as CNN Net

This proposed CNN Net is a blockbuster CNN-based architecture used for the detection and classification of Leaf Blight. The proposed SidNet model consists of 33 layers involving 8 convolutional layers, 10 ReLU layers, 4 layers of batch normalization, one dropout layer, one softmax layer, one classification output, one fully connected, and one global average pooling layer. The Input size of the Proposed CNN Net is “227 × 227 × 3” and it contains 35 connections. The stride is “1 × 1” throughout the proposed framework’s convolution layer. The number of filters is set to 96 in all convolutional layers of the framework, while the padding dimensions vary according to the convolution layer used in the different stages of architecture. The padding of the last convolution layer is 5, 5, 5, 5. Two pooling layers are used, where the stride for the two pooling layers in the architecture is “2 × 2”. While the pooling size of the first pooling layer is 3, 2 and the pooling size of the second pooling layer is 3, 3. The mean decay and variance decay for all batch normalization layers are 0.1. Figure 3 shows the architecture of the proposed SidNet.
The proposed deep model named SidNet comprises 33 convolution layers. Its Input size is “227 × 227 × 3” and it contains 35 connections.
Due to the small number of samples of the available dataset, the proposed deep model with softmax (SM) classifier is first trained on the third-party dataset named CIFAR 100 [60]. Then, the guava leaf dataset is fed to SidNet for feature extraction. The features are extracted from the fully connected (FC) layer. These features, after feature selection, are trained and tested on various classifiers (such as SVM with its variants and KNN with its variants) for evaluation. According to our model, the features in SidNet are presented as:
f e × f = {   L 1 × 1   ,   L 1 × 2   ,   L 1 × 3   ,   .. L 1 × p   }
where L (1p) represents the number of features obtained from the proposed CNN model, which is known as SidNet, and e × f is the dimension of the resultant function.
Visualization of the strongest feature maps at different convolution layers with the proposed SidNet architecture is shown in Figure 4. The visualization is performed on conv_1, conv_2, conv_3, conv_4, conv_5, conv_6, conv_7, and conv2 of the proposed architecture.

3.5. DarkNet-53

DarkNet-53 acts as the backbone of YoloV3 for object detection. DarkNet comprises 53 layers and consists of multiple convolution layers. A total of 1024 features in DarkNet are extracted with the help of the global average pooling layer. There are 25 layers in Batch normalization. The input size is 256 × 256. According to our mathematical model, the features in DarkNet are presented as:
f a × b = {   j 1 × 1   ,   j 1 × 2   ,   j 1 × 3   , .. j 1 × m   }
where J (1…m) shows the number of features extracted from DarkNet-53 and a × b is the dimension of the resultant function. Visualization of the strongest feature maps at different convolution layers with the DarkNet-53 architecture is shown in Figure 5

3.6. AlexNet

AlexNet is the simplest model that comprises 8 layers. There are 5 convolution layers and 3 fully connected layers. Its input size is 227 × 227. The Activation function used in all layers is ReLU, which is applied after every convolution layer and fully connected layer. The Drop layer is applied after the first and second fully connected layer. The Activation function in the output layer is Softmax. The AlexNet contains 4096 features with the help of a convolution layer named a fully connected layer. According to our mathematical model, the features in AlexNet are presented as:
f c × d   = {   k 1 × 1   ,   k 1 × 2   ,   k 1 × 3 ,   . k 1 × n }
where k(1 … n) shows the number of features extracted from AlexNet and c × d shows the dimensions of the resultant function. Visualization of the strongest feature maps at different convolution layers with the AlexNet architecture is shown in Figure 6

3.7. Feature Selection

After extracting features from the proposed model and other pretrained models such as DarkNet-53, AlexNet, and SidNet, the selection of features is done with the help of feature sorting using entropy. The mathematical model for the selection of features using entropy is represented as:
E a × b   ( 1 ,   2 ,   3 ,   ..   n ) =   f a × b   { j 1 × 1   ,   j 1 × 2 ,   j 1 × 3 , . j 1 × m   } l o g ( j   1 × 1   ,   j 1 × 2   ,   j 1 × 3 , j 1 × m )
E c × d   ( β 1 ,   β 2 ,   β 3 ,   ..   β n   f   c × d = { k 1 × 1 , k 1 × 2 , k 1 × 3 , . . k 1 × n } l o g l o g   ( k 1 × 1   ,   k 1 × 2   ,   k 1 × 3 ,   k 1 × n )  
E e × f   ( γ 1 ,   γ 2 ,   γ 3 ,     γ n =   f e × f = { L 1 × 1 , L 1 × 2 , L 1 × 3 , L 1 × p } l o g l o g   ( L 1 × 1   ,   L 1 × 2   ,   L 1 × 3 , .. L 1 × p )  
where a × b, c × d, e × f represents the dimension of features obtained after sorting features. Log(j(1 − n)), Log(k(1 − n)), and Log(L(1 − n)) show the prediction of probability and j(1 − n), k(1 − n) and L(1 − n) show the selected features obtained from the extracted features. (1 − n), β(1 − n), γ ( 1 n ) show the features which are sorted.

3.8. Feature Fusion

Feature Fusion is performed to select the most optimal features. According to the mathematical model, the fusion of features is represented as:
f = x = 1 L (   E a × b )   y = 1 m ( E c × d )   z = 1 n ( E e × f )
After fusing the features using the entropy algorithm, Binary Gray Wolf Optimization is performed to obtain the most optimal results. The mathematical model representing the features selected from Binary Gray Wolf Optimization are as follows:
d = 1   i f   γ o < 1 3   2   i f   1 3 γ 6 < 2 3   3   i f   ,   o t h e r w i s e  
where d is the function of BGWO and γ o and γ 6 are the adjusting parameters that are used to set the value of the most optimal features.

3.9. Classification

Different classifiers are available for classification purposes, but SVM and KNN classifiers are chosen. These two classifiers are selected to achieve high accuracy and the most optimal results. In machine learning, features are reduced by carrying out the feature vector dimension. Different classification algorithms are available, such as Minimum Mean Distance (MMD), K means clustering, Decision Trees, Shadow algorithm, and Naïve Bayes. In this work, SVM and KNN are selected to perform the classification on the guava leaf dataset. These classifiers generate better results compared to other classifiers.

4. Results and Discussion

The purpose of this study is to classify the Leaf Blight disease with the best possible results. After the processing of the dataset using YCbCr, the extraction of features is performed using two pretrained models along with one proposed net. The selection of features is performed using BGWO. For classification purposes, SVM and KNN are chosen for the evaluation of execution. This section provides details about experiments that are performed on multiple sets of features and the results are recorded accordingly. These experiments and results are shown in two sets of test cases. Using 5 folds and 10 folds, validation experiments are performed. In comparison with other classifiers, SVM and KNN are selected, as they give the best results. The set of experiments are performed on Windows 10 (64-bit) and a Core (TM) i7-8700 CPU, 3.20 GHz (12 CPUs) 3.2 GHz processor with 16 GB RAM and an LCD and keyboard from HP. Training and testing of the designed network are performed on MATLAB R2020b.

4.1. Dataset

The assembled dataset for the classification of Leaf Blight used in this analysis [61] is small. These experiments are performed on guava leaves. The chosen dataset contains 415 guava leaf images in the original. This dataset contains a small number of images; therefore, they are augmented using horizontal and vertical flipping techniques, which increases the total number of images. The total number of images is 1000 after the augmentation technique. This dataset is publicly available on the Mendeley website. It is a binary class dataset (see Figure 7 for sample images) and many researchers have used it in their studies.

4.2. Performance Evaluation Methods

The performance measures can be used to measure and detect the performance of leaf blight disease in plants. The consequences are defined as follows: a True Positive rate as TP, True Negative rate as TN, False Positive rate as FP, and False Negative rate as FN. The Table 2 illustrates the performance measures that are used in this work.
Table 3 shows a summary of the best-achieved results performed on the guava leaf dataset, which proves that the proposed methodology is efficient and robust. Here, Quadratic SVM achieves the best results, i.e., 98.9% over 5 folds in 9.2 s and 99.2% over 10 folds in 16.2 s, with 3045 features on 5-fold cross validation.
The accompanying text contains some discussion on some of the experiments.
  • Experiment_1: Using 5-Fold and 10-Fold Validation on 3045 features
This section provides details about two test cases that were performed on 3045 features using both 5 folds and 10 folds. The selected classifiers are variants of SVM and KNN, which were chosen to get robust results. After augmentation, the chosen dataset contained 1002 images. The best results were recorded with measures such as accuracy, precision, recall, F1-score, and training time.
  • Experiment_1(a): Using 5 Folds and 3045 Features (1002 × 3045 features)
This test case shows the results of 3045 features upon 1002 images using SVM and KNN classifiers with 5-fold cross validation. Details are shown in Table 4.
The best results are achieved by the Quadratic SVM classifier in comparison with all KNN classifiers, which is 98.9% in 9.2 s. Here, the confusion matrix in Figure 8 and the ROC curve in Figure 9 are shown for the best results with the Quadratic SVM classifier. In the confusion matrix, A represents the diseased class, While B depicts healthy class
Here the ROC curve is shown in Figure 9 for both classes A and B, which are presented as ROC A and ROC B.
  • Experiment_1(b): Using 10 Folds and 3045 Features (1002 × 3045 features)
This test case shows the results of 3045 features upon 1002 images using SVM and KNN classifiers with 10 folds. Details are shown in Table 5.
The Quadratic SVM classifier achieved the best result in comparison with all KNN classifiers, which is 99.2%. Here, the confusion matrix as presented in Figure 10 and the ROC curve are shown for the best results against the Quadratic SVM classifier.
Here, the ROC curve is shown in Figure 11 for both classes A and B, which are presented as ROC A and ROC B.
  • Experiment_2: Using 5-Fold and 10-Fold Validation on 200 features
This section provides details about two test cases that were performed on 200 features using both 5 folds and 10 folds. The efficiently selected classifiers are SVM and KNN, which were chosen to get robust results. The chosen dataset contains 1002 images. The best results are recorded with other measures, such as accuracy, precision, recall, F1-score, training time, etc. Results are shown in the Table 6 in detail.
  • Experiment_2(a): Using 5 Folds and 200 Features (1002 × 200 features)
This test case shows the results of 200 features upon 1002 images using SVM and KNN classifiers with 5 folds. Details are shown in Table 6.
The Fine Gaussian SVM classifier achieved the best result in comparison with all KNN classifiers, which is 81.1% in 0.8 s. Here, the confusion matrix in Figure 12 and the ROC in Figure 13 curve are shown for the best results with the Fine Gaussian SVM classifier.
The ROC curve is shown in Figure 13 for both classes A and B which are presented as ROC A and ROC B.
  • Experiment_2(b): Using 10 Folds and 200 Features (1002 × 200 features)
This test case shows the results of 200 features upon 1002 images using SVM and KNN classifiers with 10 folds. Details are shown in Table 7.
Here, the Fine Gaussian SVM classifier achieved the best result in comparison with all KNN classifiers, which is 82.6% in 1.2 s. Here, the confusion matrix also shown in Figure 14 and the ROC curve are shown for the best results against the Fine Gaussian SVM classifier.
The ROC curve is shown in Figure 15 for both classes A and B, which are presented as ROC A and ROC B.
  • Experiment_3: Using 5-Fold and 10-Fold Validation on 500 features.
This section provides details about two test cases that are performed on 500 features using both 5 folds and 10 folds. The efficiently selected classifiers are SVM and KNN, which were chosen to get robust results. The chosen dataset contains 1002 images. The best results are recorded with some other measures like Accuracy, precision, recall, F1-score, training time, and others. Results are shown in the Table 8 in detail.
  • Experiment_3(a): Using 5 Folds and 500 Features (1002 × 500 features)
This test case shows the results of 500 features upon 1002 images using SVM and KNN classifiers with 5 folds. Details are shown in Table 8.
Here, the classifier Quadratic SVM achieved the best result in comparison with all KNN classifiers, which is 84.1% in 1.2 s. The confusion matrix in Figure 16 and the ROC curve in Figure 17 are shown for the best results against the Quadratic SVM classifier.
The ROC curve is shown for both classes A and B, which are presented as ROC A and ROC B in Figure 17.
  • Experiment_3(b): Using 10 Folds and 500 Features (1002 × 500 features)
This test case shows the results of 500 features upon 1002 images using SVM and KNN classifiers with 10 folds. Details are shown in Table 9.
Here, the classifier Fine Gaussian SVM achieved the best result in comparison with all KNN classifiers, which is 85.2% in 2.2 s. Here, the confusion matrix and the ROC curve are shown for the best results against the Fine Gaussian SVM classifier. The confusion matrix for FG (SVM) is shown in Figure 18.
Here, the ROC curve is presented (in Figure 19) for both classes A and B, which are presented as ROC A and ROC B.
Figure 20 shows the 5-fold best achievable results between training time and features. It shows the consumed time for training that is required for a particular set of features.
Here, features are presented on the x-axis and the training time is shown on the y-axis. These results are taken on 5-folds for the best achievable results.
Similarly, the graph in Figure 21 shows the relation between accuracy and features at 5-fold results. It shows that the features are presented along the x-axis and the accuracy is presented along the y-axis, and the line presents the best achievable results.
The graph in Figure 22 shows the relation between training time and features upon results taken on 10 folds. This graph shows that features are shown along the x-axis and the training time is shown along the y-axis, and the time consumed by a particular set of features for training is illustrated.
The graph in Figure 23 shows the relation between training time and accuracy upon results taken on 10 folds. This graph shows that features are shown along the x-axis and the accuracy is shown along the y-axis, and the time consumed by a particular set of features for training is illustrated.

5. Conclusions

Agriculture is the key to the development and rise of emergent nations. Diseases in plants cause crop damage. Detection of leaf blight is enormously important as it affects the annual production of guava fruit. Finally, the development of an automated system becomes indispensable. In this paper, leaf blight can be detected, analyzed, and classified through the proposed methodology. In this proposed methodology, our own deep CNN is designed, containing thirty-three layers. In the first phase of image processing, preprocessing is done by using color spacing YCbCr. The Guava dataset is chosen for the identification and analysis of leaf blight. Because the dataset is small, data augmentation is performed. Horizontal and vertical flipping were performed on images of guava leaves. After preprocessing, feature extraction was performed using Darknet-53 and AlexNet, as well as the proposed SidNet. For the selection of the best features, optimization algorithms such as Entropy and Binary Gray Wolf are used. Finally, classification is performed on guava leaf images and the best results with higher accuracy and less computational cost are achieved. Multiple experiments are performed while using the set of selected features (200, 500, 750, 1000 features using 5- and 10-fold validation). Based on the selected features, 98.9% of the results are achieved using an SVM classifier, as it proves that this proposed methodology is robust and efficient.
In the future, this work can be explored with quantum deep learning for improved performance. Quantum computing-based machine learning and deep convolutional neural networks can detect and classify leaf blight at its initial stage more precisely and meticulously, which will help save crops and fruits, save plants from destruction, and increase the production of guava fruit.

Author Contributions

Conceptualization, S.M., M.R., O.D.O., S.U.R., A.E.R. and H.T.R.; methodology, S.M. and M.R.; investigation, M.R., O.D.O., S.U.R., A.E.R. and H.T.R.; resources, A.E.R. and H.T.R.; writing—original draft preparation, S.M., M.R., O.D.O., S.U.R., A.E.R. and H.T.R.; writing—review and editing, S.M., M.R., O.D.O., S.U.R., A.E.R. and H.T.R.; supervision, M.R, A.E.R. and H.T.R.; funding acquisition, A.E.R. All authors have read and agreed to the published version of the manuscript.

Funding

The authors extend their appreciation to King Saud University for funding this work through Researchers Supporting Project number (RSPD2023R711), King Saud University, Riyadh, Saudi Arabia.

Institutional Review Board Statement

The study was conducted in accordance with the Declaration of Helsinki, and approved by the Institutional Review Board.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Rehman, A.; Jingdong, L.; Shahzad, B.; Chandio, A.A.; Hussain, I.; Nabi, G.; Iqbal, M.S. Economic perspectives of major field crops of Pakistan: An empirical study. Pac. Sci. Rev. B Humanit. Soc. Sci. 2015, 1, 145–158. [Google Scholar] [CrossRef] [Green Version]
  2. Rai, M.K.; Asthana, P.; Jaiswal, V.; Jaiswal, U. Biotechnological advances in guava (Psidium guajava L.): Recent developments and prospects for further research. Trees 2010, 24, 1–12. [Google Scholar] [CrossRef]
  3. Mitra, S.; Thingreingam Irenaeus, K. Guava cultivars of the world. In International Symposia on Tropical and Temperate Horticulture—ISTTH2016; CIRAD Publications: Cairns, Queensland, Australia, 2016; pp. 905–910. [Google Scholar]
  4. Almadhor, A.; Rauf, H.T.; Lali, M.I.U.; Damaševičius, R.; Alouffi, B.; Alharbi, A. AI-Driven Framework for Recognition of Guava Plant Diseases through Machine Learning from DSLR Camera Sensor Based High Resolution Imagery. Sensors 2021, 21, 3830. [Google Scholar] [CrossRef] [PubMed]
  5. Misra, A. Guava diseases—Their symptoms, causes and management. In Diseases of Fruits and Vegetables; Springer: Berlin/Heidelberg, Germany, 2004; Volume II, pp. 81–119. [Google Scholar]
  6. Dhiman, B.; Kumar, Y.; Hu, Y.-C. A general purpose multi-fruit system for assessing the quality of fruits with the application of recurrent neural network. Soft Comput. 2021, 25, 9255–9272. [Google Scholar] [CrossRef]
  7. Ashraf, R.; Habib, M.A.; Akram, M.; Latif, M.A.; Malik, M.S.A.; Awais, M.; Dar, S.H.; Mahmood, T.; Yasir, M.; Abbas, Z. Deep convolution neural network for big data medical image classification. IEEE Access 2020, 8, 105659–105670. [Google Scholar] [CrossRef]
  8. Naranjo-Torres, J.; Mora, M.; Hernández-García, R.; Barrientos, R.J.; Fredes, C.; Valenzuela, A. A review of convolutional neural network applied to fruit image processing. Appl. Sci. 2020, 10, 3443. [Google Scholar] [CrossRef]
  9. Lloret, E.; Plaza, L.; Aker, A. The challenging task of summary evaluation: An overview. Lang. Resour. Eval. 2018, 52, 101–148. [Google Scholar] [CrossRef] [Green Version]
  10. Bojer, C.S.; Meldgaard, J.P. Kaggle forecasting competitions: An overlooked learning opportunity. Int. J. Forecast. 2021, 37, 587–603. [Google Scholar] [CrossRef]
  11. Fu, X.; Cao, X. Underwater image enhancement with global–local networks and compressed-histogram equalization. Signal Process. Image Commun. 2020, 86, 115892. [Google Scholar] [CrossRef]
  12. Tang, Z.; Gao, Y.; Karlinsky, L.; Sattigeri, P.; Feris, R.; Metaxas, D. OnlineAugment: Online Data Augmentation with Less Domain Knowledge; Springer: Berlin/Heidelberg, Germany, 2020; pp. 313–329. [Google Scholar]
  13. Liu, Z.; Lai, Z.; Ou, W.; Zhang, K.; Zheng, R. Structured optimal graph based sparse feature extraction for semi-supervised learning. Signal Process. 2020, 170, 107456. [Google Scholar] [CrossRef]
  14. Ghojogh, B.; Samad, M.N.; Mashhadi, S.A.; Kapoor, T.; Ali, W.; Karray, F.; Crowley, M. ‘Feature selection and feature extraction in pattern analysis: A literature review. arXiv 2019, arXiv:1905.02845. [Google Scholar]
  15. Zhang, J.; Liu, B. A review on the recent developments of sequence-based protein feature extraction methods. Curr. Bioinform. 2019, 14, 190–199. [Google Scholar] [CrossRef]
  16. Latif, A.; Rasheed, A.; Sajid, U.; Ahmed, J.; Ali, N.; Ratyal, N.I.; Zafar, B.; Dar, S.H.; Sajid, M.; Khalil, T. Content-based image retrieval and feature extraction: A comprehensive review. Math. Probl. Eng. 2019, 2019, 9658350. [Google Scholar] [CrossRef]
  17. Raj, R.J.S.; Shobana, S.J.; Pustokhina, I.V.; Pustokhin, D.A.; Gupta, D.; Shankar, K. Optimal feature selection-based medical image classification using deep learning model in internet of medical things. IEEE Access 2020, 8, 58006–58017. [Google Scholar] [CrossRef]
  18. Ali, L.; Wajahat, I.; Golilarz, N.A.; Keshtkar, F.; Bukhari, S.A.C. ‘LDA–GA–SVM: Improved hepatocellular carcinoma prediction through dimensionality reduction and genetically optimized support vector machine. Neural Comput. Appl. 2021, 33, 2783–2792. [Google Scholar] [CrossRef]
  19. Ozyurt, B.; Akcayol, M.A. A new topic modeling based approach for aspect extraction in aspect based sentiment analysis: SS-LDA. Expert Syst. Appl. 2021, 168, 114231. [Google Scholar] [CrossRef]
  20. Zhao, J.; Liang, J.-M.; Dong, Z.-N.; Tang, D.-Y.; Liu, Z. Accelerating information entropy-based feature selection using rough set theory with classified nested equivalence classes. Pattern Recognit. 2020, 107, 107517. [Google Scholar] [CrossRef]
  21. Maleki, N.; Zeinali, Y.; Niaki, S.T.A. A k-NN method for lung cancer prognosis with the use of a genetic algorithm for feature selection. Expert Syst. Appl. 2021, 164, 113981. [Google Scholar] [CrossRef]
  22. Hu, P.; Pan, J.-S.; Chu, S.-C. Improved binary grey wolf optimizer and its application for feature selection. Knowl. Based Syst. 2020, 195, 105746. [Google Scholar] [CrossRef]
  23. Tubishat, M.; Idris, N.; Shuib, L.; Abushariah, M.A.; Mirjalili, S. Improved Salp Swarm Algorithm based on opposition based learning and novel local search algorithm for feature selection. Expert Syst. Appl. 2020, 145, 113122. [Google Scholar] [CrossRef]
  24. Too, J.; Rahim Abdullah, A. Binary atom search optimisation approaches for feature selection. Connect. Sci. 2020, 32, 406–430. [Google Scholar] [CrossRef]
  25. Özyurt, F. A fused CNN model for WBC detection with MRMR feature selection and extreme learning machine. Soft Comput. 2020, 24, 8163–8172. [Google Scholar] [CrossRef]
  26. De Silva, K.; Jönsson, D.; Demmer, R.T. A combined strategy of feature selection and machine learning to identify predictors of prediabetes. J. Am. Med. Inform. Assoc. 2020, 27, 396–406. [Google Scholar] [CrossRef] [PubMed]
  27. Germain, P.-L.; Sonrel, A.; Robinson, M.D. pipeComp, a general framework for the evaluation of computational pipelines, reveals performant single cell RNA-seq preprocessing tools. Genome Biol. 2020, 21, 1–28. [Google Scholar] [CrossRef]
  28. Kaur, P.; Singh, A.; Chana, I. Computational techniques and tools for omics data analysis: State-of-the-art, challenges, and future directions. Arch. Comput. Methods Eng. 2021, 28, 4595–4631. [Google Scholar] [CrossRef]
  29. Ghosh, P.; Azam, S.; Jonkman, M.; Karim, A.; Shamrat, F.J.M.; Ignatious, E.; Shultana, S.; Beeravolu, A.R.; De Boer, F. Efficient Prediction of Cardiovascular Disease Using Machine Learning Algorithms With Relief and LASSO Feature Selection Techniques. IEEE Access 2021, 9, 19304–19326. [Google Scholar] [CrossRef]
  30. Deepa, N.; Prabadevi, B.; Maddikunta, P.K.; Gadekallu, T.R.; Baker, T.; Khan, M.A.; Tariq, U. An AI-based intelligent system for healthcare analysis using Ridge-Adaline Stochastic Gradient Descent Classifier. J. Supercomput. 2021, 77, 1998–2017. [Google Scholar] [CrossRef]
  31. Amini, F.; Hu, G. A two-layer feature selection method using genetic algorithm and elastic net. Expert Syst. Appl. 2021, 166, 114072. [Google Scholar] [CrossRef]
  32. Zhang, H.; Xu, H.; Xiao, Y.; Guo, X.; Ma, J. Rethinking the Image Fusion: A Fast Unified Image Fusion Network Based on Proportional Maintenance of Gradient and Intensity. Proc. AAAI Conf. Artif. Intell. 2020, 34, 12797–12804. [Google Scholar] [CrossRef]
  33. Jung, H.; Kim, Y.; Jang, H.; Ha, N.; Sohn, K. Unsupervised deep image fusion with structure tensor representations. IEEE Trans. Image Process. 2020, 29, 3845–3858. [Google Scholar] [CrossRef]
  34. Shakya, S. Analysis of artificial intelligence based image classification techniques. J. Innov. Image Process. (JIIP) 2020, 2, 44–54. [Google Scholar] [CrossRef]
  35. Hong, D.; Wu, X.; Ghamisi, P.; Chanussot, J.; Yokoya, N.; Zhu, X.X. Invariant attribute profiles: A spatial-frequency joint feature extractor for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2020, 58, 3791–3808. [Google Scholar] [CrossRef] [Green Version]
  36. Ke, Z.; Qiu, D.; Li, K.; Yan, Q.; Lau, R.W. Guided Collaborative Training for Pixel-Wise Semi-Supervised Learning; Springer: Berlin/Heidelberg, Germany, 2020; pp. 429–445. [Google Scholar]
  37. Wang, F.; Liu, H.; Guo, D.; Sun, F. Unsupervised Representation Learning by InvariancePropagation. arXiv 2020, preprint. arXiv:2010.11694. [Google Scholar]
  38. Karantanellis, E.; Marinos, V.; Vassilakis, E.; Christaras, B. Object-based analysis using unmanned aerial vehicles (UAVs) for site-specific landslide assessment. Remote Sens. 2020, 12, 1711. [Google Scholar] [CrossRef]
  39. Durmuş, H.; Güneş, E.O.; Kırcı, M. Disease Detection on the Leaves of the Tomato Plants by Using Deep Learning; IEEE: Piscataway, NJ, USA, 2017; pp. 1–5. [Google Scholar]
  40. Atila, Ü.; Uçar, M.; Akyol, K.; Uçar, E. Plant leaf disease classification using EfficientNet deep learning model. Ecol. Inform. 2021, 61, 101182. [Google Scholar] [CrossRef]
  41. Ferentinos, K.P. Deep learning models for plant disease detection and diagnosis. Comput. Electron. Agric. 2018, 145, 311–318. [Google Scholar] [CrossRef]
  42. Azarmdel, H.; Jahanbakhshi, A.; Mohtasebi, S.S.; Muñoz, A.R. Evaluation of image processing technique as an expert system in mulberry fruit grading based on ripeness level using artificial neural networks (ANNs) and support vector machine (SVM). Postharvest Biol. Technol. 2020, 166, 111201. [Google Scholar] [CrossRef]
  43. Kasinathan, T.; Singaraju, D.; Uyyala, S.R. Insect classification and detection in field crops using modern machine learning techniques. Inf. Process. Agric. 2021, 8, 446–457. [Google Scholar] [CrossRef]
  44. Yao, J.; Ye, Y. The effect of image recognition traffic prediction method under deep learning and naive Bayes algorithm on freeway traffic safety. Image Vis. Comput. 2020, 103, 103971. [Google Scholar] [CrossRef]
  45. Atouf, I.; Al Okaishi, W.Y.; Zaaran, A.; Slimani, I.; Benrabh, M. A real-time system for vehicle detection with shadow removal and vehicle classification based on vehicle features at urban roads. Int. J. Power Electron. Drive Syst. 2020, 11, 2091. [Google Scholar] [CrossRef]
  46. Khan, M.A.; Khan, M.A.; Ahmed, F.; Mittal, M.; Goyal, L.M.; Hemanth, D.J.; Satapathy, S.C. Gastrointestinal diseases segmentation and classification based on duo-deep architectures. Pattern Recognit. Lett. 2020, 131, 193–204. [Google Scholar] [CrossRef]
  47. Chen, J.; Lian, Y.; Li, Y. ‘Real-time grain impurity sensing for rice combine harvesters using image processing and decision-tree algorithm. Comput. Electron. Agric. 2020, 175, 105591. [Google Scholar] [CrossRef]
  48. Moubayed, A.; Injadat, M.; Shami, A.; Lutfiyya, H. Student engagement level in an e-learning environment: Clustering using k-means. Am. J. Distance Educ. 2020, 34, 137–156. [Google Scholar] [CrossRef]
  49. Nanglia, P.; Kumar, S.; Mahajan, A.N.; Singh, P.; Rathee, D. A hybrid algorithm for lung cancer classification using SVM and Neural Networks. ICT Express 2021, 7, 335–341. [Google Scholar] [CrossRef]
  50. Wu, W.; Li, D.; Du, J.; Gao, X.; Gu, W.; Zhao, F.; Feng, X.; Yan, H. An intelligent diagnosis method of brain MRI tumor segmentation using deep convolutional neural network and SVM algorithm. Comput. Math. Methods Med. 2020, 2020, 6789306. [Google Scholar] [CrossRef]
  51. Basha, C.Z.; Rohini, G.; Jayasri, A.V.; Anuradha, S. Enhanced and Effective Computerized Classification of X-Ray Images; IEEE: Piscataway, NJ, USA, 2020; pp. 86–91. [Google Scholar]
  52. Islam, N.; Rashid, M.M.; Wibowo, S.; Xu, C.-Y.; Morshed, A.; Wasimi, S.A.; Moore, S.; Rahman, S.M. Early Weed Detection Using Image Processing and Machine Learning Techniques in an Australian Chilli Farm. Agriculture 2021, 11, 387. [Google Scholar] [CrossRef]
  53. Sasankar, P.; Kosarkar, U. A Study for Face Recognition Using Techniques PCA and KNN. EasyChair Print. 2021. Available online: https://wwww.easychair.org/publications/preprint_download/gS7Q (accessed on 30 January 2023).
  54. Bharate, A.A.; Shirdhonkar, M. Classification of Grape Leaves Using KNN and SVM Classifiers; IEEE: Piscataway, NJ, USA, 2020; pp. 745–749. [Google Scholar]
  55. Goncharov, P.; Ososkov, G.; Nechaevskiy, A.; Uzhinskiy, A.; Nestsiarenia, I. Disease Detection on the Plant Leaves by Deep Learning; Springer: Berlin/Heidelberg, Germany, 2018; pp. 151–159. [Google Scholar]
  56. De Luna, R.G.; Dadios, E.P.; Bandala, A.A. Automated Image Capturing System for Deep Learning-Based Tomato Plant Leaf Disease Detection and Recognition; IEEE: Piscataway, NJ, USA, 2018; pp. 1414–1419. [Google Scholar]
  57. Sharma, P.; Berwal, Y.P.S.; Ghai, W. Performance analysis of deep learning CNN models for disease detection in plants using image segmentation. Inf. Process. Agric. 2020, 7, 566–574. [Google Scholar] [CrossRef]
  58. Oppenheim, D.; Shani, G.; Erlich, O.; Tsror, L. Using deep learning for image-based potato tuber disease detection. Phytopathology 2019, 109, 1083–1087. [Google Scholar] [CrossRef]
  59. Chowdhury, M.E.; Rahman, T.; Khandakar, A.; Ayari, M.A.; Khan, A.U.; Khan, M.S.; Al-Emadi, N.; Reaz, M.B.I.; Islam, M.T.; Ali, S.H.M. Automatic and Reliable Leaf Disease Detection Using Deep Learning Techniques. AgriEngineering 2021, 3, 294–312. [Google Scholar] [CrossRef]
  60. Krizhevsky, A.; Hinton, G. Learning Multiple Layers of Features from Tiny Images. 2009. Available online: http://www.cs.utoronto.ca/~kriz/learning-features-2009-TR.pdf (accessed on 30 January 2023).
  61. Rauf, H.T.; Lali, M.I.U. A Guava Fruits and Leaves Dataset for Detection and Classification of Guava Diseases through Machine Learning. Mendeley Data 2021, 1. (Dataset). [Google Scholar]
Figure 1. Proposed Framework for disease recognition in guava leaves.
Figure 1. Proposed Framework for disease recognition in guava leaves.
Agriculture 13 00667 g001
Figure 2. RGB to YCbCr conversion.
Figure 2. RGB to YCbCr conversion.
Agriculture 13 00667 g002
Figure 3. The architecture of the proposed CNN network, SidNet.
Figure 3. The architecture of the proposed CNN network, SidNet.
Agriculture 13 00667 g003
Figure 4. Visualization of the images of strong feature maps of different convolution layers. (a) conv_1 (b) conv_ 2 (c) conv_ 3 (d) conv_4 (e) conv_5 (f) conv_6 (g) conv_7 (h) conv2.
Figure 4. Visualization of the images of strong feature maps of different convolution layers. (a) conv_1 (b) conv_ 2 (c) conv_ 3 (d) conv_4 (e) conv_5 (f) conv_6 (g) conv_7 (h) conv2.
Agriculture 13 00667 g004
Figure 5. Visualization of the images of strong feature maps of different convolution layers of DarkNet-53. (a) conv3 (b) conv1 (c) conv2 (d) conv4 (e) conv5 (f) conv6.
Figure 5. Visualization of the images of strong feature maps of different convolution layers of DarkNet-53. (a) conv3 (b) conv1 (c) conv2 (d) conv4 (e) conv5 (f) conv6.
Agriculture 13 00667 g005
Figure 6. Visualization of the images of strong feature maps of different convolution layers of AlexNet. (a) conv2 (b) conv1 (c) conv3 (d) conv4 (e) conv5 (f) conv6.
Figure 6. Visualization of the images of strong feature maps of different convolution layers of AlexNet. (a) conv2 (b) conv1 (c) conv3 (d) conv4 (e) conv5 (f) conv6.
Agriculture 13 00667 g006
Figure 7. Sample dataset images of guava plant.
Figure 7. Sample dataset images of guava plant.
Agriculture 13 00667 g007
Figure 8. Confusion matrix for Q(SVM) with 3045 features and using 5-fold results.
Figure 8. Confusion matrix for Q(SVM) with 3045 features and using 5-fold results.
Agriculture 13 00667 g008
Figure 9. ROC curve for two classes A and B with 3045 features and using 5-fold results.
Figure 9. ROC curve for two classes A and B with 3045 features and using 5-fold results.
Agriculture 13 00667 g009
Figure 10. Confusion matrix for Q(SVM) with 3045 features and using 10-fold results.
Figure 10. Confusion matrix for Q(SVM) with 3045 features and using 10-fold results.
Agriculture 13 00667 g010
Figure 11. ROC curve for two classes A and B with 3045 features and using 10-fold results.
Figure 11. ROC curve for two classes A and B with 3045 features and using 10-fold results.
Agriculture 13 00667 g011
Figure 12. Confusion matrix for FG(SVM) with 200 features and using 5-fold results.
Figure 12. Confusion matrix for FG(SVM) with 200 features and using 5-fold results.
Agriculture 13 00667 g012
Figure 13. ROC curves for two classes A and B with 200 features and using 5-fold results.
Figure 13. ROC curves for two classes A and B with 200 features and using 5-fold results.
Agriculture 13 00667 g013
Figure 14. Confusion matrix for FG(SVM) with 200 features and using 10-fold results.
Figure 14. Confusion matrix for FG(SVM) with 200 features and using 10-fold results.
Agriculture 13 00667 g014
Figure 15. ROC curve for two classes A and B with 200 features and using 10-fold results.
Figure 15. ROC curve for two classes A and B with 200 features and using 10-fold results.
Agriculture 13 00667 g015
Figure 16. Confusion matrix for Q(SVM) with 500 features and using 500 fatures with 5-fold results.
Figure 16. Confusion matrix for Q(SVM) with 500 features and using 500 fatures with 5-fold results.
Agriculture 13 00667 g016
Figure 17. ROC curve for two classes A and B with 500 features and using 5-fold results.
Figure 17. ROC curve for two classes A and B with 500 features and using 5-fold results.
Agriculture 13 00667 g017
Figure 18. Confusion matrix for FG(SVM) with 500 features and using 10-fold results.
Figure 18. Confusion matrix for FG(SVM) with 500 features and using 10-fold results.
Agriculture 13 00667 g018
Figure 19. ROC curve for two classes A and B with 500 features and using 10-fold results.
Figure 19. ROC curve for two classes A and B with 500 features and using 10-fold results.
Agriculture 13 00667 g019
Figure 20. Graph showing training time and features on 5-fold.
Figure 20. Graph showing training time and features on 5-fold.
Agriculture 13 00667 g020
Figure 21. Graph showing accuracy and features on 5-fold.
Figure 21. Graph showing accuracy and features on 5-fold.
Agriculture 13 00667 g021
Figure 22. Graph showing features and training time.
Figure 22. Graph showing features and training time.
Agriculture 13 00667 g022
Figure 23. Graph showing features and accuracy on 10 folds.
Figure 23. Graph showing features and accuracy on 10 folds.
Agriculture 13 00667 g023
Table 1. An overview of recent literature regarding plants diseases analysis.
Table 1. An overview of recent literature regarding plants diseases analysis.
Ref.YearTechniquesDatasetDiseasesResults %
[41]2018Downscaling and squaring method, AlexNet, VGG-16, AlexNetOWTBn.87,848
58 classes
Apple Scab, Black Rot, Early Blight, Brown Leaf Spot99.53
[55]2018ResNet-50, Deep Siamese convolutional network, TSNE method, KNNPVDBlack Rot, Esca, Chlorosis90
[56]2018Transfer learning, F-RCNN, classification.4923Phoma Rot, Leaf Miner, Target Spot95.75
[57]2019F-CNN, S-CNN, Segmentation, annotation and labeling on region of interest (lesions), random transformation (stretch/rotation/brightness/contrast blur)Independent datasetSpider Mite, Target Spot98.6
[58]2019VGG classification, resizing and transformation of images into grayscale,2465Black dot and scurf96
[40]2021Efficient-Net (B5Ver), Alexnet, ResNet50, classification.61,486Late Blight, Bacterial Spot99.97
[59]2021Resizing, normalizing and augmentation, Efficient-Net (B7Ver), Efficient-Net (B4Ver), U-net, and modified U-net segmentation model, Score-Cam visualization technique,18,161Target Spot99.9
Table 2. Performance evaluation metrics.
Table 2. Performance evaluation metrics.
MeasuresMathematical Expressions
Accuracy True   Positive   +   True   Negative True   Positive   True   Negative   +   False   Positive   +   False   Negative
Recall True   Positive True   Positive   +   False   Positive
PRC/Positive Prediction True   Positive True   Positive   +   False   Positive
F1 Score 2 . Precision . Recall Precision   +   Recall
Table 3. Summary of results.
Table 3. Summary of results.
Test CasesExperiment #FoldsFeaturesClassifierAccuracy %Training Time (s)
11 (a)53045Quadratic SVM98.99.2
22 (a)5200Fine Gaussian SVM81.10.8
33 (a)5500Quadratic SVM84.12.2
44 (a)5750Cubic SVM85.62.4
55 (a)51000Cubic SVM87.63.2
61 (b)103045Quadratic SVM99.216.2
72 (b)10200Fine Gaussian SVM82.61.2
83 (b)10500Fine Gaussian SVM85.22.20
94 (b)10750Cubic SVM87.84.8
105 (b)101000Cubic SVM87.56.2
Table 4. Experiment_1 using 5 Folds (3045 Feature).
Table 4. Experiment_1 using 5 Folds (3045 Feature).
ClassifierAccuracy (%)Total CostPrecision Speed (obs/s)Training Time (s)PrecisionF1-ScoreRecall
Quadratic SVM98.9127609.210.990.990.99
Linear SVM98.0207809.650.980.980.98
Cubic SVM98.6147709.430.9850.9850.99
Fine Gaussian SVM80.119949016.30.780.7850.86
Medium Gaussian SVM97.3277709.760.970.9750.97
Coarse Gaussian SVM92.57574010.350.9150.9250.92
Weighted KNN95.05036012.890.9450.950.95
FINE KNN93.56536013.620.9350.9350.94
Medium KNN87.812236012.940.8650.8750.90
Coarse KNN81.318736012.810.7950.7950.87
Cosine KNN89.810133013.520.890.8950.91
Cubic KNN87.51257358.970.8650.870.90
Table 5. Experiment_1 using 10 Folds (3045 Feature).
Table 5. Experiment_1 using 10 Folds (3045 Feature).
ClassifierAccuracy (%)Total Cost Precision Speed (obs/s)Training Time (s)PrecisionRecallF1-Score
Quadratic SVM99.2846016.2210.9950.990.99
Linear SVM98.31744018.230.980.980.98
Cubic SVM99.1945016.70.990.990.99
Fine Gaussian SVM85.414631029.2650.840.890.84
Medium Gaussian SVM98.31745017.6760.980.980.98
Coarse Gaussian SVM93.66445018.6890.930.940.93
Weighted KNN94.35719023.0370.9350.950.94
FINE KNN94.25819024.3740.9350.950.94
Medium KNN88.211819023.120.870.910.87
Coarse KNN82.717319023.2780.810.880.81
Cosine KNN91.09018024.4380.90.920.90
Cubic KNN88.31175773.7130.870.910.87
Table 6. Experiment_2 using 5 Folds (200 Feature).
Table 6. Experiment_2 using 5 Folds (200 Feature).
ClassifierAccuracy (%)Total CostPrecision Speed (obs/s)Training Time (s)PrecisionRecallF1-Score
Fine Gaussian SVM81.118914,0000.815710.80.820.805
Linear SVM73.726416,0001.34860.7350.7350.73
Cubic SVM80.421015,0001.85720.8050.8050.8
Quadratic SVM78.421615,0001.74310.7850.780.785
Medium Gaussian SVM72.827315,0000.76670.710.740.715
Coarse Gaussian SVM60.539614,0000.6710.5650.710.5
Weighted KNN78.221877000.767410.7850.7850.78
FINE KNN74.825376001.19170.750.750.75
Medium KNN66.833378000.83670.660.6650.66
Coarse KNN62.437772000.790920.590.6750.555
Cosine KNN67.232975000.7666060.670.6650.67
Cubic KNN69.031112003.80960.680.6850.68
Table 7. Experiment_2 using 10 Folds (200 Feature).
Table 7. Experiment_2 using 10 Folds (200 Feature).
ClassifierAccuracy (%)Total CostPrecision Speed (obs/s)Training Time (s)PrecisionRecallF1-Score
Fine Gaussian SVM82.617483001.26640.8150.830.825
Linear SVM73.226980001.85760.7250.730.73
Cubic SVM80.819284005.65480.8150.810.81
Quadratic SVM78.221878003.42370.780.780.78
Medium Gaussian SVM73.926281001.150.7250.750.725
Coarse Gaussian SVM60.239981001.16680.5650.7150.49
Weighted KNN78.621441001.26620.790.7850.78
FINE KNN75.025042001.83580.7550.750.75
Medium KNN65.934240001.81290.6550.6550.65
Coarse KNN62.038140001.25440.590.670.55
Cosine KNN65.934239001.2740.6650.6650.655
Cubic KNN68.53169604.570.6750.680.68
Table 8. Experiment_3 using 5 Folds (500 Feature).
Table 8. Experiment_3 using 5 Folds (500 Feature).
ClassifierAccuracy (%)Total CostPrecision Speed (obs/s)Training Time
(s)
PrecisionRecallF1-Score
Quadratic SVM84.115970002.20450.840.840.84
Linear SVM79.021069001.70010.7850.790.79
Cubic SVM82.417673002.21171.641.651.65
Fine Gaussian SVM83.616469001.27681.661.681.67
Medium Gaussian SVM75.824269001.14610.7450.7650.75
Coarse Gaussian SVM67.033168001.18910.650.680.65
Weighted KNN80.419631001.58850.8050.8050.805
FINE KNN77.622432001.99720.780.7750.775
Medium KNN69.430732001.63580.6850.690.685
Coarse KNN67.433131001.5840.650.6950.64
Cosine KNN68.531630001.61650.6850.6850.685
Cubic KNN73.32684709.33250.720.730.73
Table 9. Experiment_3 using 10 Folds (500 Feature).
Table 9. Experiment_3 using 10 Folds (500 Feature).
ClassifierAccuracy (%)Total CostPrecision Speed (obs/sec)Training Time
(s)
PrecisionRecallF1-Score
Fine Gaussian SVM85.214838002.20880.8450.860.85
Linear SVM79.820236004.51370.790.80.79
Cubic SVM84.215841004.40020.8450.840.84
Quadratic SVM84.715340004.16440.8450.8450.845
Medium Gaussian SVM75.924140002.10570.750.7650.755
Coarse Gaussian SVM67.732439002.06350.660.690.655
Weighted KNN82.517518002.69160.830.8250.825
FINE KNN79.420617003.71720.7950.7950.79
Medium KNN71.528618002.70120.710.710.71
Coarse KNN66.433718002.6670.6450.6850.63
Cosine KNN67.932216002.7860.6850.680.68
Cubic KNN75.424638011.1540.750.750.75
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Mumtaz, S.; Raza, M.; Okon, O.D.; Rehman, S.U.; Ragab, A.E.; Rauf, H.T. A Hybrid Framework for Detection and Analysis of Leaf Blight Using Guava Leaves Imaging. Agriculture 2023, 13, 667. https://doi.org/10.3390/agriculture13030667

AMA Style

Mumtaz S, Raza M, Okon OD, Rehman SU, Ragab AE, Rauf HT. A Hybrid Framework for Detection and Analysis of Leaf Blight Using Guava Leaves Imaging. Agriculture. 2023; 13(3):667. https://doi.org/10.3390/agriculture13030667

Chicago/Turabian Style

Mumtaz, Sidrah, Mudassar Raza, Ofonime Dominic Okon, Saeed Ur Rehman, Adham E. Ragab, and Hafiz Tayyab Rauf. 2023. "A Hybrid Framework for Detection and Analysis of Leaf Blight Using Guava Leaves Imaging" Agriculture 13, no. 3: 667. https://doi.org/10.3390/agriculture13030667

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop