Next Article in Journal
Prediction of Sound Insulation Using Artificial Neural Networks—Part II: Lightweight Wooden Façade Structures
Previous Article in Journal
Changes in the Stress Tolerance of Dwarf Rabbits in Animal-Assisted Interventions
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Plant Disease Detection Using Deep Convolutional Neural Network

1
Computer Science & Engineering, Vel Tech Rangarajan Dr. Sagunthala R&D Institute of Science and Technology, Chennai 600062, India
2
Neuroaesthetic Lab, Bioinstrumentation and Medical Tecniques Group, Faculty of Electrical Engineering and Computer Science, Stefan cel Mare University of Suceava, 720229 Suceava, Romania
3
Electromechanical Faculty, Department of Telecomunication and Electronics, Maritime University of Constanta, 900663 Constanta, Romania
4
Department of Computer Science and Information Technology, University of Lahore, Lahore 54590, Pakistan
*
Author to whom correspondence should be addressed.
Appl. Sci. 2022, 12(14), 6982; https://doi.org/10.3390/app12146982
Submission received: 13 June 2022 / Revised: 6 July 2022 / Accepted: 8 July 2022 / Published: 10 July 2022
(This article belongs to the Section Computing and Artificial Intelligence)

Abstract

:
In this research, we proposed a novel 14-layered deep convolutional neural network (14-DCNN) to detect plant leaf diseases using leaf images. A new dataset was created using various open datasets. Data augmentation techniques were used to balance the individual class sizes of the dataset. Three image augmentation techniques were used: basic image manipulation (BIM), deep convolutional generative adversarial network (DCGAN) and neural style transfer (NST). The dataset consists of 147,500 images of 58 different healthy and diseased plant leaf classes and one no-leaf class. The proposed DCNN model was trained in the multi-graphics processing units (MGPUs) environment for 1000 epochs. The random search with the coarse-to-fine searching technique was used to select the most suitable hyperparameter values to improve the training performance of the proposed DCNN model. On the 8850 test images, the proposed DCNN model achieved 99.9655% overall classification accuracy, 99.7999% weighted average precision, 99.7966% weighted average recall, and 99.7968% weighted average F1 score. Additionally, the overall performance of the proposed DCNN model was better than the existing transfer learning approaches.

1. Introduction

The diagnosis and treatment of disease are essential to improving the growth and yield of agricultural plants. For instance, an average estimated yield loss by corn crop diseases in the United States and Ontario from 2012 to 2015 was USD 76.51 per acre [1]. Manual monitoring of plant diseases will not give accurate outcomes regularly [2]. Additionally, finding domain experts for monitoring plant diseases is highly difficult and expensive for farmers. For that reason, an intelligent plant disease diagnosis system was essential to monitor the agricultural fields regularly [3]. At present, several plant disease detection methods are proposed for automatic plant disease detection using artificial intelligence techniques with fewer human efforts [4]. Deep convolutional neural network (DCNN) is a most successful image classification technique [5]. The DCNN comprises various layers, such as convolutional, pooling and fully connected layers for learning features from the training data [6].
Transfer learning uses the pre-trained neural network from one task to a similar new task. Transfer learning techniques can minimize the time for model design and training. The standard transfer learning techniques in image classification are AlexNet, DenseNet, VGG16, Inception-v3, MobileNet and ResNet [7]. The DCNN requires more data for an efficient training process [8]. Data augmentation techniques are producing new images for the existing dataset using several data transformation techniques [9,10]. Basic image manipulation (BIM), deep convolutional generative adversarial network (DCGAN) and neural style transfer (NST) are popular data augmentation techniques [11]. The most common BIM techniques are affine transformation, scaling, cropping, flipping, padding, rotation, translation, brightness, contrast, saturation and hue. The DCGAN is an unsupervised neural network to create a new set of realistic images from the training data [12]. The GAN comprises two DCNNs, such as generator DCNN and discriminator DCNN. The generator DCNN creates new images similar to the training data; also, the discriminator network classifies the original and newly created images by the generator DCNN. DCGAN is one of the most successful image augmentation techniques in medical image processing applications [13]. The NST is an image transformation technique to produce new images using content and style reference images [12].
Additionally, the DCNNs need suitable hyperparameter values to improve the classification performance. Hyperparameters are the most significant training parameters that can influence the performance of deep learning techniques. Activation function, dropout value, epochs, filter size, learning rate, loss function and mini-batch size are the most common hyperparameters in DCNNs. The selection of the suitable value of hyperparameters is a challenging task in solving a deep learning problem. Hyperparameter tuning techniques are used to discover the most suitable hyperparameter values for the DCNNs [14]. Grid search and random search are the most popular hyperparameter tuning approaches in deep learning. High-performance computing power is needed to train the deep learning algorithm with better efficiency and less training time [15].
This article proposed a novel DCNN model for diagnosing 42 leaf diseases in 16 different plant species. This research used data augmentation and hyperparameter optimization techniques to improve the performance of the disease detection model. To produce augmented leaf images, BIM, DCGAN and NST techniques were used. There are 58 diseased and healthy plant leaf classes that were used to train the DCNN model. Random searching with the coarse-to-fine technique was used to optimize the value of the most common hyperparameters. Finally, the performance of the proposed DCNN was compared with the standard transfer learning approaches, such as AlexNet, Inception-v3-Net, ResNet-50 and VGG16Net. The rest of the article is organized as follows: Section 2 provides a detailed survey on plant leaf disease detection using artificial intelligence techniques. Section 3 describes the implementation process of the proposed DCNN model for plant leaf disease detection. Section 4 demonstrates the experimental results and related discussions of the research. Finally, Section 5 provides the conclusions and future directions of the research.

2. Related Works

The early detection of plant diseases is a significant step in the disease prevention and treatment process [6]. Accurate disease detection techniques can be used by the farmers for applying the prevention and treatment procedures [10,16]. The most recent plant disease detection techniques are reviewed in this section. In Ref. [4], the authors proposed a support vector machine (SVM) model for detecting sugar crop diseases. They used the hyperspectral images as an input of the SVM model for disease detection. The average classification accuracy of the hyperspectral image-based SVM technique was 78% in testing data. In Ref. [17], the authors proposed two huanglongbing disease identification models for citrus plants. The SVM and artificial neural network (ANN) were used to design the huanglongbing detection techniques. The classification accuracies of the SVM and ANN models in the test data were 92.8% and 92.2% respectively. Identification of tomato yellow leaf curl disease was achieved using the SVM technique with the quadratic kernel function proposed by the author in [18]. The overall classification accuracy of this algorithm was 92% in tomato yellow leaf curl disease detection.
Additionally, the authors in [19] discussed and compared numerous image processing and feature extraction techniques to identify the various plant diseases using their leaf images. The authors in [20] developed a DCNN model to detect legume plant species using vein morphological patterns. In Ref. [21], the authors proposed a region-based and single-shot multi-box detector CNN model for designing the plant disease and pest identification model. The VGGNet and ResNet were used to improve the classification performance of the model. Table 1 compares the various state-of-the-art DCNN models for plant disease detection proposed by different articles.
The data augmentation technique improves the diversity of training data without collecting new data. The authors in [11] compared the advantages of various data augmentation techniques in the DCNNs training process. The augmentation techniques are GAN, flipping, cropping, shifting, principal component analysis (PCA), color, noise and rotation. The result shows that the training performance of the cropping, flipping, GAN and rotation are higher than the other augmentation techniques. Additionally, the result proves that the combination of different augmentation techniques can give better performance than individuals. In Ref. [12], the authors introduced the BIM, GAN and NST augmentation techniques for plant leaf disease classification and compared the performance of each technique. The experimental result shows that the performance of the combined augmentation technique was better than the individual techniques. The authors in [32], discussed the advantages of the existing data augmentation techniques in deep learning applications. In Ref. [33], the authors proposed a DCNN for pest detection using GAN based image augmented dataset. The testing result shows that the classification performance of the GAN-based image augmented dataset was better than the non-augmented dataset. The authors in [13] discussed the advantages of the GAN augmentation technique in DCNN development. Moreover, the authors in [34] discussed the importance of hyperparameter tuning to achieve a better performance of DCNNs. The detailed survey indicates the importance of dataset size, augmentation techniques and selection of hyperparameter values in the plant leaf disease detection model. The following section presents the information about the proposed dataset and DCNN model for detecting various plant diseases from leaf images.

3. Materials and Methods

This section provides a complete description of the architecture and training process of the proposed DCNN model with the experimental setup and dataset preparation. The proposed plant leaf disease detection model pipeline starts with dataset preparation and ends with model prediction. Python 3.7 programming language and TensorFlow 2.9.1, numpy Version 1.19. 2, matplotlib Version 3.5.2 and OpenCV Version 4.5.5 libraries are used for dataset preparation and DCNN model implementation. Data preparation, preprocessing, model designing and prediction tasks are performed using an HP Z240 workstation with an Intel Core i7 CPU and sixteen gigabytes of random access memory. The training and testing process of the proposed DCNN and existing state-of-the-art techniques were performed using an NVidia DGX-1 deep learning server station. The deep learning server includes two Intel Xeon E5-2694 Version 4 CPUs and eight Tesla P100 GPUs for accelerating the training process of deep neural networks. In subsequent subsections, each phase of the proposed plant leaf disease detection pipeline are discussed in detail. First, the details of the dataset preparation and preprocess are discussed in the next subsection.

3.1. Dataset Preparation and Preprocessing

Diseased and healthy leaf images of various plants were collected from different standard open data repositories [9,35,36,37,38]. Sixteen different plant species were used to create the plant leaf disease dataset. Each plant contains healthy and the most common disease classes in the dataset. There are 58 different classes of plant leaves, and 1 no-leaves class is present in the dataset. The collected original dataset contains 61,459 plant leaves and no-leaves images. The list of plant names with the healthy and disease classes of the proposed dataset is shown in Table 2.
To create an even number of images in each class, data augmentation techniques were introduced. Additionally, the data augmentation techniques can increase the dataset size and reduce the overfitting during the training process of the model by adding some augmented images to the training dataset. The BIM, DCGAN and NST augmentation techniques were used to produce the augmented images in the dataset. The BIM augmentation techniques consist of image cropping, flipping, PCA color augmentation, rotation and scaling. The PCA color augmentation technique alters the intensity of the color channels using the principal component of the pixels [11]. Additionally, the image cropping, flipping, rotation and scaling techniques create augmented images by changing the color and position of the input images. There are 36,541 augmented images created by the BIM augmentation technique in the dataset.
DCGANs create augmented images that resemble the training data. The DCGAN consists of two DCNN networks, such as generator DCNN and discriminator DCNN. The generator DCNN network takes a vector of random noise and up-samples it to the training data. On the other hand, the discriminator DCNN learns to classify the real and generated images [33]. The DCGAN network was trained in the graphics processing units with training epochs of 10,000 and a mini-batch size of 64. There are 32,000 augmented images created by the DCGAN augmentation technique in the dataset. NST is another image generation technique using deep learning techniques. A modified VGG19 network was used to develop the NST augmentation model in this research. The NST model was trained with 5000 epochs on the deep learning server system. The NST models require two different input images to generate an augmented output image, such as the content image and the style reference image. At first, the content image contains the essential features to be added to the output image. Second, the style reference image contains style patterns to apply to the output image. To generate the output image, the NST augmentation model applies the style features of the style image to the content image. The NST augmentation technique creates 17,500 augmented images in the dataset. Finally, The BIM, DCGAN and NST techniques were created for the augmented images to balance the data counts in each class of the dataset. The proposed dataset is named the PlantDisease59 dataset. These augmentation techniques increased the number of images in the dataset from 61,459 to 147,500 images. Additionally, the size of individual classes increased to 2500 images in each. Leaf images on the PlantDisease59 dataset were captured in the face-up direction. Figure 1 shows the sample augmented images generated by the BIM, DCGAN and NST techniques.
The first two images in Figure 1 were created using BIM techniques. The third and fourth leaf images in Figure 1 were created using the DCGAN augmentation technique. The last two sample images in Figure 1 were generated using the NST technique. The random selection technique was used to select the images for training, validation and testing from the PlantDisease59 dataset. Table 3 illustrates the number of images in the training, validation and testing dataset.
The design and development of the Proposed DCNN model for plant leaf disease detection using the hyperparameter tuning techniques and the PlantDisease59 dataset are discussed in the following subsection.

3.2. Model Design

In this section, a DCNN model for diagnosing plant leaf diseases using the PlantDisease59 dataset is proposed. Several DCNN models with different numbers and sizes of convolutional (Conv) and pooling layers were developed, and their performance is compared. The number of Conv layers varies from three to eight in different DCNN models. At maximum, the 14-layered deep convolutional neural network (14-DCNN) gives better training performance than other developed models. Five convolutional and five max-pooling layers were used to develop the proposed 14-DCNN model. The input images of the 14-DCNN are given into the first two-dimensional Conv layer. The dimension of the Conv layer output can be calculated using Equation (1):
D i m e n s i o n ( Conv ( n , k ) ) = ( [ n w f w s + 1 ] ,   [ n h f h s + 1 ] ,   f C )
The input width ( n w ) and height ( n h ) of the first convolutional layer are 128 and 128 respectively. Additionally, the f w , f h and f c represent the width, height and channels of the kernel filter of the convolutional layer. The stride ( S ) value of this Conv layer is one. The first max-pooling layer was introduced to reduce the dimension of the first Conv layer output from 126, 126, 4 to 63, 63, 4 values. The dimension of the max-pooling layer output was calculated using Equation (2):
D i m e n s i o n ( P o o l i n g ( n , k ) ) = ( [ n w f w s + 1 ] ,   [ n h f h s + 1 ] ,   n C )
The n w ,   n h   and n c represent the width, height and channels of input n, respectively. Additionally, the   f w , f h and n c represent the width, height and channels of the filter ( f ) in the max-pooling layer. The output of the first max-pooling layer is given as an input of the second Conv layer.
Likewise, the second Conv layer uses the filter sizes of 16, 3, 3 values to extract the features from the data. The output size of the second Conv layer is 61, 61, 16 values. The second max-pooling layer reduces the output data size of the second Conv layer from 61, 61, 16 to 30, 30, 16 with the filter size of 2, 2 values. The third Conv layer was introduced after the second max-pooling layer. It extracts more features from the input data using the 32, 3, 3 size kernel and produces the 28, 28, 32 sized output data. The third max-pooling layer was used to reduce the size of the output data from the third Conv layer with 2, 2 sized kernels. It reduces the size of the data to 14, 14, 32 values from 28, 28, 32 values. Additionally, the fourth Conv layer uses the 64, 3, 3 sized kernels to extract the additional features from the third pooled data and generates the output data with the size of 12, 12, 64 values from the input size of 14, 14, 32 values. In addition, the fourth max pooling layer has a filter size of 2, 2 matrices and stride value of 1 step. The fourth max-pooling layer reduces the size from 12, 12, 64 values to 6, 6, 64 values. The fifth Conv layer was introduced with a filter size of 128, 3, 3 matrices. The size of the input matrix is 6, 6, 64 and the output size of the fifth Conv layer is 4, 4, 128 matrices. After the fifth Conv layer, the fifth pooing layer was used with max-pooling function and stride value of 1 step. The input size of the fifth max-pooling layer is 4, 4, 128 values, and the output size is 2, 2, 128 values. It uses the single-step stride values to apply the kernel to the input data from the fifth Conv layer. The ReLu activation function was used on all the above Conv layers. The ReLu activation function was performed by using Equation (3).
R e L u ( x ) = M a x ( 0 ,   x )
Moreover, the flatten layer was introduced after the fifth convolutional and pooling layer. It reduces the three-dimensional data to one-dimensional data for a traditional neural network approach. The flatten layer converts the output of the fifth max-pooling layer from 2, 2, 128 values to 512 values. The first dense layer was introduced after the flatten layer in the 14-DCNN. The first dense layer increases input value from 512 to 2048 values. Equation (4) represents the individual neuron output ( z j ) of the first dense layer. The i represents the number of inputs of the first dense layer, and it ranges from 1 to 512. Additionally, the j denotes the number of outputs of the layer its range from 1 to 2048 values.
z j = R e L u ( 0 ,   i 512 b j + x i w i )  
The x i and w i represent the value and weight of the ith input of the jth output. Additionally, the b j denotes the bias value of the jth node. The dropout layer was used between the first dense and second dense layer in the 14-DCNN to avoid the overfitting issue. The second dense layer was initiated after the first dense layer and dropout layer. Similarly, the number of inputs of the second dense layer is 2048 and the output values of the neural network is 59 values. Additionally, this layer uses the softmax activation function to classify the plant leaves. The softmax function ( σ ) value of ith neuron of the dense layer can be calculated using Equation (5).
s o f t m a x ( σ ( z i ) ) = e z i j = 1 59 e z j  
The output class of the input image can be discovered using Equation (6).
O u t p u t   C l a s s   ( z o u t ) = m a x ( z 1 ,   z 2 ,   z 59 )
This output value from z 1 to z 59 represents the number of diseases and healthy plant leaf and non-leaves classes in the PlantDisease59 dataset. The total number of training parameters is 5,424,583 in the 14-DCNN model. The layered structure of the 14-DCNN model is shown in Figure 2.
After designing the 14-DCNN model, the most suitable hyperparameter values were identified using hyperparameter tuning techniques. The random search and coarse-to-fine techniques were used to discover the suitable values of the optimizer function, mini-batch size and dropout probability of the 14-DCNN. The most common optimizers considered for the hyperparameter searching are adaptive moment estimation (Adam), stochastic gradient descent (SGD) and root mean square propagation (RMSprob). The range of the mini-batch size is between 8 and 256, incremented by 8 per value. Additionally, the dropout range varies between 0.0 and 0.5, incremented by 0.1 per value. The random searching technique chooses the random combination of hyperparameter values from the search space. The selected combination of hyperparameter values is applied to the 14-DCNN and trained with 100 epochs in parallel. From the training result, the coarse-to-fine process helps to identify the most possible hyperparameter values from the search space. Finally, the hyperparameter tuning technique discovers the SGD optimizer with a mini-batch size of 32 and the dropout probability of 0.2 gives a better performance than other values. To identify the learning rate (Lr) and momentum of the SGD optimizer, a similar random search approach was used with a possible combination of the values. Table 4 shows the most common hyperparameters of the 14-DCNN model for the plant leaf disease detection model with their values.
The random search with the coarse-to-fine technique offers significantly improved searching performance than the grid search and simple random search optimization techniques. The optimized hyperparameter values and the PlantDisease59 were used to train the proposed 14-DCNN model for diagnosing the diseases from the plant leaf images.

3.3. Model Training

The proposed 14-DCNN was trained using the optimized hyperparameters and augmented dataset in the deep learning server environment. The model was trained with different epoch values between 100 and 3000. The 1000 epochs gave the maximum validation accuracy and minimum loss. The training time of the proposed model with 1000 epochs was 7452 s in Nvidia DGX-1 deep learning server. Figure 3 demonstrates the training and validation performance of the proposed 14-DCNN model for the identification of plant leaf diseases.
The 14-DCNN model achieved a training accuracy of 99.993% and a validation accuracy of 99.985%. Training and validation accuracies of the proposed 14-DCNN are higher than the other proposed DCNN models. The proposed 14-DCNN took 7452 s for the training process in the MGPUs environment. The training time of the proposed 14-DCNN was smaller than the transfer learning techniques since the number of convolutional and pooling operations of the proposed 14-DCNN are lesser than the transfer learning techniques. Finally, the architecture and weights of the proposed 14-DCNN model were stored as a hierarchical data (H5) file for the further prediction process.

3.4. Model Prediction

The saved 14-DCNN model architecture and weights were used to detect the diseases of the various plants from the input images. The real-time plant disease images were given as an input of the 14-DCNN model. The 14-DCNN model successfully predicted the plant name and disease from the input images. The matplotlib package was used to visualize the predictions of the model. Figure 4 shows the random sample prediction from the proposed 14-DCNN model.
Finally, the 14-DCNN model was converted as a TensorFlow lite (tflite) file using a TensorFlow lite converter with a latency optimization approach. The tflite file can be used to deploy the model in mobile and embedded devices for real-time prediction.

4. Results

This section examines the performance of the proposed 14-DCNN model, using various performance evaluation approaches and testing datasets. Additionally, the performance of the proposed 14-DCNN model is compared with other state-of-the-art techniques. The state-of-the-art techniques are AlexNet [25], Inception-v3-Net, ResNet-50 and VGG16Net [7,8]. These models were trained on a deep learning server system using the PlantDisease59 dataset and tested using the testing dataset. The proposed and existing models were trained only on face-up leaf images. So, the models will give the best performance on the face-up direction positioned images.
Occlusion sensitivity can visualize the most important part of the input image for classification identified by the trained model. It can measure the sensitivity of the neural network to occlusion in different regions of the data, using small perturbations of the data. This region is known as the occluding region [7]. The white and light blue color pixels of the image illustrate the most essential parts of being classified into the expected class. The dark blue color region of the image has minimum features for the classification. Occluding stride and occluding size of the selected sensitivity map are 10 and 30, respectively. Figure 5 illustrates the occlusion sensitivity of the proposed 14-DCNN model on sample test data.
The confusion matrix is a summary of predictions made by the classification techniques. The confusion matrix of the classification technique represents the true positive (TP), true negative (TN), false positive (FP) and false negative (FN) values of every single class [3]. The area under the receiver operating characteristic (AUC-ROC) curve is one of the popular metrics that is used to evaluate the performance of learning algorithms. The ROC curve plots the difference between the true positive rate (TPR) and false positive rate (FPR) [15]. The TPR and FPR are calculated using Equations (7) and (8).
TPR = TP TP + FN  
FPR = FP FP + TN
The AUC-ROC curve of the cherry healthy and strawberry leaf scorch classes shows the classification advantage of the proposed 14-DCNN. Figure 6 illustrates the AUC-ROC curve of the cherry healthy and strawberry leaf scorch classes, using the proposed 14-DCNN model.
The performance of the proposed 14-DCNN model and state-of-the-art techniques is compared, using the most common performance metrics, such as classification accuracy, precision, recall and F1-score [3]. At first, the classification accuracy is defined as the correctly classified images divided by the total number of testing images. Precision is the second most important performance evaluation metric in classification techniques. Precision is defined as the number of correctly identified results divided by the number of correctly identified and correctly rejected results that are predicted by the model. Precision is used to find the correct proportion of the classification.
The third most important performance evaluation technique is recall. The recall is the number of correctly identified results divided by the number of correctly identified and incorrectly rejected results. The recalls are used to determine the proportion of actual positives that were correctly identified. The F1 score is one of the widely used metrics for the performance evaluation of machine learning algorithms. The F1 score is defined as the harmonic mean between precision and recall. The F1 score value represents the prediction advantage of the classification techniques. The following Equations (9)–(12) were used to calculate the accuracy, weighted average precision, weighted average recall and weighted average F1 score of the classification techniques.
Classification   accuracy = TP + TN TP + TN + FP + FN
Precision = TP TP + FP
Recall = TP TP + FN
F 1   Score = 2 × ( precision   ×   recall ) ( precision   +   recall )
The range of the classification accuracy, precision, recall and F1 score is between 0 (0%) and 1 (100%). The classification performance of the proposed 14-DCNN model on individual classes in the dataset is illustrated in Table 5.
Figure 7 compares the performance of the proposed 14-DCNN and existing state-of-the-art classification techniques using accuracy, weighted average precision, weighted average recall and weighted average F1 score.
The comparison results show the accuracy, precision, recall, and F1 scores of the proposed 14-DCNN are higher than the AlexNet, Inception-v3-Net, ResNet-50 and VGG16Net. Additionally, the complexity of the proposed 14-DCNN and other transfer learning techniques is illustrated in Table 6.
The complexity analysis result shows that the number of trainable parameters and model size of the proposed 14-DCNN model are lesser than the existing transfer learning techniques. The smaller number of trainable parameters and small model size will reduce the complexity of the model prediction process. The comparison results illustrate that the performance of the proposed 14-DCNN is higher than the AlexNet, Inception-v3-Net, ResNet-50 and VGG16Net on plant leaf disease classification.

5. Conclusions

A novel DCNN model was proposed to detect plant leaf diseases from leaf images in this research. The proposed 14-DCNN model was designed and trained to detect 42 leaf diseases in 16 plants through leaf images. The data augmentation and hyperparameter optimization techniques were also used to enhance the performance of the 14-DCNN in this research. Three augmentation techniques were used to enhance the dataset size to 147,500 images. The augmentation techniques are NST, DCGAN, and BIM. The individual class size, including original and augmented images, of the dataset, was 2500 images. The 14-DCNN comprises five Conv and five max-pooling layers. The random search with the coarse-to-fine technique was used to optimize the value of the hyperparameter for training the proposed 14-DCNN model. Training of the most successful 14-DCNN model was completed with the training and validation dataset of 139,000 images and optimized hyperparameter values. The proposed 14-DCNN model achieved a classification accuracy of 99.9655%, a precision value of 99.7999%, a recall value of 99.7966%, and an F1 score of 99.7968% on the training dataset. The optimized hyperparameter values and the data augmentation techniques had a considerable influence on the results of the proposed DCNN model. Compared with standard transfer learning techniques, the proposed 14-DCNN model has higher classification performance. An extension of this research will be adding new classes of plant diseases and an increasing number of training images in the dataset and modifying the architecture of the DCNN model using more convolutional and other layers. In the future, we plan to estimate the possibility of plant disease and analyze the severity using the deep learning technique. Moreover, we will extend disease detection from plant leaves to other parts of the plants, such as flowers, fruits, and stems.

Author Contributions

Conceptualization, J.A.P. and V.D.K.; writing—original draft preparation, J.A.P. and V.D.K.; writing—review, designing, analysis and editing, J.A.P., V.D.K., O.G., M.H., M.A. and K.K.; supervision, V.D.K.; funding acquisition, M.H. All the authors contributed in writing, reviewing, and structuring of the work. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by a grant of the Romanian National Authority for Scientific Research and Innovation, CCCDI-UEFISCDI, project number 203, COFUND-ICT-AGRI-FOOD-MERIAVINO-1, within PNCDI III.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Pandian, J.A.; Kanchanadevi, K.; Kumar, V.D.; Jasińska, E.; Goňo, R.; Leonowicz, Z.; Jasiński, M. A Five Convolutional Layer Deep Convolutional Neural Network for Plant Leaf Disease Detection. Electronics 2022, 11, 1266. [Google Scholar] [CrossRef]
  2. Sladojevic, S.; Arsenovic, M.; Anderla, A.; Culibrk, D.; Stefanovic, D. Deep Neural Networks Based Recognition of Plant Diseases by Leaf Image Classification. Comput. Intell. Neurosci. 2016, 2016, 3289801. [Google Scholar] [CrossRef] [Green Version]
  3. Geetharamani, G.; Pandian, A. Identification of plant leaf diseases using a nine-layer deep convolutional neural network. Comput. Electr. Eng. 2019, 76, 323–338. [Google Scholar]
  4. Rumpf, T.; Mahlein, A.-K.; Steiner, U.; Oerke, E.-C.; Dehne, H.-W.; Plümer, L. Early detection and classification of plant diseases with Support Vector Machines based on hyperspectral reflectance. Comput. Electron. Agric. 2010, 74, 91–99. [Google Scholar] [CrossRef]
  5. Chen, H.-C.; Widodo, A.M.; Wisnujati, A.; Rahaman, M.; Lin, J.C.-W.; Chen, L.; Weng, C.-E. AlexNet Convolutional Neural Network for Disease Detection and Classification of Tomato Leaf. Electronics 2022, 11, 951. [Google Scholar] [CrossRef]
  6. Lee, S.H.; Chan, C.S.; Wilkin, P.; Remagnino, P. Deep-plant: Plant identification with convolutional neural networks. In Proceedings of the 2015 IEEE International Conference on Image Processing (ICIP), Quebec City, QC, Canada, 27–30 September 2015; pp. 452–456. [Google Scholar]
  7. Brahimi, M.; Arsenovic, M.; Laraba, S.; Sladojevic, S.; Boukhalfa, K.; Moussaoui, A. Deep Learning for Plant Diseases: Detection and Saliency Map Visualisation. In Human and Machine Learning: Visible, Explainable, Trustworthy and Transparent; Zhou, J., Chen, F., Eds.; Springer International Publishing: Cham, Switzerland, 2018; pp. 93–117. [Google Scholar]
  8. Ferentinos, K.P. Deep learning models for plant disease detection and diagnosis. Comput. Electron. Agric. 2018, 145, 311–318. [Google Scholar] [CrossRef]
  9. Arun Pandian, J.; Geetharamani, G. Data for: Identification of Plant Leaf Diseases Using a 9-Layer Deep Convolutional Neural Network. Mendeley Data. 2019. Available online: https://data.mendeley.com/datasets/tywbtsjrjv/1 (accessed on 29 March 2020).
  10. Almadhor, A.; Rauf, H.T.; Lali, M.I.U.; Damaševičius, R.; Alouffi, B.; Alharbi, A. AI-Driven Framework for Recognition of Guava Plant Diseases through Machine Learning from DSLR Camera Sensor Based High Resolution Imagery. Sensors 2021, 21, 3830. [Google Scholar] [CrossRef]
  11. Shijie, J.; Ping, W.; Peiyi, J.; Siping, H. Research on data augmentation for image classification based on convolution neural networks. In Proceedings of the 2017 Chinese Automation Congress (CAC), Jinan, China, 20–22 October 2017; pp. 4165–4170. [Google Scholar]
  12. Pandian, J.A.; Geetharamani, G.; Annette, B. Data augmentation on plant leaf disease image dataset using image manipulation and deep learning techniques. In Proceedings of the 2019 IEEE 9th International Conference on Advanced Computing (IACC), Tiruchirappalli, India, 13–14 December 2019; pp. 199–204. [Google Scholar]
  13. Bang, S.; Baek, F.; Park, S.; Kim, W.; Kim, H. Image augmentation to improve construction resource detection using generative adversarial networks, cut-and-paste and image transformation techniques. Autom. Constr. 2020, 115, 103198. [Google Scholar] [CrossRef]
  14. Trivedi, N.K.; Gautam, V.; Anand, A.; Aljahdali, H.M.; Villar, S.G.; Anand, D.; Goyal, N.; Kadry, S. Early Detection and Classification of Tomato Leaf Disease Using High-Performance Deep Neural Network. Sensors 2021, 21, 7987. [Google Scholar] [CrossRef]
  15. Agarwal, M.; Gupta, S.K.; Biswas, K.K. Development of Efficient CNN model for Tomato crop disease identification. Sustain. Comput. Inform. Syst. 2020, 28, 100407. [Google Scholar] [CrossRef]
  16. Ebrahimi, M.A.; Khoshtaghaza, M.H.; Minaei, S.; Jamshidi, B. Vision-based pest detection based on SVM classification method. Comput. Electron. Agric. 2017, 137, 52–58. [Google Scholar] [CrossRef]
  17. Wetterich, C.B.; Neves, R.F.D.O.; Belasque, J.; Ehsani, R.; Marcassa, L.G. Detection of Huanglongbing in Florida using fluorescence imaging spectroscopy and machine-learning methods. Appl. Opt. 2017, 56, 15–23. [Google Scholar] [CrossRef]
  18. Mokhtar, U.; Ali, M.A.; Hassanien, A.E.; Hefny, H. Identifying Two of Tomatoes Leaf Viruses Using Support Vector Machine. In Information Systems Design and Intelligent Applications; Mandal, J.K., Satapathy, S.C., Sanyal, M.K., Sarkar, P.P., Mukhopadhyay, A., Eds.; Springer India: New Delhi, India, 2015; pp. 771–782. [Google Scholar]
  19. Bharate, A.A.; Shirdhonkar, M.S. A review on plant disease detection using image processing. In Proceedings of the 2017 International Conference on Intelligent Sustainable Systems (ICISS), Palladam, India, 7–8 December 2017; pp. 103–109. [Google Scholar]
  20. Grinblat, G.L.; Uzal, L.C.; Larese, M.G.; Granitto, P.M. Deep learning for plant identification using vein morphological patterns. Comput. Electron. Agric. 2016, 127, 418–424. [Google Scholar] [CrossRef] [Green Version]
  21. Fuentes, A.; Yoon, S.; Kim, S.C.; Park, D.S. A Robust Deep-Learning-Based Detector for Real-Time Tomato Plant Diseases and Pests Recognition. Sensors 2017, 17, 2022. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  22. DeChant, C.; Wiesner-Hanks, T.; Chen, S.; Stewart, E.L.; Yosinski, J.; Gore, M.A.; Nelson, R.J.; Lipson, H. Automated Identification of Northern Leaf Blight-Infected Maize Plants from Field Imagery Using Deep Learning. Phytopathology 2017, 107, 1426–1432. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  23. Johannes, A.; Picon, A.; Alvarez-Gila, A.; Echazarra, J.; Rodriguez-Vaamonde, S.; Navajas, A.D.; Ortiz-Barredo, A. Automatic plant disease diagnosis using mobile capture devices, applied on a wheat use case. Comput. Electron. Agric. 2017, 138, 200–209. [Google Scholar] [CrossRef]
  24. Kawasaki, Y.; Uga, H.; Kagiwada, S.; Iyatomi, H. Basic study of automated diagnosis of viral plant diseases using convolutional neural networks. In Advances in Visual Computing; Bebis, G., Boyle, R., Parvin, B., Koracin, D., McMahan, R., Jerald, J., Eds.; Springer International Publishing: Cham, Switzerland, 2015; pp. 638–645. [Google Scholar]
  25. Nachtigall, L.G.; Araujo, R.M.; Nachtigall, G.R. Classification of apple tree disorders using convolutional neural networks. In Proceedings of the 2016 IEEE 28th International Conference on Tools with Artificial Intelligence (ICTAI), San Jose, CA, USA, 6–8 November 2016; pp. 472–476. [Google Scholar]
  26. Rangarajan, A.K.; Purushothaman, R.; Ramesh, A. Tomato crop disease classification using pre-trained deep learning algorithm. Procedia Comput. Sci. 2018, 133, 1040–1047. [Google Scholar] [CrossRef]
  27. Zhang, X.; Qiao, Y.; Meng, F.; Fan, C.; Zhang, M. Identification of Maize Leaf Diseases Using Improved Deep Convolutional Neural Networks. IEEE Access 2018, 6, 30370–30377. [Google Scholar] [CrossRef]
  28. Brahimi, M.; Boukhalfa, K.; Moussaoui, A. Deep Learning for Tomato Diseases: Classification and Symptoms Visualization. Appl. Artif. Intell. 2017, 31, 299–315. [Google Scholar] [CrossRef]
  29. Lu, Y.; Yi, S.; Zeng, N.; Liu, Y.; Zhang, Y. Identification of rice diseases using deep convolutional neural networks. Neurocomputing 2017, 267, 378–384. [Google Scholar] [CrossRef]
  30. Mohanty, S.P.; Hughes, D.P.; Salathé, M. Using Deep Learning for Image-Based Plant Disease Detection. Front. Plant Sci. 2016, 7, 1419. [Google Scholar] [CrossRef] [Green Version]
  31. Arnal Barbedo, J.G. Plant disease identification from individual lesions and spots using deep learning. Biosyst. Eng. 2019, 180, 96–107. [Google Scholar] [CrossRef]
  32. Shorten, C.; Khoshgoftaar, T.M. A survey on Image Data Augmentation for Deep Learning. J. Big Data 2019, 6, 60. [Google Scholar] [CrossRef]
  33. Lu, C.-Y.; Rustia, D.J.A.; Lin, T.-T. Generative Adversarial Network Based Image Augmentation for Insect Pest Classification Enhancement. IFAC-PapersOnLine 2019, 52, 1–5. [Google Scholar] [CrossRef]
  34. Saranyaraj, D.; Manikandan, M.; Maheswari, S. A deep convolutional neural network for the early detection of breast carcinoma with respect to hyper- parameter tuning. Multimed. Tools Appl. 2020, 79, 11013–11038. [Google Scholar]
  35. Hu, G.; Wu, H.; Zhang, Y.; Wan, M. Data for: A Low Shot Learning Method for Tea Leaf’s Disease Identification. Mendeley Data. 2019. Available online: https://data.mendeley.com/datasets/dbjyfkn6jr/1 (accessed on 29 March 2020).
  36. Kour, V.P.; Arora, S. Plantaek: A Leaf Database of Native Plants of Jammu and Kashmir. Mendeley Data. 2019. Available online: https://data.mendeley.com/datasets/t6j2h22jpx/2 (accessed on 29 March 2020).
  37. Krohling, R.A.; Esgario, J.; Ventura, J.A. BRACOL—A Brazilian Arabica Coffee Leaf Images Dataset to Identification and Quantification of Coffee Diseases and Pests. Mendeley Data. 2019. Available online: https://data.mendeley.com/datasets/yy2k5y8mxg/1 (accessed on 29 March 2020).
  38. Parraga-Alava, J.; Cusme, K.; Loor, A.; Santander, E. RoCoLe: A Robusta Coffee Leaf Images Dataset. Mendeley Data. 2019. Available online: https://data.mendeley.com/datasets/c5yvn32dzg/2 (accessed on 29 March 2020).
Figure 1. Sample augmented images using BIM, DCGAN and NST techniques.
Figure 1. Sample augmented images using BIM, DCGAN and NST techniques.
Applsci 12 06982 g001
Figure 2. Layered structure of 14-DCNN model.
Figure 2. Layered structure of 14-DCNN model.
Applsci 12 06982 g002
Figure 3. (a) Accuracy and (b) loss of the proposed 14-DCNN model.
Figure 3. (a) Accuracy and (b) loss of the proposed 14-DCNN model.
Applsci 12 06982 g003
Figure 4. Sample predicted images using the 14-DCNN model.
Figure 4. Sample predicted images using the 14-DCNN model.
Applsci 12 06982 g004
Figure 5. (a) Original image, (b) occlusion sensitivity of 14-DCNN.
Figure 5. (a) Original image, (b) occlusion sensitivity of 14-DCNN.
Applsci 12 06982 g005
Figure 6. Sample AUC-ROC curves of 14-DCNN.
Figure 6. Sample AUC-ROC curves of 14-DCNN.
Applsci 12 06982 g006
Figure 7. Performance comparison of proposed 14-DCNN and existing model.
Figure 7. Performance comparison of proposed 14-DCNN and existing model.
Applsci 12 06982 g007
Table 1. Comparison of different DCNN architecture.
Table 1. Comparison of different DCNN architecture.
ArticleYearSpecieNumber of ClassesNumber of ImagesArchitectureAccuracy (%)
[22]2017Maize21796Custom96.7
[23]2017Wheat33500Custom81.04
[24]2015Cucumber3800Custom94.9
[25]2016Apple51450AlexNet97.3
[26]2018Tomato713,262VGG16Net97.29
[27]2018Maize93060GoogLeNet98.9
[28]2017Tomato914,828GoogLeNet99.18
[29]2017Rice10500AlexNet95.48
[2]2016Multiple154483CaffeNet96.3
[30]2016Multiple3854,306GoogLeNet99.35
[7]2018Multiple3854,323InceptionV3Net99.76
[3]2019Multiple3961,486Custom96.46
[8]2018Multiple5887,848VGG16Net99.53
[31]2019Multiple7946,409GoogLeNet86.5
[15]2020Tomato1018,160Custom98.7
[14]2021Tomato103000Custom98.49
[5]2021Tomato1018,345AlexNet98.0
[1]2022Multiple38240,000Custom98.41
Table 2. List of classes in the proposed dataset.
Table 2. List of classes in the proposed dataset.
S. NoPlant NameClass Names
1Aloe VeraHealthy
2Leaf Rot
3Leaf Rust
4AppleHealthy
5Leaf Scab
6Black Rot
7Leaf Rust
8BananaHealthy
9Bacterial Wilt
10Black Sigatoka
11CherryHealthy
12Powdery Mildew
13CitrusHealthy
14Black Spot
15Canker
16Greening
17Melanose
18CornHealthy
19Common Rust
20Leaf Spot
21Northern Leaf Blight
22CoffeeHealthy
23Cercospora Leaf Spot
24Leaf Rust
25Red Spider Mite
26GrapeHealthy
27Black Measles
28Black Rot
29Leaf Blight
30PaddyHealthy
31Brown Spot
32Hispa
33Leaf Blast
34PeachHealthy
35Bacterial Spot
36PepperHealthy
37Bacterial Spot
38PotatoHealthy
39Early Blight
40Late Blight
41StrawberryHealthy
42Leaf Scorch
43TeaHealthy
44Leaf Blight
45Red Leaf Spot
46Red Scab
47TomatoHealthy
48Bacterial Spot
49Early Blight
50Late Blight
51Leaf Mold
52Leaf Spot
53Spider Mite
54Target Spot
55Mosaic Virus
56Yellow Leaf Curl Virus
57WheatHealthy
58Leaf Rust
59no-leavesno-leaves
Table 3. Training, validation and test dataset size.
Table 3. Training, validation and test dataset size.
Dataset NameNumber of ImagesNumber of Images in Each Class
Training Set132,7502250
Validation Set5900100
Testing Set8850150
Table 4. Optimized hyperparameters of the 14-DCNN.
Table 4. Optimized hyperparameters of the 14-DCNN.
HyperparameterValue
Batch Sizes32
Dropout Value0.2
LossCategorical Cross entropy
OptimizerSGD with Lr = 0.0001 and momentum = 0.9
Activation function for Conv layerReLu
Table 5. Class-wise performance of proposed 14-DCNN model.
Table 5. Class-wise performance of proposed 14-DCNN model.
Plant NameClass NamesPRECISIONRECALLF1-SCORE
Aloe VeraHealthy111
Leaf Rot10.986670.99329
Leaf Rust0.9868410.99338
AppleHealthy111
Leaf Scab111
Black Rot10.986670.99329
Leaf Rust111
BananaHealthy111
Bacterial Wilt0.9933810.99668
Black Sigatoka10.993330.99666
CherryHealthy111
Powdery Mildew111
CitrusHealthy111
Black Spot0.9868410.99338
Canker111
Greening111
Melanose10.986670.99329
CornHealthy111
Common Rust111
Leaf Spot111
Northern Leaf Blight111
CoffeeHealthy111
Cercospora Leaf Spot111
Leaf Rust111
Red Spider Mite111
GrapeHealthy111
Black Measles10.980.9899
Black Rot0.9677410.98361
Leaf Blight111
PaddyHealthy111
Brown Spot0.9803910.9901
Hispa111
Leaf Blast111
PeachHealthy111
Bacterial Spot0.986580.980.98328
PepperHealthy111
Bacterial Spot10.986670.99329
PotatoHealthy111
Early Blight111
Late Blight111
StrawberryHealthy111
Leaf Scorch111
TeaHealthy111
Leaf Blight10.986670.99329
Red Leaf Spot0.9868410.99338
Red Scab111
TomatoHealthy111
Bacterial Spot111
Early Blight111
Late Blight10.993330.99666
Leaf Mold0.9933810.99668
Leaf Spot111
Spider Mite111
Target Spot111
Mosaic Virus111
Yellow Leaf Curl Virus111
WheatHealthy111
Leaf Rust111
No LeavesNo Leaves111
Table 6. Complexity comparison of proposed 14-DCNN and existing model.
Table 6. Complexity comparison of proposed 14-DCNN and existing model.
AlexNetInception-v3-NetResNet-50VGG16Net14-DCNN
No. of Parameters44,752,73924,937,28326,722,21139,443,04317,928,571
Model Size (MB)133929812837
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Pandian, J.A.; Kumar, V.D.; Geman, O.; Hnatiuc, M.; Arif, M.; Kanchanadevi, K. Plant Disease Detection Using Deep Convolutional Neural Network. Appl. Sci. 2022, 12, 6982. https://doi.org/10.3390/app12146982

AMA Style

Pandian JA, Kumar VD, Geman O, Hnatiuc M, Arif M, Kanchanadevi K. Plant Disease Detection Using Deep Convolutional Neural Network. Applied Sciences. 2022; 12(14):6982. https://doi.org/10.3390/app12146982

Chicago/Turabian Style

Pandian, J. Arun, V. Dhilip Kumar, Oana Geman, Mihaela Hnatiuc, Muhammad Arif, and K. Kanchanadevi. 2022. "Plant Disease Detection Using Deep Convolutional Neural Network" Applied Sciences 12, no. 14: 6982. https://doi.org/10.3390/app12146982

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop