You are currently viewing a new version of our website. To view the old version click .
Applied Sciences
  • Article
  • Open Access

10 July 2022

Plant Disease Detection Using Deep Convolutional Neural Network

,
,
,
,
and
1
Computer Science & Engineering, Vel Tech Rangarajan Dr. Sagunthala R&D Institute of Science and Technology, Chennai 600062, India
2
Neuroaesthetic Lab, Bioinstrumentation and Medical Tecniques Group, Faculty of Electrical Engineering and Computer Science, Stefan cel Mare University of Suceava, 720229 Suceava, Romania
3
Electromechanical Faculty, Department of Telecomunication and Electronics, Maritime University of Constanta, 900663 Constanta, Romania
4
Department of Computer Science and Information Technology, University of Lahore, Lahore 54590, Pakistan
This article belongs to the Section Computing and Artificial Intelligence

Abstract

In this research, we proposed a novel 14-layered deep convolutional neural network (14-DCNN) to detect plant leaf diseases using leaf images. A new dataset was created using various open datasets. Data augmentation techniques were used to balance the individual class sizes of the dataset. Three image augmentation techniques were used: basic image manipulation (BIM), deep convolutional generative adversarial network (DCGAN) and neural style transfer (NST). The dataset consists of 147,500 images of 58 different healthy and diseased plant leaf classes and one no-leaf class. The proposed DCNN model was trained in the multi-graphics processing units (MGPUs) environment for 1000 epochs. The random search with the coarse-to-fine searching technique was used to select the most suitable hyperparameter values to improve the training performance of the proposed DCNN model. On the 8850 test images, the proposed DCNN model achieved 99.9655% overall classification accuracy, 99.7999% weighted average precision, 99.7966% weighted average recall, and 99.7968% weighted average F1 score. Additionally, the overall performance of the proposed DCNN model was better than the existing transfer learning approaches.

1. Introduction

The diagnosis and treatment of disease are essential to improving the growth and yield of agricultural plants. For instance, an average estimated yield loss by corn crop diseases in the United States and Ontario from 2012 to 2015 was USD 76.51 per acre [1]. Manual monitoring of plant diseases will not give accurate outcomes regularly [2]. Additionally, finding domain experts for monitoring plant diseases is highly difficult and expensive for farmers. For that reason, an intelligent plant disease diagnosis system was essential to monitor the agricultural fields regularly [3]. At present, several plant disease detection methods are proposed for automatic plant disease detection using artificial intelligence techniques with fewer human efforts [4]. Deep convolutional neural network (DCNN) is a most successful image classification technique [5]. The DCNN comprises various layers, such as convolutional, pooling and fully connected layers for learning features from the training data [6].
Transfer learning uses the pre-trained neural network from one task to a similar new task. Transfer learning techniques can minimize the time for model design and training. The standard transfer learning techniques in image classification are AlexNet, DenseNet, VGG16, Inception-v3, MobileNet and ResNet [7]. The DCNN requires more data for an efficient training process [8]. Data augmentation techniques are producing new images for the existing dataset using several data transformation techniques [9,10]. Basic image manipulation (BIM), deep convolutional generative adversarial network (DCGAN) and neural style transfer (NST) are popular data augmentation techniques [11]. The most common BIM techniques are affine transformation, scaling, cropping, flipping, padding, rotation, translation, brightness, contrast, saturation and hue. The DCGAN is an unsupervised neural network to create a new set of realistic images from the training data [12]. The GAN comprises two DCNNs, such as generator DCNN and discriminator DCNN. The generator DCNN creates new images similar to the training data; also, the discriminator network classifies the original and newly created images by the generator DCNN. DCGAN is one of the most successful image augmentation techniques in medical image processing applications [13]. The NST is an image transformation technique to produce new images using content and style reference images [12].
Additionally, the DCNNs need suitable hyperparameter values to improve the classification performance. Hyperparameters are the most significant training parameters that can influence the performance of deep learning techniques. Activation function, dropout value, epochs, filter size, learning rate, loss function and mini-batch size are the most common hyperparameters in DCNNs. The selection of the suitable value of hyperparameters is a challenging task in solving a deep learning problem. Hyperparameter tuning techniques are used to discover the most suitable hyperparameter values for the DCNNs [14]. Grid search and random search are the most popular hyperparameter tuning approaches in deep learning. High-performance computing power is needed to train the deep learning algorithm with better efficiency and less training time [15].
This article proposed a novel DCNN model for diagnosing 42 leaf diseases in 16 different plant species. This research used data augmentation and hyperparameter optimization techniques to improve the performance of the disease detection model. To produce augmented leaf images, BIM, DCGAN and NST techniques were used. There are 58 diseased and healthy plant leaf classes that were used to train the DCNN model. Random searching with the coarse-to-fine technique was used to optimize the value of the most common hyperparameters. Finally, the performance of the proposed DCNN was compared with the standard transfer learning approaches, such as AlexNet, Inception-v3-Net, ResNet-50 and VGG16Net. The rest of the article is organized as follows: Section 2 provides a detailed survey on plant leaf disease detection using artificial intelligence techniques. Section 3 describes the implementation process of the proposed DCNN model for plant leaf disease detection. Section 4 demonstrates the experimental results and related discussions of the research. Finally, Section 5 provides the conclusions and future directions of the research.

3. Materials and Methods

This section provides a complete description of the architecture and training process of the proposed DCNN model with the experimental setup and dataset preparation. The proposed plant leaf disease detection model pipeline starts with dataset preparation and ends with model prediction. Python 3.7 programming language and TensorFlow 2.9.1, numpy Version 1.19. 2, matplotlib Version 3.5.2 and OpenCV Version 4.5.5 libraries are used for dataset preparation and DCNN model implementation. Data preparation, preprocessing, model designing and prediction tasks are performed using an HP Z240 workstation with an Intel Core i7 CPU and sixteen gigabytes of random access memory. The training and testing process of the proposed DCNN and existing state-of-the-art techniques were performed using an NVidia DGX-1 deep learning server station. The deep learning server includes two Intel Xeon E5-2694 Version 4 CPUs and eight Tesla P100 GPUs for accelerating the training process of deep neural networks. In subsequent subsections, each phase of the proposed plant leaf disease detection pipeline are discussed in detail. First, the details of the dataset preparation and preprocess are discussed in the next subsection.

3.1. Dataset Preparation and Preprocessing

Diseased and healthy leaf images of various plants were collected from different standard open data repositories [9,35,36,37,38]. Sixteen different plant species were used to create the plant leaf disease dataset. Each plant contains healthy and the most common disease classes in the dataset. There are 58 different classes of plant leaves, and 1 no-leaves class is present in the dataset. The collected original dataset contains 61,459 plant leaves and no-leaves images. The list of plant names with the healthy and disease classes of the proposed dataset is shown in Table 2.
Table 2. List of classes in the proposed dataset.
To create an even number of images in each class, data augmentation techniques were introduced. Additionally, the data augmentation techniques can increase the dataset size and reduce the overfitting during the training process of the model by adding some augmented images to the training dataset. The BIM, DCGAN and NST augmentation techniques were used to produce the augmented images in the dataset. The BIM augmentation techniques consist of image cropping, flipping, PCA color augmentation, rotation and scaling. The PCA color augmentation technique alters the intensity of the color channels using the principal component of the pixels [11]. Additionally, the image cropping, flipping, rotation and scaling techniques create augmented images by changing the color and position of the input images. There are 36,541 augmented images created by the BIM augmentation technique in the dataset.
DCGANs create augmented images that resemble the training data. The DCGAN consists of two DCNN networks, such as generator DCNN and discriminator DCNN. The generator DCNN network takes a vector of random noise and up-samples it to the training data. On the other hand, the discriminator DCNN learns to classify the real and generated images [33]. The DCGAN network was trained in the graphics processing units with training epochs of 10,000 and a mini-batch size of 64. There are 32,000 augmented images created by the DCGAN augmentation technique in the dataset. NST is another image generation technique using deep learning techniques. A modified VGG19 network was used to develop the NST augmentation model in this research. The NST model was trained with 5000 epochs on the deep learning server system. The NST models require two different input images to generate an augmented output image, such as the content image and the style reference image. At first, the content image contains the essential features to be added to the output image. Second, the style reference image contains style patterns to apply to the output image. To generate the output image, the NST augmentation model applies the style features of the style image to the content image. The NST augmentation technique creates 17,500 augmented images in the dataset. Finally, The BIM, DCGAN and NST techniques were created for the augmented images to balance the data counts in each class of the dataset. The proposed dataset is named the PlantDisease59 dataset. These augmentation techniques increased the number of images in the dataset from 61,459 to 147,500 images. Additionally, the size of individual classes increased to 2500 images in each. Leaf images on the PlantDisease59 dataset were captured in the face-up direction. Figure 1 shows the sample augmented images generated by the BIM, DCGAN and NST techniques.
Figure 1. Sample augmented images using BIM, DCGAN and NST techniques.
The first two images in Figure 1 were created using BIM techniques. The third and fourth leaf images in Figure 1 were created using the DCGAN augmentation technique. The last two sample images in Figure 1 were generated using the NST technique. The random selection technique was used to select the images for training, validation and testing from the PlantDisease59 dataset. Table 3 illustrates the number of images in the training, validation and testing dataset.
Table 3. Training, validation and test dataset size.
The design and development of the Proposed DCNN model for plant leaf disease detection using the hyperparameter tuning techniques and the PlantDisease59 dataset are discussed in the following subsection.

3.2. Model Design

In this section, a DCNN model for diagnosing plant leaf diseases using the PlantDisease59 dataset is proposed. Several DCNN models with different numbers and sizes of convolutional (Conv) and pooling layers were developed, and their performance is compared. The number of Conv layers varies from three to eight in different DCNN models. At maximum, the 14-layered deep convolutional neural network (14-DCNN) gives better training performance than other developed models. Five convolutional and five max-pooling layers were used to develop the proposed 14-DCNN model. The input images of the 14-DCNN are given into the first two-dimensional Conv layer. The dimension of the Conv layer output can be calculated using Equation (1):
D i m e n s i o n ( Conv ( n , k ) ) = ( [ n w f w s + 1 ] ,   [ n h f h s + 1 ] ,   f C )
The input width ( n w ) and height ( n h ) of the first convolutional layer are 128 and 128 respectively. Additionally, the f w , f h and f c represent the width, height and channels of the kernel filter of the convolutional layer. The stride ( S ) value of this Conv layer is one. The first max-pooling layer was introduced to reduce the dimension of the first Conv layer output from 126, 126, 4 to 63, 63, 4 values. The dimension of the max-pooling layer output was calculated using Equation (2):
D i m e n s i o n ( P o o l i n g ( n , k ) ) = ( [ n w f w s + 1 ] ,   [ n h f h s + 1 ] ,   n C )
The n w ,   n h   and n c represent the width, height and channels of input n, respectively. Additionally, the   f w , f h and n c represent the width, height and channels of the filter ( f ) in the max-pooling layer. The output of the first max-pooling layer is given as an input of the second Conv layer.
Likewise, the second Conv layer uses the filter sizes of 16, 3, 3 values to extract the features from the data. The output size of the second Conv layer is 61, 61, 16 values. The second max-pooling layer reduces the output data size of the second Conv layer from 61, 61, 16 to 30, 30, 16 with the filter size of 2, 2 values. The third Conv layer was introduced after the second max-pooling layer. It extracts more features from the input data using the 32, 3, 3 size kernel and produces the 28, 28, 32 sized output data. The third max-pooling layer was used to reduce the size of the output data from the third Conv layer with 2, 2 sized kernels. It reduces the size of the data to 14, 14, 32 values from 28, 28, 32 values. Additionally, the fourth Conv layer uses the 64, 3, 3 sized kernels to extract the additional features from the third pooled data and generates the output data with the size of 12, 12, 64 values from the input size of 14, 14, 32 values. In addition, the fourth max pooling layer has a filter size of 2, 2 matrices and stride value of 1 step. The fourth max-pooling layer reduces the size from 12, 12, 64 values to 6, 6, 64 values. The fifth Conv layer was introduced with a filter size of 128, 3, 3 matrices. The size of the input matrix is 6, 6, 64 and the output size of the fifth Conv layer is 4, 4, 128 matrices. After the fifth Conv layer, the fifth pooing layer was used with max-pooling function and stride value of 1 step. The input size of the fifth max-pooling layer is 4, 4, 128 values, and the output size is 2, 2, 128 values. It uses the single-step stride values to apply the kernel to the input data from the fifth Conv layer. The ReLu activation function was used on all the above Conv layers. The ReLu activation function was performed by using Equation (3).
R e L u ( x ) = M a x ( 0 ,   x )
Moreover, the flatten layer was introduced after the fifth convolutional and pooling layer. It reduces the three-dimensional data to one-dimensional data for a traditional neural network approach. The flatten layer converts the output of the fifth max-pooling layer from 2, 2, 128 values to 512 values. The first dense layer was introduced after the flatten layer in the 14-DCNN. The first dense layer increases input value from 512 to 2048 values. Equation (4) represents the individual neuron output ( z j ) of the first dense layer. The i represents the number of inputs of the first dense layer, and it ranges from 1 to 512. Additionally, the j denotes the number of outputs of the layer its range from 1 to 2048 values.
z j = R e L u ( 0 ,   i 512 b j + x i w i )  
The x i and w i represent the value and weight of the ith input of the jth output. Additionally, the b j denotes the bias value of the jth node. The dropout layer was used between the first dense and second dense layer in the 14-DCNN to avoid the overfitting issue. The second dense layer was initiated after the first dense layer and dropout layer. Similarly, the number of inputs of the second dense layer is 2048 and the output values of the neural network is 59 values. Additionally, this layer uses the softmax activation function to classify the plant leaves. The softmax function ( σ ) value of ith neuron of the dense layer can be calculated using Equation (5).
s o f t m a x ( σ ( z i ) ) = e z i j = 1 59 e z j  
The output class of the input image can be discovered using Equation (6).
O u t p u t   C l a s s   ( z o u t ) = m a x ( z 1 ,   z 2 ,   z 59 )
This output value from z 1 to z 59 represents the number of diseases and healthy plant leaf and non-leaves classes in the PlantDisease59 dataset. The total number of training parameters is 5,424,583 in the 14-DCNN model. The layered structure of the 14-DCNN model is shown in Figure 2.
Figure 2. Layered structure of 14-DCNN model.
After designing the 14-DCNN model, the most suitable hyperparameter values were identified using hyperparameter tuning techniques. The random search and coarse-to-fine techniques were used to discover the suitable values of the optimizer function, mini-batch size and dropout probability of the 14-DCNN. The most common optimizers considered for the hyperparameter searching are adaptive moment estimation (Adam), stochastic gradient descent (SGD) and root mean square propagation (RMSprob). The range of the mini-batch size is between 8 and 256, incremented by 8 per value. Additionally, the dropout range varies between 0.0 and 0.5, incremented by 0.1 per value. The random searching technique chooses the random combination of hyperparameter values from the search space. The selected combination of hyperparameter values is applied to the 14-DCNN and trained with 100 epochs in parallel. From the training result, the coarse-to-fine process helps to identify the most possible hyperparameter values from the search space. Finally, the hyperparameter tuning technique discovers the SGD optimizer with a mini-batch size of 32 and the dropout probability of 0.2 gives a better performance than other values. To identify the learning rate (Lr) and momentum of the SGD optimizer, a similar random search approach was used with a possible combination of the values. Table 4 shows the most common hyperparameters of the 14-DCNN model for the plant leaf disease detection model with their values.
Table 4. Optimized hyperparameters of the 14-DCNN.
The random search with the coarse-to-fine technique offers significantly improved searching performance than the grid search and simple random search optimization techniques. The optimized hyperparameter values and the PlantDisease59 were used to train the proposed 14-DCNN model for diagnosing the diseases from the plant leaf images.

3.3. Model Training

The proposed 14-DCNN was trained using the optimized hyperparameters and augmented dataset in the deep learning server environment. The model was trained with different epoch values between 100 and 3000. The 1000 epochs gave the maximum validation accuracy and minimum loss. The training time of the proposed model with 1000 epochs was 7452 s in Nvidia DGX-1 deep learning server. Figure 3 demonstrates the training and validation performance of the proposed 14-DCNN model for the identification of plant leaf diseases.
Figure 3. (a) Accuracy and (b) loss of the proposed 14-DCNN model.
The 14-DCNN model achieved a training accuracy of 99.993% and a validation accuracy of 99.985%. Training and validation accuracies of the proposed 14-DCNN are higher than the other proposed DCNN models. The proposed 14-DCNN took 7452 s for the training process in the MGPUs environment. The training time of the proposed 14-DCNN was smaller than the transfer learning techniques since the number of convolutional and pooling operations of the proposed 14-DCNN are lesser than the transfer learning techniques. Finally, the architecture and weights of the proposed 14-DCNN model were stored as a hierarchical data (H5) file for the further prediction process.

3.4. Model Prediction

The saved 14-DCNN model architecture and weights were used to detect the diseases of the various plants from the input images. The real-time plant disease images were given as an input of the 14-DCNN model. The 14-DCNN model successfully predicted the plant name and disease from the input images. The matplotlib package was used to visualize the predictions of the model. Figure 4 shows the random sample prediction from the proposed 14-DCNN model.
Figure 4. Sample predicted images using the 14-DCNN model.
Finally, the 14-DCNN model was converted as a TensorFlow lite (tflite) file using a TensorFlow lite converter with a latency optimization approach. The tflite file can be used to deploy the model in mobile and embedded devices for real-time prediction.

4. Results

This section examines the performance of the proposed 14-DCNN model, using various performance evaluation approaches and testing datasets. Additionally, the performance of the proposed 14-DCNN model is compared with other state-of-the-art techniques. The state-of-the-art techniques are AlexNet [25], Inception-v3-Net, ResNet-50 and VGG16Net [7,8]. These models were trained on a deep learning server system using the PlantDisease59 dataset and tested using the testing dataset. The proposed and existing models were trained only on face-up leaf images. So, the models will give the best performance on the face-up direction positioned images.
Occlusion sensitivity can visualize the most important part of the input image for classification identified by the trained model. It can measure the sensitivity of the neural network to occlusion in different regions of the data, using small perturbations of the data. This region is known as the occluding region [7]. The white and light blue color pixels of the image illustrate the most essential parts of being classified into the expected class. The dark blue color region of the image has minimum features for the classification. Occluding stride and occluding size of the selected sensitivity map are 10 and 30, respectively. Figure 5 illustrates the occlusion sensitivity of the proposed 14-DCNN model on sample test data.
Figure 5. (a) Original image, (b) occlusion sensitivity of 14-DCNN.
The confusion matrix is a summary of predictions made by the classification techniques. The confusion matrix of the classification technique represents the true positive (TP), true negative (TN), false positive (FP) and false negative (FN) values of every single class [3]. The area under the receiver operating characteristic (AUC-ROC) curve is one of the popular metrics that is used to evaluate the performance of learning algorithms. The ROC curve plots the difference between the true positive rate (TPR) and false positive rate (FPR) [15]. The TPR and FPR are calculated using Equations (7) and (8).
TPR = TP TP + FN  
FPR = FP FP + TN
The AUC-ROC curve of the cherry healthy and strawberry leaf scorch classes shows the classification advantage of the proposed 14-DCNN. Figure 6 illustrates the AUC-ROC curve of the cherry healthy and strawberry leaf scorch classes, using the proposed 14-DCNN model.
Figure 6. Sample AUC-ROC curves of 14-DCNN.
The performance of the proposed 14-DCNN model and state-of-the-art techniques is compared, using the most common performance metrics, such as classification accuracy, precision, recall and F1-score [3]. At first, the classification accuracy is defined as the correctly classified images divided by the total number of testing images. Precision is the second most important performance evaluation metric in classification techniques. Precision is defined as the number of correctly identified results divided by the number of correctly identified and correctly rejected results that are predicted by the model. Precision is used to find the correct proportion of the classification.
The third most important performance evaluation technique is recall. The recall is the number of correctly identified results divided by the number of correctly identified and incorrectly rejected results. The recalls are used to determine the proportion of actual positives that were correctly identified. The F1 score is one of the widely used metrics for the performance evaluation of machine learning algorithms. The F1 score is defined as the harmonic mean between precision and recall. The F1 score value represents the prediction advantage of the classification techniques. The following Equations (9)–(12) were used to calculate the accuracy, weighted average precision, weighted average recall and weighted average F1 score of the classification techniques.
Classification   accuracy = TP + TN TP + TN + FP + FN
Precision = TP TP + FP
Recall = TP TP + FN
F 1   Score = 2 × ( precision   ×   recall ) ( precision   +   recall )
The range of the classification accuracy, precision, recall and F1 score is between 0 (0%) and 1 (100%). The classification performance of the proposed 14-DCNN model on individual classes in the dataset is illustrated in Table 5.
Table 5. Class-wise performance of proposed 14-DCNN model.
Figure 7 compares the performance of the proposed 14-DCNN and existing state-of-the-art classification techniques using accuracy, weighted average precision, weighted average recall and weighted average F1 score.
Figure 7. Performance comparison of proposed 14-DCNN and existing model.
The comparison results show the accuracy, precision, recall, and F1 scores of the proposed 14-DCNN are higher than the AlexNet, Inception-v3-Net, ResNet-50 and VGG16Net. Additionally, the complexity of the proposed 14-DCNN and other transfer learning techniques is illustrated in Table 6.
Table 6. Complexity comparison of proposed 14-DCNN and existing model.
The complexity analysis result shows that the number of trainable parameters and model size of the proposed 14-DCNN model are lesser than the existing transfer learning techniques. The smaller number of trainable parameters and small model size will reduce the complexity of the model prediction process. The comparison results illustrate that the performance of the proposed 14-DCNN is higher than the AlexNet, Inception-v3-Net, ResNet-50 and VGG16Net on plant leaf disease classification.

5. Conclusions

A novel DCNN model was proposed to detect plant leaf diseases from leaf images in this research. The proposed 14-DCNN model was designed and trained to detect 42 leaf diseases in 16 plants through leaf images. The data augmentation and hyperparameter optimization techniques were also used to enhance the performance of the 14-DCNN in this research. Three augmentation techniques were used to enhance the dataset size to 147,500 images. The augmentation techniques are NST, DCGAN, and BIM. The individual class size, including original and augmented images, of the dataset, was 2500 images. The 14-DCNN comprises five Conv and five max-pooling layers. The random search with the coarse-to-fine technique was used to optimize the value of the hyperparameter for training the proposed 14-DCNN model. Training of the most successful 14-DCNN model was completed with the training and validation dataset of 139,000 images and optimized hyperparameter values. The proposed 14-DCNN model achieved a classification accuracy of 99.9655%, a precision value of 99.7999%, a recall value of 99.7966%, and an F1 score of 99.7968% on the training dataset. The optimized hyperparameter values and the data augmentation techniques had a considerable influence on the results of the proposed DCNN model. Compared with standard transfer learning techniques, the proposed 14-DCNN model has higher classification performance. An extension of this research will be adding new classes of plant diseases and an increasing number of training images in the dataset and modifying the architecture of the DCNN model using more convolutional and other layers. In the future, we plan to estimate the possibility of plant disease and analyze the severity using the deep learning technique. Moreover, we will extend disease detection from plant leaves to other parts of the plants, such as flowers, fruits, and stems.

Author Contributions

Conceptualization, J.A.P. and V.D.K.; writing—original draft preparation, J.A.P. and V.D.K.; writing—review, designing, analysis and editing, J.A.P., V.D.K., O.G., M.H., M.A. and K.K.; supervision, V.D.K.; funding acquisition, M.H. All the authors contributed in writing, reviewing, and structuring of the work. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by a grant of the Romanian National Authority for Scientific Research and Innovation, CCCDI-UEFISCDI, project number 203, COFUND-ICT-AGRI-FOOD-MERIAVINO-1, within PNCDI III.

Institutional Review Board Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Pandian, J.A.; Kanchanadevi, K.; Kumar, V.D.; Jasińska, E.; Goňo, R.; Leonowicz, Z.; Jasiński, M. A Five Convolutional Layer Deep Convolutional Neural Network for Plant Leaf Disease Detection. Electronics 2022, 11, 1266. [Google Scholar] [CrossRef]
  2. Sladojevic, S.; Arsenovic, M.; Anderla, A.; Culibrk, D.; Stefanovic, D. Deep Neural Networks Based Recognition of Plant Diseases by Leaf Image Classification. Comput. Intell. Neurosci. 2016, 2016, 3289801. [Google Scholar] [CrossRef] [Green Version]
  3. Geetharamani, G.; Pandian, A. Identification of plant leaf diseases using a nine-layer deep convolutional neural network. Comput. Electr. Eng. 2019, 76, 323–338. [Google Scholar]
  4. Rumpf, T.; Mahlein, A.-K.; Steiner, U.; Oerke, E.-C.; Dehne, H.-W.; Plümer, L. Early detection and classification of plant diseases with Support Vector Machines based on hyperspectral reflectance. Comput. Electron. Agric. 2010, 74, 91–99. [Google Scholar] [CrossRef]
  5. Chen, H.-C.; Widodo, A.M.; Wisnujati, A.; Rahaman, M.; Lin, J.C.-W.; Chen, L.; Weng, C.-E. AlexNet Convolutional Neural Network for Disease Detection and Classification of Tomato Leaf. Electronics 2022, 11, 951. [Google Scholar] [CrossRef]
  6. Lee, S.H.; Chan, C.S.; Wilkin, P.; Remagnino, P. Deep-plant: Plant identification with convolutional neural networks. In Proceedings of the 2015 IEEE International Conference on Image Processing (ICIP), Quebec City, QC, Canada, 27–30 September 2015; pp. 452–456. [Google Scholar]
  7. Brahimi, M.; Arsenovic, M.; Laraba, S.; Sladojevic, S.; Boukhalfa, K.; Moussaoui, A. Deep Learning for Plant Diseases: Detection and Saliency Map Visualisation. In Human and Machine Learning: Visible, Explainable, Trustworthy and Transparent; Zhou, J., Chen, F., Eds.; Springer International Publishing: Cham, Switzerland, 2018; pp. 93–117. [Google Scholar]
  8. Ferentinos, K.P. Deep learning models for plant disease detection and diagnosis. Comput. Electron. Agric. 2018, 145, 311–318. [Google Scholar] [CrossRef]
  9. Arun Pandian, J.; Geetharamani, G. Data for: Identification of Plant Leaf Diseases Using a 9-Layer Deep Convolutional Neural Network. Mendeley Data. 2019. Available online: https://data.mendeley.com/datasets/tywbtsjrjv/1 (accessed on 29 March 2020).
  10. Almadhor, A.; Rauf, H.T.; Lali, M.I.U.; Damaševičius, R.; Alouffi, B.; Alharbi, A. AI-Driven Framework for Recognition of Guava Plant Diseases through Machine Learning from DSLR Camera Sensor Based High Resolution Imagery. Sensors 2021, 21, 3830. [Google Scholar] [CrossRef]
  11. Shijie, J.; Ping, W.; Peiyi, J.; Siping, H. Research on data augmentation for image classification based on convolution neural networks. In Proceedings of the 2017 Chinese Automation Congress (CAC), Jinan, China, 20–22 October 2017; pp. 4165–4170. [Google Scholar]
  12. Pandian, J.A.; Geetharamani, G.; Annette, B. Data augmentation on plant leaf disease image dataset using image manipulation and deep learning techniques. In Proceedings of the 2019 IEEE 9th International Conference on Advanced Computing (IACC), Tiruchirappalli, India, 13–14 December 2019; pp. 199–204. [Google Scholar]
  13. Bang, S.; Baek, F.; Park, S.; Kim, W.; Kim, H. Image augmentation to improve construction resource detection using generative adversarial networks, cut-and-paste and image transformation techniques. Autom. Constr. 2020, 115, 103198. [Google Scholar] [CrossRef]
  14. Trivedi, N.K.; Gautam, V.; Anand, A.; Aljahdali, H.M.; Villar, S.G.; Anand, D.; Goyal, N.; Kadry, S. Early Detection and Classification of Tomato Leaf Disease Using High-Performance Deep Neural Network. Sensors 2021, 21, 7987. [Google Scholar] [CrossRef]
  15. Agarwal, M.; Gupta, S.K.; Biswas, K.K. Development of Efficient CNN model for Tomato crop disease identification. Sustain. Comput. Inform. Syst. 2020, 28, 100407. [Google Scholar] [CrossRef]
  16. Ebrahimi, M.A.; Khoshtaghaza, M.H.; Minaei, S.; Jamshidi, B. Vision-based pest detection based on SVM classification method. Comput. Electron. Agric. 2017, 137, 52–58. [Google Scholar] [CrossRef]
  17. Wetterich, C.B.; Neves, R.F.D.O.; Belasque, J.; Ehsani, R.; Marcassa, L.G. Detection of Huanglongbing in Florida using fluorescence imaging spectroscopy and machine-learning methods. Appl. Opt. 2017, 56, 15–23. [Google Scholar] [CrossRef]
  18. Mokhtar, U.; Ali, M.A.; Hassanien, A.E.; Hefny, H. Identifying Two of Tomatoes Leaf Viruses Using Support Vector Machine. In Information Systems Design and Intelligent Applications; Mandal, J.K., Satapathy, S.C., Sanyal, M.K., Sarkar, P.P., Mukhopadhyay, A., Eds.; Springer India: New Delhi, India, 2015; pp. 771–782. [Google Scholar]
  19. Bharate, A.A.; Shirdhonkar, M.S. A review on plant disease detection using image processing. In Proceedings of the 2017 International Conference on Intelligent Sustainable Systems (ICISS), Palladam, India, 7–8 December 2017; pp. 103–109. [Google Scholar]
  20. Grinblat, G.L.; Uzal, L.C.; Larese, M.G.; Granitto, P.M. Deep learning for plant identification using vein morphological patterns. Comput. Electron. Agric. 2016, 127, 418–424. [Google Scholar] [CrossRef] [Green Version]
  21. Fuentes, A.; Yoon, S.; Kim, S.C.; Park, D.S. A Robust Deep-Learning-Based Detector for Real-Time Tomato Plant Diseases and Pests Recognition. Sensors 2017, 17, 2022. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  22. DeChant, C.; Wiesner-Hanks, T.; Chen, S.; Stewart, E.L.; Yosinski, J.; Gore, M.A.; Nelson, R.J.; Lipson, H. Automated Identification of Northern Leaf Blight-Infected Maize Plants from Field Imagery Using Deep Learning. Phytopathology 2017, 107, 1426–1432. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  23. Johannes, A.; Picon, A.; Alvarez-Gila, A.; Echazarra, J.; Rodriguez-Vaamonde, S.; Navajas, A.D.; Ortiz-Barredo, A. Automatic plant disease diagnosis using mobile capture devices, applied on a wheat use case. Comput. Electron. Agric. 2017, 138, 200–209. [Google Scholar] [CrossRef]
  24. Kawasaki, Y.; Uga, H.; Kagiwada, S.; Iyatomi, H. Basic study of automated diagnosis of viral plant diseases using convolutional neural networks. In Advances in Visual Computing; Bebis, G., Boyle, R., Parvin, B., Koracin, D., McMahan, R., Jerald, J., Eds.; Springer International Publishing: Cham, Switzerland, 2015; pp. 638–645. [Google Scholar]
  25. Nachtigall, L.G.; Araujo, R.M.; Nachtigall, G.R. Classification of apple tree disorders using convolutional neural networks. In Proceedings of the 2016 IEEE 28th International Conference on Tools with Artificial Intelligence (ICTAI), San Jose, CA, USA, 6–8 November 2016; pp. 472–476. [Google Scholar]
  26. Rangarajan, A.K.; Purushothaman, R.; Ramesh, A. Tomato crop disease classification using pre-trained deep learning algorithm. Procedia Comput. Sci. 2018, 133, 1040–1047. [Google Scholar] [CrossRef]
  27. Zhang, X.; Qiao, Y.; Meng, F.; Fan, C.; Zhang, M. Identification of Maize Leaf Diseases Using Improved Deep Convolutional Neural Networks. IEEE Access 2018, 6, 30370–30377. [Google Scholar] [CrossRef]
  28. Brahimi, M.; Boukhalfa, K.; Moussaoui, A. Deep Learning for Tomato Diseases: Classification and Symptoms Visualization. Appl. Artif. Intell. 2017, 31, 299–315. [Google Scholar] [CrossRef]
  29. Lu, Y.; Yi, S.; Zeng, N.; Liu, Y.; Zhang, Y. Identification of rice diseases using deep convolutional neural networks. Neurocomputing 2017, 267, 378–384. [Google Scholar] [CrossRef]
  30. Mohanty, S.P.; Hughes, D.P.; Salathé, M. Using Deep Learning for Image-Based Plant Disease Detection. Front. Plant Sci. 2016, 7, 1419. [Google Scholar] [CrossRef] [Green Version]
  31. Arnal Barbedo, J.G. Plant disease identification from individual lesions and spots using deep learning. Biosyst. Eng. 2019, 180, 96–107. [Google Scholar] [CrossRef]
  32. Shorten, C.; Khoshgoftaar, T.M. A survey on Image Data Augmentation for Deep Learning. J. Big Data 2019, 6, 60. [Google Scholar] [CrossRef]
  33. Lu, C.-Y.; Rustia, D.J.A.; Lin, T.-T. Generative Adversarial Network Based Image Augmentation for Insect Pest Classification Enhancement. IFAC-PapersOnLine 2019, 52, 1–5. [Google Scholar] [CrossRef]
  34. Saranyaraj, D.; Manikandan, M.; Maheswari, S. A deep convolutional neural network for the early detection of breast carcinoma with respect to hyper- parameter tuning. Multimed. Tools Appl. 2020, 79, 11013–11038. [Google Scholar]
  35. Hu, G.; Wu, H.; Zhang, Y.; Wan, M. Data for: A Low Shot Learning Method for Tea Leaf’s Disease Identification. Mendeley Data. 2019. Available online: https://data.mendeley.com/datasets/dbjyfkn6jr/1 (accessed on 29 March 2020).
  36. Kour, V.P.; Arora, S. Plantaek: A Leaf Database of Native Plants of Jammu and Kashmir. Mendeley Data. 2019. Available online: https://data.mendeley.com/datasets/t6j2h22jpx/2 (accessed on 29 March 2020).
  37. Krohling, R.A.; Esgario, J.; Ventura, J.A. BRACOL—A Brazilian Arabica Coffee Leaf Images Dataset to Identification and Quantification of Coffee Diseases and Pests. Mendeley Data. 2019. Available online: https://data.mendeley.com/datasets/yy2k5y8mxg/1 (accessed on 29 March 2020).
  38. Parraga-Alava, J.; Cusme, K.; Loor, A.; Santander, E. RoCoLe: A Robusta Coffee Leaf Images Dataset. Mendeley Data. 2019. Available online: https://data.mendeley.com/datasets/c5yvn32dzg/2 (accessed on 29 March 2020).
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.