Next Article in Journal
Factors Influencing Genomic Prediction Accuracies of Tropical Maize Resistance to Fall Armyworm and Weevils
Previous Article in Journal
Transmission of Engineered Plastids in Sugarcane, a C4 Monocotyledonous Plant, Reveals that Sorting of Preprogrammed Progenitor Cells Produce Heteroplasmy
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Convolutional Neural Network for Automatic Identification of Plant Diseases with Limited Data

1
Department of Computer Science, College of Computer Science and Information Technology, King Faisal University, P.O. Box 400, Al-Ahsa 31982, Saudi Arabia
2
Faculty of Computers and Information, Menoufia University, Shibin Al Kawm 32511, Menoufia, Egypt
*
Author to whom correspondence should be addressed.
Plants 2021, 10(1), 28; https://doi.org/10.3390/plants10010028
Submission received: 2 December 2020 / Revised: 19 December 2020 / Accepted: 21 December 2020 / Published: 24 December 2020
(This article belongs to the Section Plant Protection and Biotic Interactions)

Abstract

:
Automated identification of plant diseases is very important for crop protection. Most automated approaches aim to build classification models based on leaf or fruit images. These approaches usually require the collection and annotation of many images, which is difficult and costly process especially in the case of new or rare diseases. Therefore, in this study, we developed and evaluated several methods for identifying plant diseases with little data. Convolutional Neural Networks (CNNs) are used due to their superior ability to transfer learning. Three CNN architectures (ResNet18, ResNet34, and ResNet50) were used to build two baseline models, a Triplet network and a deep adversarial Metric Learning (DAML) approach. These approaches were trained from a large source domain dataset and then tuned to identify new diseases from few images, ranging from 5 to 50 images per disease. The proposed approaches were also evaluated in the case of identifying the disease and plant species together or only if the disease was identified, regardless of the affected plant. The evaluation results demonstrated that a baseline model trained with a large set of source field images can be adapted to classify new diseases from a small number of images. It can also take advantage of the availability of a larger number of images. In addition, by comparing it with metric learning methods, we found that baseline model has better transferability when the source domain images differ from the target domain images significantly or are captured in different conditions. It achieved an accuracy of 99 % when the shift from source domain to target domain was small and 81 % when that shift was large and outperformed all other competitive approaches.

1. Introduction

Since plant diseases may cause crop damage, they pose a major threat to food security and sustainability and may lead to food shortages. Therefore, early detection and control of plant diseases is very important. This process often requires trained human expertise to make the correct diagnosis. However, this expertise is not always available, especially in remote locations and small farms in developing counties. With the advancement of consumer devices, such as smartphones, that can capture high-quality images, the development of an effective image-based diagnostic system can greatly contribute to early disease diagnosis and waste reduction [1]. In the literature, several studies have been proposed to diagnose plant diseases using computer vision, machine learning, or deep learning. In this study we are interested in deep learning-based approaches that have proven effective in many visual recognition tasks.
A number of the techniques that have been proposed focus on diagnosing diseases of a single crop. Esgario et al. [2] proposed an approach to classify and quantify the biotic stress of coffee leaves. They used several Convolutional Neural Network (CNN) models to diagnose and quantify the severity of four coffee plant diseases. They also applied various data augmentation techniques to achieve high diagnostic results. In [3], a deep learning based algorithm for image-based detection of cassava plant disease has been proposed. The authors applied the transfer learning of a pretrained CNN model to train a linear classifier, support vector machine (SVM), and a nearest neighbor (KNN) classifier using a dataset of cassava disease images captured in the field. Their best model achieved an average accuracy of 93 % for three diseases and two types of pest damage. A similar approach has also been proposed for detection and diagnosis of cassava plant diseases in [4]. The tomato crop has attracted the interest of many researchers and several techniques have been proposed to diagnose its diseases [5,6,7]. An approach for diagnosing corn leaf diseases has been proposed in [8]. A dataset collected with a smartphone camera was used to train a custom CNN model and a good accuracy was obtained using a limited number of test images. Likewise, approaches have been proposed using CNNs trained from scratch or using transfer learning to classify apple plant diseases [9,10], grape plant diseases [11], and potato plant diseases [12].
Another class of methods aims to provide general solutions that can be used to diagnose several diseases simultaneously. Too et al. [13] performed a comparative study to fine-tune several CNN models for identifying plant diseases. The results of this study using the PlantVillage dataset [14] indicated that DenseNet121 [15] outperformed all other models and achieved an accuracy score of 99.75 % . P D 2 S E N e t was proposed in [16], based on ResNet50 [17], to identify plant species, diagnose diseases, and estimate disease severity. This approach achieved an accuracy score of 98 % for plant disease classification. Many similar methods of classifying multicrop diseases have been also proposed in the literature as presented in [18,19]. Lee et al. [20] approached the problem differently, by comparing a number of techniques for classifying plant diseases based on the common disease name. This formulation allowed them to build a more generalizable model. In this study, we will also evaluate this formulation in the case of learning from little data.
If we look at the literature, we can realize the effectiveness of deep learning methods for diagnosing plant diseases. However, on the other hand, the success of these techniques depends greatly on the availability of a good amount of data to train the models, and the lack of data may hinder the development of these techniques. This data collection and labeling are often challenging, either because of the novelty of the disease or because of the high cost or lack of appropriate expertise. Therefore, the focus is on learning techniques that can learn from small datasets, formally called few-shot learning (FSL) [21,22,23]. There is a variety of FSL learning approaches that can be organized into four main categories: metalearning, metric learning, data augmentation, and transfer learning. Algorithms in metalearning category learn a learning strategy or a good model initialization to adjust well to a novel, few-shot learning task [24,25,26]. The metric learning approaches learn a semantic embedding space using a distance loss function. Accordingly, they map images in a space where similar classes are close together while different classes are further apart. Metric learning is an active area of research in which several algorithms have been proposed such as Siamese network [27], Triplet network [28], matching network [29], prototypical network [30], and relation network [31]. The data augmentation approaches generate more data from novel classes to facilitate the regular learning [32,33]. This area of research is under active development and there are a number of studies that have focused on comparing different FSL methods. Chen et al. [23] conducted a study to compare several metalearning, metric learning, and transfer learning methods using natural images. In another recent study [34], few-shot learning methods were compared when there is a shift between source and target domains. These studies showed that the performance of few-shot learning approaches may vary depending on the source of the images and the problem formulation. It also indicated that few-shot learning techniques based on metalearning poorly perform when there is a significant shift between source and target domains.
Recently, some FSL-based approaches have been proposed for identifying plant species and diagnosing crop diseases. Hu et al. [35] generate synthetic leaf images to facilitate tea leaves disease identification. The Siamese network was used in [36] to build a classification approach for diagnosing citrus diseases and in [37] to build a plant leaf classification approach. In [38], Argueso et al. introduced an enhanced approach to classify plant leaf diseases through the use of Triplet loss [28] and support vector machines (SVM) [39]. They created the source and target domains datasets from the PlantVillage [14] dataset and fine-tuned the last fifty layers of the Inception V3.0 model. Their results using the Triplet loss to create an embedding space and SVM classifier for FSL outperformed the fine-tuned model. However, these results were influenced by their fine-tuning setup because it is difficult to optimize a large number of parameters with just a few images.
Since selection of an appropriate learning strategy and models, as well as the right formulation of the problem, are key factors to the success of the FSL approach, the main purpose of this study is to provide an effective plant disease diagnosis approach that can learn from little data. The contributions of this study can be summarized as follows:
  • Development and evaluation of several FSL approaches to classify plant diseases based on Triplet network, Deep Adversarial Metric Learning (DAML) [40] and transfer learning using both linear and cosine-similarity classifiers [23]. All approaches were evaluated using images captured under the same as well as different conditions. We have found that fine-tuning of a pretrained model outperforms all other approaches. Figure 1 shows an overview of the approaches developed in this study.
  • We investigate the impact of model complexity on the performance of the FSL approaches, ResNet18, ResNet34, and ResNet50 were used in this study. The results showed that fine-tuning using linear classifiers benefits from increased model complexity while this cannot be confirmed for other approaches.
  • We examined two different formulations in this study. In the first, we classify plants and diseases together, while in the second we focus on diseases only. The second formulation helped the model to achieve better results in identifying new diseases and to better adapt to images captured under different conditions.

2. Materials and Methods

2.1. Datasets

Two datasets were used for training and evaluating all few-shot learning algorithms in this study, namely the PlantVillage dataset [14] and the coffee leaf dataset [2]. PlantVillage dataset has 54,305 leaf images of 14 crop species and 26 diseases distributed among 38 crop-disease pairs. This dataset contains clear images of plant leaves and each image contains only one leaf. It also comes with preset training/testing subsets which we follow in this study. Table 1 gives a summary of these classes and the number of images in each class. We use this dataset in two different ways, first to classify disease-crop pairs and second to classify disease regardless of the affected crop. In the first configuration, 32 classes (C7 to C38) are used as the source domain D s and 6 classes (C1 to C6) as the target domain D t similar to [38]. In this case the source domain has 43,444 samples while the target/novel domain has 10,861 samples. In the second configuration, we rearrange the dataset according to the common disease name as in [20], resulting in 20 disease classes and one healthy class. For diseases that affect more than one plant, all samples are combined in one class under the name of this disease, and for the healthy class, about 5000 samples were collected from all available plants. The summery of the rearranged dataset is shown in Table 2. The three diseases with the fewest images (CD1, CD4, and CD19) were selected from the rearranged dataset as target (novel) domain classes and the rest as the source domain classes. The number of samples in the source domain D s is 44,081 and in the target domain D t is 1278.
The coffee leaves Dataset [2] contains 1747 Images of Arabica coffee leaves captured using different mobile phones. It contains healthy leaves as well as leaves affected by one or more diseases (leaf miner, rust, brown leaf spot, cercospora leaf spot). Each leaf is labeled with the predominant disease. In this study, we did not use this dataset for training and all classes were used as target/novel domain classes. This configuration allows us to fairly evaluate the performance of the developed models in a more realistic situation, when the novel dataset is captured under different condition. The summary of this dataset’s classes is presented in Table 3. Samples from PlantVillage and coffee leaf datasets are shown in Figure 2 and Figure 3, respectively.

2.2. Learning Approaches

In this study, we aim to develop an approach that can learn from little data. This problem in the community is known as few-shot learning (FSL), which can leverage the information learned from a large source domain dataset, D s , to build a model that can classify novel classes from target domain, D t , using few samples. In FSL, the target domain has a label space that differs from the source domain and it is formally defined as C w a y / K s h o t s classification problem; Classify C novel classes using K sample from each class. The idea behind all FSL algorithms is to build a generic feature extractor or embedder, f, to map the image X i to a low-dimensional feature vector f i = f ( X i , θ ) , θ is the embedder parameters. This embedding trained from source data, D s , should be general enough to be used for classifying new classes in the target domain, D t . Since the classification of plant diseases differs from the classification of digits or general objects, we will develop several classification models using little data and compare them to build our final model. Details of learning algorithms and methodologies will be presented in the following subsections.

2.3. Transfer Learning: Baseline and Baseline++

In transfer learning, a CNN model is trained using the source domain dataset and then fine-tuned using few samples from the target domain dataset as shown in Figure 4. In the training phase, both the feature extractor f and the classifier G s are trained from scratch while in the fine-tuning phase, the feature extractor is fixed and a new classifier G t is trained using a small target domain dataset. Similar to [23], in this study, we use two baseline models. The first uses a linear classifier and we refer to it as Baseline and the other uses a cosine-similarity based classifier and we refer to it as Baseline++.
The Baseline model uses a linear classifier G s ( . | W s ) , W s R d × c is the weight matrix, d is the dimension of the feature vector and c is the number of classes. This classifier has a linear layer followed by a Softmax function σ as defined in Equation (1).
Y ˜ = σ ( W s T f X i , θ )
In the Baseline++ classifier, the weight matrix W s consists of c, d-dimensional weight vectors [ w s 1 , w s 2 , , w s c ] . Each weight vector, w s j , can be considered as a prototype for one class. During the training process, the cosine similarity c s i j between the feature vector f i and the weight vector w s j is calculated as in Equation (2) and the final classification probability is obtained by normalizing the similarity vector, [ c s i 1 , c s i 2 , , c s i c ] using a Softmax function. In this study, we use cross-entropy loss to train and fine-tune both the Baseline and Baseline++ models.
c s i j = f ( X i , θ ) T w s j f ( X i , θ ) w s j

2.4. Metric Learning Using Triplet Network

The Triplet network [28] uses three instances of the same feature embedder with shared weights to embed the input triplet which contains an anchor sample X i , a positive sample X i + and a negative sample X i . Anchor and positive samples are sampled from the same class while the negative sample is sampled from a different class. This network is trained from scratch using the source domain dataset, D s , by minimizing the triplet loss function [41] defined in Equation (3). The structure of the Triplet network is shown in Figure 5. During the training phase, the network reduces the distance between the anchor and its positive pair while increasing the distance to the negative sample.
L t ( f i , f i + , f i ) = m a x 0 , D ( f i , f i + ) 2 D ( f i , f i ) 2 + m ,
where, f i , f i + and f i are the feature vectors corresponding to X i , X i + and X i , respectively. D ( . , . ) is the Euclidean distance and m is the margin.
After training from source domain dataset, to classify plant diseases with little data, a metric learning approach based on [22] is utilized. In this approach, the trained embedder is used to extract the features of few target domain samples, D t , and a multiclass Support Vector Machine (SVM) [39] classifier is used for final classification. SVM solves the classification problem by finding a set of hyper planes in the d dimensional space that separates samples from different classes. It tries to find the planes with the maximum margin, or distance, from data points on both sides. In this study, one versus all multiclass SVM classifier with linear kernel was used.

2.5. Deep Adversarial Metric Learning (DAML)

DAML [40] attempts to improve traditional metric learning (using triplet loss) by generating synthetic hard negatives from easy negatives. This allows the learning algorithm to take advantage of the large number of easy negatives and increases the diversity and representation of negative samples near the margin. It jointly trains a hard negative generator and a distance metric by minimizing the loss function defined in Equation (4).
J ( f i , f i + , f i ) = J g e n ( f i , f i + , f i ) + λ J m ( f i , f i + , f i ) ,
where f i , f i + , f i are the feature vectors corresponding to the anchor, positive and negative samples and λ parameter represents the balance between the metric loss J m and the adversarial loss J g e n . The generator receives triplets ( f i , f i + , f i ) and generates hard negatives f ˜ i and it is trained by minimizing the objective function defined in Equation (5). Accordingly, the generator aims to create negative images close to the anchor image and similar to other negatives and at the same time may fool the metric- learning. The triplet loss defined in Equation (6) is utilized here for metric learning. In this study, three-layer fully connected network was used as a feature generator and the input to the generator is a combination of anchor, positive and negative samples and the output is a synthetic hard negative.
J g e n ( f i , f i + , f i ) =   J h a r d + λ 1 J r e g + λ 2 J a d v =   f ˜ i f i 2 2 +   λ 1 f ˜ i f i 2 2 + λ 2 m a x 0 , D ( f i , f ˜ i ) 2 D ( f i , f i + ) 2 α ,
λ 1 and λ 2 are two balancing parameters.
J m ( f i , f i + , f i ) = m a x 0 , D ( f i , f i + ) 2 D ( f i , f ˜ i ) 2 + m ,
Similar to the metric-learning using Triplet network, feature embedder and generator were trained with the source domain data and a multiclass SVM was used for FSL in the target domain.

2.6. Deep Architectures and Experimental Setup

In all experiments, we used three CNN models having different complexities, namely, ResNet18, ResNet34, and ResNet50 [17]. We use models of the same architecture (Deep residual learning) but with different complexity to study the impact of model complexity on the classification accuracy. The main idea behind deep residual learning is to use identity shortcut connections to force each network block to fit a residual mapping H ( X ) = F ( X ) + X , where X is the input feature and F ( X ) is the output of this block. This formulation, as demonstrated by the original authors, can solve the problem of vanishing/exploding gradient and alleviate accuracy degradation in deep networks. All residual architectures share the same structure but have different number of blocks and layers. ResNet18 has about 12 million trainable parameters and an output size of 512, ResNet34 has about 21 million trainable parameters and an output size of 512 and ResNet50 has about 26 million trainable parameters and an output size of 2048. For metric learning, we use a fully connected layer with ReLU activation function to map the backbone output to a 128-dimensional vector.
All models were trained using the entire source domain dataset, after that each model was fine-tuned using a small set of target domain images ranging from 5 to 50 in increments of 5. Baseline and Baseline + + models were trained in source domain using Adam optimizer [42] with a learning rate of 0.0001 for 30 epochs and the best model was used for the fine-tuning phase. For metric learning using triplet loss, a margin of 1 was used in the loss function and the model was trained using Adam optimizer with a learning rate of 0.0001 for 15 epochs. The Deep Adversarial Metric Learning used Adam optimizer for backbone model and generator optimization with a learning rate of 0.0001 and a margin of 1 for the triplet loss. The training dataset was augmented using random horizontal and vertical flipping, rotation in a range of 0 to 30, color change (hue, saturation, brightness and contrast) by a factor of 0.1 and center cropping.
In target domain, the baseline and baseline++ models were fine-tuned using Stochastic Gradient Decent (SGD) optimizer with a learning rate of 0.01, momentum of 0.9, and weigh decay of 0.001 for 100 epochs. SVM with linear kernel was used for triplet loss and DAML methods and the soft margin parameter was set to C = 1 . All experiments were repeated 100 times and the average accuracy was calculated using a query set of 50 images per class. All experiments were conducted using PyTorch [43] deep learning framework.

3. Results and Discussion

In this study, the performance of four learning approaches and three backbone models for classifying crop diseases from limited data was evaluated. We use two different formulations; one considers both the classification of crops and diseases as a single task as [38] and the other focuses on the classification of diseases only using the common disease name as recommended by [20]. The results of all experiments will be presented in this section.

3.1. Crop and Disease Classification

In this set of experiments, 32 classes of the PlantVillage dataset [14] were used as the source domain data, D s , as detailed previously. Two datasets with different characteristics were used as the target domain data. The first one contains six different classes of the PlantVillage dataset and the other is the coffee leaf dataset [2] which contains six classes of coffee plant diseases and was collected under different conditions from the source domain dataset. Figure 6 shows the results for the first target domain dataset. This figure clearly indicates that the Baseline model outperforms all other learning approaches with all backbone models and for all few-shot settings. It achieved the lowest error rate of 2.48 % for 50 shots and 14.42 % for five shots with the ResNet50 backbone model slightly outperforming the ResNet34 backbone model. DAML comes in the second place, with a lower error rate than Triplet network for most of the few-shot settings especially when the number of samples is very low. Both approaches work better with the ResNet34, which is slightly outperformed the ResNet50. The Baseline++ lags behind other approaches significantly. The results of these experiments are summarized in Table 4.
The results of the coffee leaf dataset are shown in Figure 7 and Table 5. These results are consistent with the previous results, as the Baseline model achieved the lowest error rate of 20.22 for 50 shots and 35.54 for five shots with the ResNet50 backbone model and it outperformed all other approaches. DAML, with ResNet50, is second, while Baseline++ is the last. This figure also indicates that the large shift between the source domain and the target domain data leads to some degradation in performance.

3.2. Disease Classification Based on Common Disease Name

Since the same disease may affect more than one crop, the focus of these experiments will be on classifying diseases regardless of which plant is affected. This formulation is more realistic in the real world, where the farmer knows which crops he is growing but wants to learn about the diseases that afflict them. Eighteen classes of the rearranged PlantVillage dataset, as in Table 2, were used as the source domain dataset. Three new diseases from the PlantVillage dataset were used as the first target domain dataset and coffee leaf dataset as the second target domain dataset.
The results shown in Figure 8 for the first target domain dataset (based on disease name) indicate that the baseline model outperformed all other approaches. It achieved a very low error rate even with a very few shots. The DAML and Triplet network exchange are second place based on the number of samples and the backbone model, but they fall behind the baseline model by a large margin, especially in the case of very few shots. The summary of these results is shown in Table 6. For the coffee leaf dataset, the classification results based on the disease name are shown in Figure 9. Here, we can see the consistent performance of the baseline model and the DAML approach. Baseline model achieved the lowest error rate of 19.44 for 50 shots and 34.28 for 5 shots while DAML achieved an error rate of 22.2 for 50 shots and 42.6 for 5 shots. Table 7 shows the classification results for this dataset.
The results presented in this study show that we can use a good baseline model trained from a large source domain dataset to build a model that can learn from little data. Compared to the other approach, the baseline model can take advantage of the complexity of the backbone model to build a good representation of the features. Choosing the right backbone model and learning strategy is also important. Formulating the problem based on the name of the common disease can lead to better and more consistent results and enhances generalization of the model.
In a previous study [38], Argueso et al. indicated that metric learning using triplet loss outperformed the transfer learning strategy by a large margin. However, in their study they have tuned 50 Inception V3 [44] layers which cannot be achieved with a few samples and we believe that this is the main reason this model fails. Table 8 shows the comparison of the baseline model, the triple network with the ResNet50 backbone that was used in this study and the work of Argueso et al. [38]. This comparison shows that although a ResNet50 backbone trained from scratch using source domain data enhances Triplet network performance, the baseline model outperforms all other approaches. This is due to the fact that many diseases affect a small portion of the plant leaf, which hinders accurate representation of these diseases in the embedding space of the metric learning techniques. However, the baseline model directly extracts the features that allow it to distinguish different diseases, especially in the second formulation.

4. Conclusions

In this study, we developed several approaches for classifying plant diseases that can learn from little data. Transfer learning, Triplet network, and Deep Adversarial Metric learning (DAML) were used to build these approaches. The evaluation of these approaches demonstrated the efficiency of transfer learning using a good baseline model. It achieved a very high accuracy of 99 % for new classes when the source and target domain data are captured under the same condition and a reasonable accuracy of 81 % for novel dataset that is captured under different conditions. It can generalize well and beat all competitive approaches. We also found that DAML can enhance the traditional metric learning by generating hard samples and increases the data diversity. Selection of the appropriate model and learning strategy is also essential to the success of the FSL approach. Therefore, developing a disease classification method regardless of the plant can perform better and it is more appropriate for real applications than the usual method, which is to classify both disease and plant type together. Finally, we believe that focusing only on the affected part of the plant might lead to better classification results, especially for diseases that affect a small portion of the leaf.

Author Contributions

Conceptualization, A.A. (Ahmed Afifi) and A.A. (Amira Abdelwahab); methodology, A.A. (Ahmed Afifi); software, A.A. (Ahmed Afifi); validation, A.A. (Ahmed Afifi) and A.A. (Abdulaziz Alhumam); formal analysis, A.A. (Ahmed Afifi); investigation, A.A. (Amira Abdelwahab); resources, A.A. (Ahmed Afifi); data curation, A.A. (Ahmed Afifi) and A.A. (Amira Abdelwahab); writing—original draft preparation, A.A. (Ahmed Afifi); writing—review and editing, A.A. (Amira Abdelwahab) and A.A. (Abdulaziz Alhumam); visualization, A.A. (Ahmed Afifi); supervision, A.A. (Abdulaziz Alhumam); project administration, A.A. (Ahmed Afifi); funding acquisition, A.A. (Ahmed Afifi) and A.A. (Abdulaziz Alhumam). All authors have read and agreed to the published version of the manuscript.

Funding

The authors extend their appreciation to the Deputyship for Research and Innovation, Ministry of Education in Saudi Arabia for funding this research work through the project number IFT20139.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data is available from its original source as cited in the article. The PlantVillage dataset is available at https://github.com/spMohanty/PlantVillage-Dataset and the Coffee dataset at https://github.com/esgario/lara2018/.

Acknowledgments

The authors would like to thank the Deanship of Scientific Research and the College of Computer Science and Information Technology at King Faisal University for their continuous cooperation and support

Conflicts of Interest

The authors declare no conflict of interest. The funding sponsors had no role in the design of the study, in the collection, analyses, or interpretation of data, in the writing of the manuscript, or in the decision to publish the results.

References

  1. Johannes, A.; Picon, A.; Alvarez-Gila, A.; Echazarra, J.; Rodriguez-Vaamonde, S.; Navajas, A.D.; Ortiz-Barredo, A. Automatic plant disease diagnosis using mobile capture devices, applied on a wheat use case. Comput. Electron. Agric. 2017, 138, 200–209. [Google Scholar] [CrossRef]
  2. Esgario, J.G.; Krohling, R.A.; Ventura, J.A. Deep learning for classification and severity estimation of coffee leaf biotic stress. Comput. Electron. Agric. 2020, 169. [Google Scholar] [CrossRef] [Green Version]
  3. Ramcharan, A.; Baranowski, K.; McCloskey, P.; Ahmed, B.; Legg, J.; Hughes, D.P. Deep learning for image-based cassava disease detection. Front. Plant Sci. 2017, 8. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  4. Sambasivam, G.; Opiyo, G.D. A predictive machine learning application in agriculture: Cassava disease detection and classification with imbalanced dataset using convolutional neural networks. Egypt. Inform. J. 2020. [Google Scholar] [CrossRef]
  5. Fuentes, A.F.; Yoon, S.; Lee, J.; Park, D.S. High-performance deep neural network-based tomato plant diseases and pests diagnosis system with refinement filter bank. Front. Plant Sci. 2018, 9. [Google Scholar] [CrossRef] [Green Version]
  6. Wspanialy, P.; Moussa, M. A detection and severity estimation system for generic diseases of tomato greenhouse plants. Comput. Electron. Agric. 2020, 178. [Google Scholar] [CrossRef]
  7. Fuentes, A.; Yoon, S.; Kim, S.C.; Park, D.S. A robust deep-learning-based detector for real-time tomato plant diseases and pests recognition. Sensors 2017, 17, 2022. [Google Scholar] [CrossRef] [Green Version]
  8. Sibiya, M.; Sumbwanyambe, M. A Computational Procedure for the Recognition and Classification of Maize Leaf Diseases Out of Healthy Leaves Using Convolutional Neural Networks. AgriEngineering 2019, 1, 119–131. [Google Scholar] [CrossRef] [Green Version]
  9. Liu, B.; Zhang, Y.; He, D.J.; Li, Y. Identification of apple leaf diseases based on deep convolutional neural networks. Symmetry 2018, 10, 11. [Google Scholar] [CrossRef] [Green Version]
  10. Thapa, R.; Snavely, N.; Belongie, S.; Khan, A. The plant pathology 2020 challenge dataset to classify foliar disease of apples. arXiv 2020, arXiv:2004.11958. [Google Scholar]
  11. Liu, B.; Ding, Z.; Tian, L.; He, D.; Li, S.; Wang, H. Grape Leaf Disease Identification Using Improved Deep Convolutional Neural Networks. Front. Plant Sci. 2020, 11. [Google Scholar] [CrossRef]
  12. Tiwari, D.; Ashish, M.; Gangwar, N.; Sharma, A.; Patel, S.; Bhardwaj, S. Potato Leaf Diseases Detection Using Deep Learning. In Proceedings of the International Conference on Intelligent Computing and Control Systems (ICICCS 2020), Madurai, India, 13–15 May 2020; pp. 461–466. [Google Scholar] [CrossRef]
  13. Too, E.C.; Yujian, L.; Njuki, S.; Yingchun, L. A comparative study of fine-tuning deep learning models for plant disease identification. Comput. Electron. Agric. 2019, 161, 272–279. [Google Scholar] [CrossRef]
  14. Hughes, D.P.; Salathe, M. An open access repository of images on plant health to enable the development of mobile disease diagnostics. arXiv 2015, arXiv:1511.08060. [Google Scholar]
  15. Huang, G.; Liu, Z.; Van Der Maaten, L.; Weinberger, K.Q. Densely connected convolutional networks. In Proceedings of the 30th IEEE Conference on Computer Vision and Pattern Recognition (CVPR 2017), Honolulu, HI, USA, 21–26 July 2017; pp. 2261–2269. [Google Scholar] [CrossRef] [Green Version]
  16. Liang, Q.; Xiang, S.; Hu, Y.; Coppola, G.; Zhang, D.; Sun, W. PD 2 SE-Net: Computer-assisted plant disease diagnosis and severity estimation network. Comput. Electron. Agric. 2019, 157, 518–529. [Google Scholar] [CrossRef]
  17. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar] [CrossRef] [Green Version]
  18. Saleem, M.H.; Potgieter, J.; Arif, K.M. Plant disease detection and classification by deep learning. Plants 2019, 8, 468. [Google Scholar] [CrossRef] [Green Version]
  19. Hasan, R.I.; Yusuf, S.M.; Alzubaidi, L. Review of the state of the art of deep learning for plant diseases: A broad analysis and discussion. Plants 2020, 9, 1302. [Google Scholar] [CrossRef]
  20. Lee, S.H.; Goëau, H.; Bonnet, P.; Joly, A. New perspectives on plant disease characterization based on deep learning. Comput. Electron. Agric. 2020, 170. [Google Scholar] [CrossRef]
  21. Wang, Y.; Yao, Q.; Kwok, J.T.; Ni, L.M. Generalizing from a Few Examples: A Survey on Few-shot Learning. ACM Comput. Surv. 2020, 53. [Google Scholar] [CrossRef]
  22. Medela, A.; Picon, A.; Saratxaga, C.L.; Belar, O.; Cabezon, V.; Cicchi, R.; Bilbao, R.; Glover, B. Few shot learning in histopathological images: Reducing the need of labeled data on biological datasets. In Proceedings of the International Symposium on Biomedical Imaging (ISBI 2019), Venice, Italy, 8–11 April 2019; pp. 1860–1864. [Google Scholar] [CrossRef] [Green Version]
  23. Chen, W.Y.; Wang, Y.C.F.; Liu, Y.C.; Kira, Z.; Huang, J.B. A closer look at few-shot classification. In Proceedings of the 7th International Conference on Learning Representations (ICLR 2019), New Orleans, LA, USA, 6–9 May 2019. [Google Scholar]
  24. Finn, C.; Abbeel, P.; Levine, S. Model-agnostic meta-learning for fast adaptation of deep networks. In Proceedings of the 34th International Conference on Machine Learning (ICML 2017), Sydney, Australia, 6–11 August 2017; Volume 3, pp. 1856–1868. [Google Scholar]
  25. Nichol, A.; Achiam, J.; Schulman, J. On first-order meta-learning algorithms. arXiv 2018, arXiv:1803.02999. [Google Scholar]
  26. Mishra, N.; Rohaninejad, M.; Chen, X.; Abbeel, P. A simple neural attentive meta-learner. In Proceedings of the 6th International Conference on Learning Representations (ICLR 2018), Vancouver, BC, Canada, 30 April–3 May 2018. [Google Scholar]
  27. Van der Spoel, E.; Rozing, M.P.; Houwing-Duistermaat, J.J.; Eline Slagboom, P.; Beekman, M.; de Craen, A.J.M.; Westendorp, R.G.J.; van Heemst, D. Siamese Neural Networks for One-Shot Image Recognition. In Proceedings of the ICML—Deep Learning Workshop, Lille Grande Palais, Lille, France, 10–11 July 2015; Volume 7, pp. 956–963. [Google Scholar]
  28. Hoffer, E.; Ailon, N. Deep metric learning using triplet network. In Proceedings of the 3rd International Conference on Learning Representations (ICLR 2015)—Workshop Track Proceedings, Copenhagen, Denmark,, 12–14 October 2015. [Google Scholar]
  29. Oriol, V.; Charles, B.; Tim, L.; Koray, K.; Daan, W. Matching Networks for One Shot Learning. In Proceedings of the 30th International Conference on Neural Information Processing Systems, Barcelona, Spain, 5–10 December 2016; pp. 3630–3638. [Google Scholar]
  30. Snell, J.; Swersky, K.; Zemel, R.S. Prototypical Networks for Few-shot Learning. Adv. Neural Inf. Process. Syst. 2017, 30, 4077–4087. [Google Scholar]
  31. Sung, F.; Yang, Y.; Zhang, L.; Xiang, T.; Torr, P.H.; Hospedales, T.M. Learning to Compare: Relation Network for Few-Shot Learning. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 1199–1208. [Google Scholar] [CrossRef] [Green Version]
  32. Wang, Y.X.; Girshick, R.; Hebert, M.; Hariharan, B. Low-Shot Learning from Imaginary Data. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 7278–7286. [Google Scholar] [CrossRef] [Green Version]
  33. Antoniou, A.; Storkey, A.; Edwards, H. Data augmentation generative adversarial networks. arXiv 2017, arXiv:1711.04340. [Google Scholar]
  34. Guo, Y.; Codella, N.C.; Karlinsky, L.; Codella, J.V.; Smith, J.R.; Saenko, K.; Rosing, T.; Feris, R. A Broader Study of Cross-Domain Few-Shot Learning; Vedaldi, A., Bischof, H., Brox, T., Frahm, J.M., Eds.; Computer Vision—ECCV 2020. ECCV 2020. Lecture Notes in Computer Science; Springer: Cham, Switzerland, 2020; Volume 12372, pp. 124–141. [Google Scholar] [CrossRef]
  35. A low shot learning method for tea leaf’s disease identification. Comput. Electron. Agric. 2019, 163, 104852. [CrossRef]
  36. Janarthan, S.; Thuseethan, S.; Rajasegarar, S.; Lyu, Q.; Zheng, Y.; Yearwood, J. Deep Metric Learning Based Citrus Disease Classification With Sparse Data. IEEE Access 2020, 8, 162588–162600. [Google Scholar] [CrossRef]
  37. Wang, B.; Wang, D. Plant leaves classification: A few-shot learning method based on siamese network. IEEE Access 2019, 7, 151754–151763. [Google Scholar] [CrossRef]
  38. Argüeso, D.; Picon, A.; Irusta, U.; Medela, A.; San-Emeterio, M.G.; Bereciartua, A.; Alvarez-Gila, A. Few-Shot Learning approach for plant disease classification using images taken in the field. Comput. Electron. Agric. 2020, 175. [Google Scholar] [CrossRef]
  39. Hsu, C.W.; Lin, C.J. A comparison of methods for multiclass support vector machines. IEEE Trans. Neural Netw. 2002, 13, 415–425. [Google Scholar]
  40. Duan, Y.; Zheng, W.; Lin, X.; Lu, J.; Zhou, J. Deep Adversarial Metric Learning. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 2780–2789. [Google Scholar] [CrossRef]
  41. Weinberger, K.Q.; Blitzer, J.; Saul, L.K. Distance metric learning for large margin nearest neighbor classification. Adv. Neural Inf. Process. Syst. 2005, 1473–1480. [Google Scholar] [CrossRef]
  42. Kingma, D.P.; Ba, J.L. Adam: A method for stochastic optimization. In Proceedings of the 3rd International Conference on Learning Representations (ICLR 2015)—Conference Track Proceedings, Copenhagen, Denmark, 12–14 October 2015. [Google Scholar]
  43. Paszke, A.; Gross, S.; Massa, F.; Lerer, A.; Bradbury, J.; Chanan, G.; Killeen, T.; Lin, Z.; Gimelshein, N.; Antiga, L.; et al. PyTorch: An imperative style, high-performance deep learning library. Adv. Neural Inf. Process. Syst. 2019, 32, 8026–8037. [Google Scholar]
  44. Szegedy, C.; Vanhoucke, V.; Ioffe, S.; Shlens, J.; Wojna, Z. Rethinking the Inception Architecture for Computer Vision. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 2818–2826. [Google Scholar] [CrossRef] [Green Version]
Figure 1. Block diagram of the developed plant disease classification approaches.
Figure 1. Block diagram of the developed plant disease classification approaches.
Plants 10 00028 g001
Figure 2. Sample images from PlantVillage dataset. For the details of different classes, check Table 1.
Figure 2. Sample images from PlantVillage dataset. For the details of different classes, check Table 1.
Plants 10 00028 g002
Figure 3. Sample images from coffee leaf dataset. For the details of different classes, check Table 3.
Figure 3. Sample images from coffee leaf dataset. For the details of different classes, check Table 3.
Plants 10 00028 g003
Figure 4. An overview of the transfer learning methodology.
Figure 4. An overview of the transfer learning methodology.
Plants 10 00028 g004
Figure 5. The structure of the Triplet network, the three instances have the shared weights.
Figure 5. The structure of the Triplet network, the three instances have the shared weights.
Plants 10 00028 g005
Figure 6. Classification based on both crop and disease types (PlantVillage target domain dataset), mean error as a function of the number of shots per class, K, for 4 learning algorithms and 3 backbone models.
Figure 6. Classification based on both crop and disease types (PlantVillage target domain dataset), mean error as a function of the number of shots per class, K, for 4 learning algorithms and 3 backbone models.
Plants 10 00028 g006
Figure 7. Classification based on both crop and disease types (coffee leaf [2] target domain dataset), mean error as a function of the number of shots per class, K, for 4 learning algorithms and 3 backbone models.
Figure 7. Classification based on both crop and disease types (coffee leaf [2] target domain dataset), mean error as a function of the number of shots per class, K, for 4 learning algorithms and 3 backbone models.
Plants 10 00028 g007
Figure 8. Classification based on the common disease name (rearranged PlantVillage target domain dataset—see Table 2), mean error as a function of the number of shots per class, K, for 4 learning algorithms and 3 backbone models.
Figure 8. Classification based on the common disease name (rearranged PlantVillage target domain dataset—see Table 2), mean error as a function of the number of shots per class, K, for 4 learning algorithms and 3 backbone models.
Plants 10 00028 g008
Figure 9. Classification based on the common disease name (coffee leaf [2] target domain dataset, mean error as a function of the number of shots per class, K, for 4 learning algorithms and 3 backbone models.
Figure 9. Classification based on the common disease name (coffee leaf [2] target domain dataset, mean error as a function of the number of shots per class, K, for 4 learning algorithms and 3 backbone models.
Plants 10 00028 g009
Table 1. Summary of the PlantVillage datase [14].
Table 1. Summary of the PlantVillage datase [14].
ClassCropDiseaseSamples
TrainTest
C1AppleApple cab504126
C2AppleBlack rot496125
C3AppleCedar apple rust22055
C4AppleHealthy1316329
C5BlueberryHealthy1202300
C6Cherry (including sour)Healthy684170
C7Cherry (including sour)Powdery mildew842210
C8Corn (maize)Cercospora leaf spot Gray leaf spot410103
C9Corn (maize)Common rust953239
C10Corn (maize)Healthy929233
C11Corn (maize)Northern Leaf Blight788197
C12GrapeBlack rot944236
C13GrapeEsca (Black Measles)1107276
C14GrapeHealthy33984
C15GrapeLeaf blight (Isariopsis Leaf Spot)861215
C16OrangeHaunglongbing (Citrus greening)44051102
C17PeachBacterial spot1838459
C18PeachHealthy28872
C19Pepper bellBacterial spot797200
C20Pepper bellHealthy1183295
C21PotatoEarly blight800200
C22PotatoHealthy12131
C23PotatoLate blight800200
C24RaspberryHealthy29774
C25SoybeanHealthy40721018
C26SquashPowdery mildew1468367
C27StrawberryHealthy36492
C28StrawberryLeaf scorch887222
C29TomatoBacterial spot1702425
C30TomatoEarly blight800200
C31TomatoHealthy1273318
C32TomatoLate blight1527382
C33TomatoLeaf mold761191
C34TomatoSeptoria leaf spot1417354
C35TomatoSpider mites Two-spotted spider mite1341335
C36TomatoTarget spot1123281
C37TomatoTomato mosaic virus29974
C38TomatoTomato Yellow Leaf Curl Virus42861071
Table 2. Summary of the PlantVillage dataset [14] based on disease name.
Table 2. Summary of the PlantVillage dataset [14] based on disease name.
ClassDiseaseAffected PlantsSamples
TrainTest
CD1Apple scabApple504126
CD2Bacterial spotPeach, Pepper bell, Tomato43371084
CD3Black rotApple, Grape1140361
CD4Cedar apple rustApple22055
CD5Cercospora leaf spot Gray leaf spotcorn440103
CD6Common rustcorn953239
CD7Early blightPotato, Tomato1600400
CD8Esca black measlesGrape1107276
CD9Haunglongbing Citrus greeningOrange44051102
CD10Late blightPotato, Tomato2327582
CD11Leaf blight Isariopsis Leaf SpotGrape861215
CD12Leaf moldTomato761191
CD13Leaf scorchStrawberry887222
CD14Northern Leaf blightCorn817197
CD15Powdery mildewCherry, Squash2310577
CD16Septoria leaf spotTomato1417354
CD17Spider mites Two spotted spider miteTomato1341335
CD18Target spotTomato1123281
CD19Tomato mosaic virusTomato29974
CD20Tomato Yellow Leaf Curl VirusTomato42861071
CD21Healthy49091200
Table 3. Summary of the Coffee Leaf dataset [2].
Table 3. Summary of the Coffee Leaf dataset [2].
ClassPredominant DiseaseSamples
CF1Leaf miner387
CF2Rust531
CF3Brown leaf spot348
CF4Cercospora leaf spot147
CF5(Several with same severity)62
CF6Healthy272
Table 4. Classification based on both crop and disease types (PlantVillage target domain dataset), the lowest error for 5, 25, and 50 shots is highlighted in bold.
Table 4. Classification based on both crop and disease types (PlantVillage target domain dataset), the lowest error for 5, 25, and 50 shots is highlighted in bold.
ApproachResNet18/(Shots)ResNet34/(Shots)ResNet50/(Shots)
525505255052550
Baseline17.785.584.2414.64.382.5214.423.982.48
Baseline++39.4820.9214.5844.2218.413.1643.6219.9214.02
Triplet Network18.068.045.0618.17.84.2818.146.744.8
DAML18.387.95.5216.546.74.4616.847.824.92
Table 5. Classification based on both crop and disease types (Coffee leaf [2] target domain dataset), the lowest error for 5, 25, and 50 shots is highlighted in bold.
Table 5. Classification based on both crop and disease types (Coffee leaf [2] target domain dataset), the lowest error for 5, 25, and 50 shots is highlighted in bold.
ApproachResNet18/(Shots)ResNet34/(Shots)ResNet50/(Shots)
525505255052550
Baseline44.7829.4625.0836.6825.7622.6635.5422.6220.22
Baseline++54.9643.8438.3856.738.533.3260.1238.3232.7
Triplet Network43.0227.623.545.928.923.9647.1430.7424.48
DAML4328.223.5842.5227.1421.7843.8228.5823.14
Table 6. Classification based on the common disease name (rearranged PlantVillage target domain dataset—see Table 2), the lowest error for 5, 25, and 50 shots is highlighted in bold.
Table 6. Classification based on the common disease name (rearranged PlantVillage target domain dataset—see Table 2), the lowest error for 5, 25, and 50 shots is highlighted in bold.
ApproachResNet18/(Shots)ResNet34/(Shots)ResNet50/(Shots)
525505255052550
Baseline2.520.90.523.460.560.285.081.0180.66
Baseline++16.79.66.9419.089.686.5623.588.243.98
Triplet Network11.645.062.413.044.142.489.13.181.32
DAML13.845.13.1810.182.741.6211.663.442.32
Table 7. Classification based on the common disease name (coffee leaf [2] target domain dataset), the lowest error for 5, 25, and 50 shots is highlighted in bold.
Table 7. Classification based on the common disease name (coffee leaf [2] target domain dataset), the lowest error for 5, 25, and 50 shots is highlighted in bold.
ApproachResNet18/(Shots)ResNet34/(Shots)ResNet50/(Shots)
525505255052550
Baseline38.76625.8823.0836.724.521.1434.2822.5619.44
Baseline++47.3836.3833.3445.536.2433.4649.4235.2830.76
Triplet Network47.2231.1826.0847.0629.4824.5850.7231.2625.3
DAML42.5428.5623.8642.627.1222.245.7829.325.34
Table 8. Comparison of the baseline model (the best performing model), Triplet network (ResNet50), and Triplet network Argueso et al. [38].
Table 8. Comparison of the baseline model (the best performing model), Triplet network (ResNet50), and Triplet network Argueso et al. [38].
5-Shots25-Shots50-Shots
Baseline (ResNet50) 85.58 % 96.02 % 97.52 %
Triplet Network (ResNet50) 81.86 % 93.26 % 95.20 %
Argueso et al. [38] 72 % 85 % 88 %
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Afifi, A.; Alhumam, A.; Abdelwahab, A. Convolutional Neural Network for Automatic Identification of Plant Diseases with Limited Data. Plants 2021, 10, 28. https://doi.org/10.3390/plants10010028

AMA Style

Afifi A, Alhumam A, Abdelwahab A. Convolutional Neural Network for Automatic Identification of Plant Diseases with Limited Data. Plants. 2021; 10(1):28. https://doi.org/10.3390/plants10010028

Chicago/Turabian Style

Afifi, Ahmed, Abdulaziz Alhumam, and Amira Abdelwahab. 2021. "Convolutional Neural Network for Automatic Identification of Plant Diseases with Limited Data" Plants 10, no. 1: 28. https://doi.org/10.3390/plants10010028

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop