Next Article in Journal
Prototype System for Measuring and Analyzing Movements of the Upper Limb for the Detection of Occupational Hazards
Next Article in Special Issue
Monitoring the Growth and Yield of Fruit Vegetables in a Greenhouse Using a Three-Dimensional Scanner
Previous Article in Journal
A Fully-Printed CRLH Dual-Band Dipole Antenna Fed by a Compact CRLH Dual-Band Balun
Previous Article in Special Issue
Estimation of a New Canopy Structure Parameter for Rice Using Smartphone Photography
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Classification Accuracy Improvement for Small-Size Citrus Pests and Diseases Using Bridge Connections in Deep Neural Networks

Center for Advanced Image and Information Technology, School of Electronics & Information Engineering, Chon Buk National University, Jeonju, Chon Buk 54896, Korea
*
Author to whom correspondence should be addressed.
Sensors 2020, 20(17), 4992; https://doi.org/10.3390/s20174992
Submission received: 3 August 2020 / Revised: 21 August 2020 / Accepted: 30 August 2020 / Published: 3 September 2020
(This article belongs to the Special Issue Smart Agriculture Sensors)

Abstract

:
Due to the rich vitamin content in citrus fruit, citrus is an important crop around the world. However, the yield of these citrus crops is often reduced due to the damage of various pests and diseases. In order to mitigate these problems, several convolutional neural networks were applied to detect them. It is of note that the performance of these selected models degraded as the size of the target object in the image decreased. To adapt to scale changes, a new feature reuse method named bridge connection was developed. With the help of bridge connections, the accuracy of baseline networks was improved at little additional computation cost. The proposed BridgeNet-19 achieved the highest classification accuracy (95.47%), followed by the pre-trained VGG-19 (95.01%) and VGG-19 with bridge connections (94.73%). The use of bridge connections also strengthens the flexibility of sensors for image acquisition. It is unnecessary to pay more attention to adjusting the distance between a camera and pests and diseases.

1. Introduction

Pests and diseases are major causes of huge economic loss in agricultural production. Timely detection and identification of these pests and diseases is essential to control their impact. Farmers used to rely on experienced experts to perform these tasks; unfortunately, this is a time-consuming, error-ridden, and costly process. To improve recognition efficiency, image processing and computer vision techniques are widely implemented. Bashish et al. [1] proposed a solution for automatic detection and classification of plant leaf diseases. Their method was based on image processing, which used a K-means clustering technique to segment RGB images and then used a single-layered artificial neural network model for classification. Ali et al. [2] presented a citrus disease recognition system in which a ∆E color difference algorithm was adopted to separate the disease affected areas, and a color histogram and texture features were extracted for classification. The features needed to recognize agricultural pests are more complex than those needed for diseases. They are sensitive to affine transformation, illumination, and viewpoint change. Wen et al. [3] introduced a local feature-based identification method to account for variations in insect appearance. Xie et al. [4] fused multiple features of insect species together in order to enhance recognition performance. Compared with other feature-combination methods, their approach produced higher accuracy, but needed more time to train.
The above research all required pretreatment to select features for classifiers. In contrast, deep convolutional neural networks (CNN) are representation-learning models [5]. They can receive raw image data as input and automatically discover the useful features for classification and detection. Various pieces of research were conducted with them, and they achieved satisfactory results. Mohanty et al. [6] employed two pre-trained CNN architectures to classify plant disease images in a public dataset. The deeper GoogleNet performed better than AlexNet in their experiment. The same image dataset was used by Too et al. [7], as they extended the types of CNN model for comparison. Their experimental results showed that DenseNet with more layers got the highest test accuracy. The pre-trained models also play an important role in pest recognition tasks. Cheng et al. [8] used deep residual networks and a transfer learning method to identify crop pests under complex farmland background. Shen et al. [9] developed an improved inception network to extract features and exploited Faster R-CNN to detect insect areas. Pre-trained models can be easily accessed on the internet [10]. It is unnecessary to make massive adjustments to them to get high classification accuracy [11], which facilitates the widespread use of deep CNNs. However, most pre-trained models, such as VGG, ResNet, and Inception, were designed based on ImageNet [12]. Overdependence on them without considering the actual scale of the target image dataset can result in a waste of computational resources.
Feature reuse is an indispensable component of many deep network architectures. It was first employed in the Highway Network [13] which defined two gating units to express how much output is produced by transforming and carrying input. ResNet [14] simplified the reuse method by adding input features directly to output features. DenseNet [15] boosted the frequency of skip-connections. Each layer in it can receive feature maps of all the preceding layers as inputs. This dense connectivity pattern improves the flow of information and gradients throughout the network. However, it also increases the probability of reusing low-level features that contribute little to classification accuracy [16]. The concatenation operation employed by DenseNet was still followed in this paper. Meanwhile, to reduce the effect of dense connections on parameter efficiency, only short-distance connections between layers were retained. It was proven that the classification layer of CNN models focuses on exploiting higher-level features [15,17]. We also observed that the utilization of different hierarchical features by the classification layer varies with the scale of the object in the image. Based on these facts, we propose a new feature reuse method that adapts to scale changes.
A new citrus pest and disease image dataset was recently created [18]. The images in the dataset were collected from a tangerine orchard on Jeju Island and open databases on the internet. Our model and several other benchmark networks were selected to recognize them. The proposed BridgeNet achieved higher classification accuracy than our previous model (Weakly DenseNet [18]), which proves the effectiveness of the new feature reuse method. Compared with training from scratch, ImageNet pre-training sped up convergence and improved performance. However, the network architectures designed for ImageNet demonstrated very serious over-fitting problems without using the transfer learning method [18], which indicates that they are over-complicated in fitting this small-scale image dataset.
The rest of this paper is organized as follows: Section 2 reviews the related work. Section 3 introduces the citrus pest and disease image dataset. Section 4 details the proposed CNN model. Comparison results and analysis are provided in Section 5, and Section 6 presents the conclusion.

2. Related Work

Compared with traditional machine learning methods, deep learning is an end-to-end model which does not require the design of handcrafted features for classifiers [19]. In addition, the development of computer hardware allows researchers to train networks with more hidden layers. Due to these advantages, deep learning became a popular data analysis method in many research fields. Esteva et al. [20] used the GoogleNet Inception v3 CNN architecture to classify skin cancer images. The model achieved dermatologist-level accuracy. Dorafshan et al. [21] employed four common edge-detection methods and AlexNet to detect cracks in concrete. Their experimental results showed that the deep CNN model performed much better than other approaches. Alves et al. [22] adopted deep residual networks to identify cotton pests. They provided a field-based image dataset containing 1600 images of 15 pests. The improved ResNet-34 achieved the highest classification accuracy. Chen et al. [23] created a CNN architecture named INC-VGGN to detect rice and maize diseases. They replaced the last few layers of VGG-19 with two Inception modules to further enhance the feature extraction ability. Compared with original networks, this new model showed the best performance. We designed a new feature reuse block (bridge connection) to adapt to the scale changes of citrus pests and diseases in images.
Network architecture has a considerable effect on model performance. In general, a deeper network constructed results in a higher accuracy being delivered [24]. However, as depth increases, training becomes more difficult. To alleviate this problem, different feature reuse methods were proposed [13,14,15]. Inception [25] and Fractal Networks [26] provide sub-paths of different lengths for each building unit. They can be designed to be very deep without needing skip-connections. More recently, the application of attention mechanisms was regarded as another efficient way to improve network precision [27,28,29]. These attention mechanisms can help a model focus on using salient features. These network architectures achieved great success in large-scale image classification tasks.
ImageNet pre-training enables people to easily access encouraging results [11]. For this reason, models designed for ImageNet are frequently used without considering the characteristics of the target dataset. Recht et al. [30] proved that the accuracy of the same network will show a large drop on a new dataset that has a similar distribution to the original data. This phenomenon indicates that any network architecture should be optimized when facing new datasets.

3. Image Dataset Description

Our dataset has 12,561 images. It covers 17 species of citrus pest and seven types of citrus disease. Each class contains over 350 images. Figure 1 shows the number of samples in each category. The image collection methods for citrus pests and diseases were described in our previous research [18]. To balance the data distribution in the dataset and improve the model’s generalization ability, several data augmentation approaches were adopted. Table 1 depicts the parameter settings for each data augmentation. Instead of using only one operation, we randomly selected three operations and performed them sequentially to produce a new image. This method can considerably increase the diversity of generated images (Figure 2).
Image samples used in earlier studies were gathered under laboratory conditions [6,31], which reduced the robustness of the trained model to realistic conditions [32]. In contrast, for this study, we collected images with variable, realistic backgrounds. In addition, to further adapt to real-world scenes, the original image was not excessively cropped to only keep the target object region [33]. Figure 3 shows images with varying distances between the camera sensor and the citrus pest or disease.

4. Network Architecture

Computer vision competitions greatly promoted the development of deep learning. Many advanced design methods were proposed to improve network performance, which changed the situation of simply stacking convolutional layers. We followed the strategy of SqueezeNet [34] to design the network structure from micro to macro. To save computational cost, our network depth was gradually increased until the accuracy was not significantly improved.

4.1. Microstructure of Building Unit

Attention mechanisms usually produce an attention map to highlight the important features, which brings an additional computation overhead and increases the optimization difficulty. We followed the micro-construction of the Network in Network [35] to enhance the features generated by each 3 × 3 convolution (Figure 4). This structure is compatible with the mainframe of a network with no need for extra branches. Furthermore, the Mlpconv layer receives each whole feature map as an input, avoiding over-compressing information like SE (Squeeze-and-Excitation) blocks [27].

4.2. Macro Connection between Building Blocks

In order to address the degradation problem, He et al. [14] developed a residual learning framework to add input features to output features. Huang et al. [15] adopted a concatenation operation to increase the frequency of feature reuse. Compared with the add operation, the concatenation feature reuse method is easier to use, which does not require a 1 × 1 convolution to align input and output channels [34]. In addition, concatenation takes less computation time than element-wise addition [17]. For these reasons, we reused previous layer features using concatenation. ShuffleNet V2 [17] proved that the amount of feature reuse decayed exponentially with the distance between two blocks. To avoid introducing redundancy, we only established connections between adjacent layers (Figure 5).

4.3. Adaption to Object Scale in the Image

It is well known that CNN models are very sensitive to translations and rotations [36]. We observed that scale changes to an object in an image can also affect neural network performance (Figure 6). In order to find the reason for this, we borrowed SE blocks [27] to monitor the contribution of features from different building blocks to the classification (Figure 7). After training, we selected several groups of images to validate the feature importance distribution in each SE block (Figure 8). For easy comparison, we divided features into three levels based on their importance values. Table 2 presents the number of features in each level. It can be seen that, as the object scale is reduced, the number of high-level features decreases while the number of mid-level features increases. This means that a network has to use more mid-level features to identify the class of smaller-size objects in images.
There are many mid-level and high-quality features in the intermediate layers. However, direct reuse of them will increase the computational complexity of the classification layer. In addition, the proportion of low-level features in shallower layers was greater than in deeper layers (Table 2). Based on these facts, we proposed a new feature reuse method to improve parameter efficiency. The 1 × 1 convolution in Figure 9 has two purposes.
  • Channel compression: To reduce the number of useless features, the number of output channels from the 1 × 1 convolution is fewer than the number of input channels. This function is similar to the transition layer of DenseNet.
  • Feature retention: Unlike a 3 × 3 convolution, a 1 × 1 convolution performs a simple linear transformation, which can largely preserve input feature information. To further strengthen output feature quality, two 1 × 1 convolutions were stacked after the concatenation operation.
Conventional feature reuse strategies (addition and concatenation) do not consider the discrepancy between features from different layers. More specifically, the difference between shallow features and deep features is not only in the distribution characteristic, but also in the representation complexity. This complexity difference between adjacent layers can be measured by Equation (1). As the distance between layers increases, the difference mentioned will be more significant. The progressive feature reuse method shown in Figure 9 ensures a strong correlation between concatenated features. In addition, the two 1 × 1 convolutions in each feature reuse block reduce the complexity difference between features from long-distance layers.
F n i   =   j = 1 k ( F n 1 j *   W j   +   b j ) ,
where F n i represents the i-th feature map of the n-th layer, W j and b j denote the corresponding convolutional kernel and bias to the F n   1 j feature map, and k presents the number of feature maps in the (n−1)-th layer.

5. Experiments and Results

5.1. Experiment Preparation

The overall architecture of BridgeNet for citrus pests and disease recognition is shown in Table 3. Several baseline networks and their variants that have a similar depth to BridgeNet were selected for comparison. We replaced the two 1 × 1 convolutions of the Mlpconv block with CBAM (Convolutional Block Attention Module) [28] to compare their performance. In addition, the bridge connection was compared with deformable convolution. We followed the suggestion of Dai et al. [29] to apply these deformable convolutions in the last three convolutional layers (with kernel size > 1). All these networks shared the same classification block (Figure 10) and were trained with identical optimization schemes. Before training, the original image dataset was divided into a training set, a validation set, and a test set in the ratio of 4:1:1. Then, each model was trained and tested based on them. The three parts did not contain the same samples, and data augmentation was performed for each model and only on the training set. We saved the models that had the highest validation accuracy and examined their generalization ability on the test set. The model hyper-parameters presented in Table 4 were determined by trial and error.

5.2. Classification Performance

Table 5 displays the classification accuracy for each model. BridgeNet-19 achieved the highest validation accuracy, followed by VGG-19 with bridge connections and then pre-trained VGG-19. The test accuracy of the models followed the same trend as the validation accuracy, except that the pre-trained VGG-19 ranked second, followed by VGG-19 with bridge connections. Obviously, the models trained from scratch produced lower accuracy than their ImageNet pre-trained counterparts. However, the pre-trained models consumed much more computing resources than their competitors.
It is of note that the use of deformable convolutions does not improve VGG-16 performance. In contrast, the application of bridge connections considerably increased validation and test accuracy. As for additional computational cost, the bridge connection created much less of a computational burden than deformable convolution; the use of bridge connections in the model costed 5.8 MB and the cost of using deformable convolutions was 40.8 MB. Weakly DenseNet-19 performed better than CBAMNet, which indicates that the two 1 × 1 convolutional layers used for feature enhancement are more effective than the attention mechanism. The accuracy of Weakly DenseNet-19 was further increased by using features from the middle layers for classification. In terms of additional computational cost, BridgeNet-19 spent 12.3 MB and MSN-19 consumed 10.3 MB. However, BridgeNet-19 obtained better performance than MSN-19, proving the higher parameter efficiency of using bridge connections. The smaller-size models took less training time per batch except for MSN-19, which had a slower training speed than BridgeNet-19. Figure 11 depicts the training details of each model. Models trained with ImageNet pre-training displayed the fastest convergence. However, models with branch structures required more epochs to reach the final convergence state. This phenomenon indicates that simpler structural models are easier to train.
To validate the effectiveness of the new feature reuse method for adapting to object scale changes, the confusion matrices of Weakly DenseNet-19 and BridgeNet-19 were compared (Figure 12). Using the comparison results, Figure 13 presents the images that were correctly identified by BridgeNet-19 but misclassified by Weakly DenseNet-19. It can be seen that, with the help of bridge connections, BridgeNet-19 has an enhanced ability to correctly classify images with small target objects. The use of bridge connections also improves the discrimination of similar categories, for example, the citrus anthracnose and canker (the difference between them is not obvious without close observation).

5.3. Ablation Study

We considered the number of bridge connections as a hyper-parameter and explored its impact on network performance. Bridge connections were introduced from top to bottom as the overall number of connections was increased. Weakly DenseNet-19 was used as the backbone architecture. Table 6 reports the comparison results. It can be observed that, as the number of bridge connections increased, the model performance showed initial cumulative growth. When the number was increased to four, the accuracy then decreased. This indicates that excessive use of shallow information will bring more redundancy to the classification layer.

6. Conclusion and Future Work

In this study, a new CNN model was developed to identify common pests and diseases in citrus plantations. Each building block of the network contained two 1 × 1 convolutions that were used to enhance the features generated by 3 × 3 convolutional layers. Concatenation operations were used for feature reuse. To reduce redundancy, only features from adjacent layers were concatenated. We observed that, as the size of the target object in the image decreased, the use of mid-level features by the classification layer increased. Using this insight to adapt the model to scale changes, a new feature reuse method called bridge connection was designed. Experimental results show that the proposed BridgeNet-19 achieved the highest classification accuracy (95.47%). Compared with pre-trained models, our network also presented higher parameter efficiency; its model size (68.9 MB) was half of the pre-trained VGG-16 and VGG-19 networks.
Training of deep CNN models usually requires a large-scale image dataset. However, it is very difficult and expensive to collect so many high-quality, close-up sample images in some specific fields such as medicine and biology. Although ImageNet pre-trained models can allow researchers to achieve satisfactory results on different types of datasets quickly and easily, they are designed too bulky and ill-suited to fit small datasets. We hope to find a better solution to solve this problem in the future.

Author Contributions

Data curation, S.X.; formal analysis, S.X.; funding acquisition, M.L.; methodology, S.X.; project administration, M.L.; software, S.X.; writing—original draft, S.X.; writing—review and editing, M.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by “Research Base Construction Fund Support Program” funded by Chon Buk National University in 2019, and a grant (No: 2019R1A2C108690412) of National Research Foundation of Korea (NRF) funded by the Korea government (MSP).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Al Bashish, D.; Braik, M.; Bani-Ahmad, S. A framework for detection and classification of plant leaf and stem diseases. In Proceedings of the 2010 International Conference on Signal and Image Processing, Chennai, India, 15–17 December 2010. [Google Scholar]
  2. Ali, H.; Lali, M.I.; Nawaz, M.Z.; Sharif, M.; Saleem, B.A. Symptom based automated detection of citrus diseases using color histogram and textural descriptors. Comput. Electron. Agric. 2017, 138, 92–104. [Google Scholar] [CrossRef]
  3. Wen, C.; Guyer, D.E.; Li, W. Local feature-based identification and classification for orchard insects. Biosyst. Eng. 2009, 104, 299–307. [Google Scholar] [CrossRef]
  4. Xie, C.; Zhang, J.; Li, R.; Li, J.; Hong, P.; Xia, J.; Chen, P. Automatic classification for field crop insects via multiple-task sparse representation and multiple-kernel learning. Comput. Electron. Agric. 2015, 119, 123–132. [Google Scholar] [CrossRef]
  5. LeCun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef] [PubMed]
  6. Mohanty, S.P.; Hughes, D.P.; Salathé, M. Using deep learning for image-based plant disease detection. Front. Plant Sci. 2016, 7, 1419. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  7. Too, E.C.; Yujian, L.; Njuki, S.; Yingchun, L. A comparative study of fine-tuning deep learning models for plant disease identification. Comput. Electron. Agric. 2019, 161, 272–279. [Google Scholar] [CrossRef]
  8. Cheng, X.; Zhang, Y.; Chen, Y.; Wu, Y.; Yue, Y. Pest identification via deep residual learning in complex background. Comput. Electron. Agric. 2017, 141, 351–356. [Google Scholar] [CrossRef]
  9. Shen, Y.; Zhou, H.; Li, J.; Jian, F.; Jayas, D.S. Detection of stored-grain insects using deep learning. Comput. Electron. Agric. 2018, 145, 319–325. [Google Scholar] [CrossRef]
  10. Alom, M.; Tha, T.; Yakopcic, C.; Westberg, S.; Sidike, P.; Nasrin, M.; Hasan, M.; Essen, B.; Awwal, A.; Asari, V. A State-of-the-Art Survey on Deep Learning Theory and Architectures. Electronics 2019, 8, 292. [Google Scholar] [CrossRef] [Green Version]
  11. He, K.; Girshick, R.; Dollár, P. Rethinking imagenet pre-training. In Proceedings of the IEEE International Conference on Computer Vision, Seoul, Korea, 27 October–2 November 2019. [Google Scholar]
  12. Deng, J.; Dong, W.; Socher, R.; Li, L.J.; Li, K.; Li, F.F. ImageNet: A large-scale hierarchical image database. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Miami, FL, USA, 20–25 June 2009. [Google Scholar]
  13. Srivastava, R.K.; Greff, K.; Schmidhuber, J. Highway Networks. arXiv 2015, arXiv:1505.00387. [Google Scholar]
  14. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. arXiv 2015, arXiv:1512.03385. [Google Scholar]
  15. Huang, G.; Liu, Z.; van der Maaten, L.; Weinberger, K.Q. Densely Connected Convolutional Networks. arXiv 2016, arXiv:1608.06993v5. [Google Scholar]
  16. Yosinski, J.; Clune, J.; Bengio, Y.; Lipson, H. How transferable are features in deep neural networks? In Proceedings of the 28th International Conference on Neural Information Processing Systems (NeurIPS), Montreal, PQ, Canada, 8–13 December 2014. [Google Scholar]
  17. Ma, N.; Zhang, X.; Zheng, H.; Sun, J. ShuffleNet V2: Practical Guidelines for Efficient CNN Architecture Design. arXiv 2018, arXiv:1807.11164. [Google Scholar]
  18. Xing, S.; Lee, M.; Lee, K.K. Citrus Pests and Diseases Recognition Model Using Weakly Dense Connected Convolution Network. Sensors 2019, 19, 3195. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  19. Nanni, L.; Ghidoni, S.; Brahnam, S. Handcrafted vs. non-handcrafted features for computer vision classification. Pattern Recognit. 2017, 71, 158–172. [Google Scholar] [CrossRef]
  20. Esteva, A.; Kuprel, B.; Novoa, R.A.; Ko, J.; Swetter, S.M.; Blau, H.M.; Thrun, S. Dermatologist-level classification of skin cancer with deep neural networks. Nature 2017, 542, 115–118. [Google Scholar] [CrossRef]
  21. Dorafshan, S.; Thomas, R.J.; Maguire, M. Comparison of deep convolutional neural networks and edge detectors for image-based crack detection in concrete. Constr. Build. Mater. 2018, 186, 1031–1045. [Google Scholar] [CrossRef]
  22. Alves, A.N.; Souza, W.S.; Borges, D.L. Cotton pests classification in field-based images using deep residual networks. Comput. Electron. Agric. 2020, 174, 105488. [Google Scholar] [CrossRef]
  23. Chen, J.; Chen, J.; Zhang, D.; Sun, Y.; Nanehkaran, Y.A. Using deep transfer learning for image-based plant disease identification. Comput. Electron. Agric. 2020, 173, 105393. [Google Scholar] [CrossRef]
  24. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  25. Szegedy, C.; Ioffe, S.; Vanhoucke, V.; Alemi, A.A. Inception-v4, inception-resnet and the impact of residual connections on learning. In Proceedings of the Thirty-First AAAI Conference on Artificial Intelligence, San Francisco, CA, USA, 4–9 February 2017. [Google Scholar]
  26. Larsson, G.; Maire, M.; Shakhnarovich, G. Fractalnet: Ultra-deep neural networks without residuals. In Proceedings of the 5th International Conference on Learning Representations, Toulon, France, 24–26 April 2017. [Google Scholar]
  27. Hu, J.; Shen, L.; Albanie, S.; Sun, G.; Wu, E. Squeeze-and-Excitation Networks. arXiv 2017, arXiv:1709.01507. [Google Scholar]
  28. Woo, S.; Park, J.; Lee, J.Y.; Kweon, I.S. CBAM: Convolutional Block Attention Module. arXiv 2018, arXiv:1807.06521. [Google Scholar]
  29. Dai, J.; Qi, H.; Xiong, Y.; Li, Y.; Zhang, G.; Hu, H.; Wei, Y. Deformable convolutional networks. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017. [Google Scholar]
  30. Recht, B.; Roelofs, R.; Schmidt, L.; Shankar, V. Do cifar-10 classifiers generalize to cifar-10? arXiv 2018, arXiv:1806.00451. [Google Scholar]
  31. Wang, G.; Sun, Y.; Wang, J. Automatic image-based plant disease severity estimation using deep learning. Comput. Intell. Neurosci. 2017, 2017, 2917536. [Google Scholar] [CrossRef] [Green Version]
  32. Barbedo, J.G.; Castro, G.B. Influence of image quality on the identification of psyllids using convolutional neural networks. Biosyst. Eng. 2019, 182, 151–158. [Google Scholar] [CrossRef]
  33. Thenmozhi, K.; Reddy, U.S. Crop pest classification based on deep convolutional neural network and transfer learning. Comput. Electron. Agric. 2019, 164, 104906. [Google Scholar] [CrossRef]
  34. Iandola, F.N.; Han, S.; Moskewicz, M.W.; Ashraf, K.; Dally, W.J.; Keutzer, K. SqueezeNet: AlexNet-level accuracy with 50x fewer parameters and <0.5 MB model size. arXiv 2016, arXiv:1602.07360. [Google Scholar]
  35. Lin, M.; Chen, Q.; Yan, S. Network in network. arXiv 2013, arXiv:1312.4400v3. [Google Scholar]
  36. Engstrom, L.; Tsipras, D.; Schmidt, L.; Madry, A. A rotation and a translation suffice: Fooling cnns with simple transformations. arXiv 2017, arXiv:1712.02779. [Google Scholar]
  37. He, K.; Zhang, X.; Ren, S.; Sun, J. Delving Deep into Rectifiers: Surpassing Human-Level Performance on ImageNet Classification. In Proceedings of the 2015 International Conference on Computer Vision (ICCV), Santiago, Chile, 11–18 December 2015. [Google Scholar]
  38. Ng, A.Y. Feature selection, L1 vs. L2 regularization, and rotational invariance. In Proceedings of the Twenty-First International Conference on Machine Learning, Banff, AB, Canada, 4–8 July 2004. [Google Scholar]
  39. Sutskever, I.; Martens, J.; Dahl, G.E.; Hinton, G.E. On the importance of initialization and momentum in deep learning. In Proceedings of the 30th International Conference on Machine Learning (ICML), Atlanta, GA, USA, 16–21 June 2013. [Google Scholar]
  40. Srivastava, N.; Hinton, G.; Krizhevsky, A.; Sutskever, I.; Salakhutdinov, R. Dropout: A simple way to prevent neural networks from overfitting. J. Mach. Learn. Res. 2014, 15, 1929–1958. [Google Scholar]
Figure 1. Data distribution of citrus pests and diseases.
Figure 1. Data distribution of citrus pests and diseases.
Sensors 20 04992 g001
Figure 2. Results of the proposed data augmentation method: (a) original image; (b–d) generated images.
Figure 2. Results of the proposed data augmentation method: (a) original image; (b–d) generated images.
Sensors 20 04992 g002
Figure 3. Photos taken with subject at different distances from camera: (ac) citrus canker; (df) southern green stink bug.
Figure 3. Photos taken with subject at different distances from camera: (ac) citrus canker; (df) southern green stink bug.
Sensors 20 04992 g003
Figure 4. Microstructure of a building block.
Figure 4. Microstructure of a building block.
Sensors 20 04992 g004
Figure 5. Connection across building blocks.
Figure 5. Connection across building blocks.
Sensors 20 04992 g005
Figure 6. Prediction accuracy of Weakly DenseNet-19 for the different scales of target objects: (a,b) fruit fly; (c,d) root weevil.
Figure 6. Prediction accuracy of Weakly DenseNet-19 for the different scales of target objects: (a,b) fruit fly; (c,d) root weevil.
Sensors 20 04992 g006
Figure 7. Application of SE (Squeeze-and-Excitation) blocks in testing the importance of middle layer features for classification. This network architecture is called Multi-Scale-Net (MSN) and is compared with other benchmark networks in Section 5.
Figure 7. Application of SE (Squeeze-and-Excitation) blocks in testing the importance of middle layer features for classification. This network architecture is called Multi-Scale-Net (MSN) and is compared with other benchmark networks in Section 5.
Sensors 20 04992 g007
Figure 8. Visualization of each SE block; mu and sigma represent mean and standard deviation. (a–c) and (k–m) are feature importance distributions for large-size canker (Figure 3a) and large-size southern green stink bug (Figure 3d); (d–f) and (n–p) are feature importance distributions for middle-size canker (Figure 3b) and middle-size southern green stink bug (Figure 3e); (h–j) and (q–s) are feature importance distributions for small-size canker (Figure 3c) and small-size southern green stink bug (Figure 3f); (a), (d), (h), (k), (n), and (q) are the output of SE block 1; (b), (e), (i), (l), (o), and (r) are the output of SE block 2; (c), (f), (j), (m), (p), and (s) are the output of SE block 3.
Figure 8. Visualization of each SE block; mu and sigma represent mean and standard deviation. (a–c) and (k–m) are feature importance distributions for large-size canker (Figure 3a) and large-size southern green stink bug (Figure 3d); (d–f) and (n–p) are feature importance distributions for middle-size canker (Figure 3b) and middle-size southern green stink bug (Figure 3e); (h–j) and (q–s) are feature importance distributions for small-size canker (Figure 3c) and small-size southern green stink bug (Figure 3f); (a), (d), (h), (k), (n), and (q) are the output of SE block 1; (b), (e), (i), (l), (o), and (r) are the output of SE block 2; (c), (f), (j), (m), (p), and (s) are the output of SE block 3.
Sensors 20 04992 g008aSensors 20 04992 g008b
Figure 9. Proposed feature reuse method.
Figure 9. Proposed feature reuse method.
Sensors 20 04992 g009
Figure 10. Classification block for each model.
Figure 10. Classification block for each model.
Sensors 20 04992 g010
Figure 11. Training process of selected models: “—” denotes training accuracy; “…” represents validation accuracy. (a) Red lines are VGG-16 trained with ImageNet pre-training, and blue lines are VGG-16 trained from scratch; (b) red lines are VGG-16 with bridge connections, and blue lines are VGG-16 with deformable convolutions; (c) red lines are Weakly DenseNet, and blue lines are CBAMNet; (d) red lines are BridgeNet-19, and blue lines are MSN-19; (e) red lines are pre-trained VGG-19, blue lines are VGG-19 with bridge connections, and magenta lines are VGG-19 trained from scratch.
Figure 11. Training process of selected models: “—” denotes training accuracy; “…” represents validation accuracy. (a) Red lines are VGG-16 trained with ImageNet pre-training, and blue lines are VGG-16 trained from scratch; (b) red lines are VGG-16 with bridge connections, and blue lines are VGG-16 with deformable convolutions; (c) red lines are Weakly DenseNet, and blue lines are CBAMNet; (d) red lines are BridgeNet-19, and blue lines are MSN-19; (e) red lines are pre-trained VGG-19, blue lines are VGG-19 with bridge connections, and magenta lines are VGG-19 trained from scratch.
Sensors 20 04992 g011
Figure 12. Comparison of confusion matrix: (a) Weakly DenseNet-19; (b) BridgeNet-19.
Figure 12. Comparison of confusion matrix: (a) Weakly DenseNet-19; (b) BridgeNet-19.
Sensors 20 04992 g012
Figure 13. Examples of images misclassified by Weakly DenseNet-19: only images that were misclassified due to the small subject size are presented.
Figure 13. Examples of images misclassified by Weakly DenseNet-19: only images that were misclassified due to the small subject size are presented.
Sensors 20 04992 g013
Table 1. Data augmentation operations.
Table 1. Data augmentation operations.
Operation Name.Value
Rotation[0°, 30°]
Width shift[0, 0.2]
Height shift[0, 0.2]
Zoom[0.8, 1.2]
Horizontal flip-
Brightness[0.6, 1.4]
Table 2. Statistics for different levels of features.
Table 2. Statistics for different levels of features.
The Number of Low-Level Features (Value < 0.6)The Number of Mid-Level Features (0.6 Value < 0.95)The Number of High-Level Features (Value 0.95)
Large-size cankerSE block 1102260
SE block 23501620
SE block 3138138236
Total590326236
Middle-size cankerSE block 1104240
SE block 23431690
SE block 3139198175
Total586391175
Small-size cankerSE block 1108200
SE block 23261860
SE block 3137222153
Total571428153
Large-size green stink bugSE block 1107210
SE block 23341780
SE block 3142184186
Total583383186
Middle-size green stink bugSE block 1107210
SE block 23361760
SE block 3147208157
Total590405157
Small-size green stink bugSE block 1103250
SE block 23231890
SE block 3143238131
Total569452131
Table 3. Structure of BridgeNet-19. The input size is 224 × 224 × 3. The initial block followed the setting of ResNet-50.
Table 3. Structure of BridgeNet-19. The input size is 224 × 224 × 3. The initial block followed the setting of ResNet-50.
Building UnitOutput Dimension
Initial Block56 × 56 × 32
Normal Block56 × 56 × 96
Reduction Block28 × 28 × 192
Normal Block28 × 28 × 384
Reduction BlockFeature Reuse Block14 × 14 × 76814 × 14 × 256
Mlpconv Layer with
Max Pooling
Feature Reuse Block7 × 7 × 5127 × 7 × 512
Feature Reuse Block without Max Pooling7 × 7 × 512
Classification Block1 × 24
Table 4. Hyper-parameters for training models.
Table 4. Hyper-parameters for training models.
Model Hyper-ParameterMethod/Value
Weight initializationHe normal [37]
Batch size16
Weight decayL2/0.005 [38]
OptimizerSGD
MomentumNesterov/0.9 [39]
Dropout rate0.5 [40]
Initial learning rateTraining from scratch/0.001
ImageNet pre-training/0.0001
Learning rate scheduleAdaptive [18]
Training epoch500
Table 5. Comparison of model performance.
Table 5. Comparison of model performance.
Model NameTraining Accuracy (%)Validation Accuracy (%)Test Accuracy
(%)
Model Size
(MB)
Training Speed
ms/Batch Size
VGG-1699.829392.93120.2303
Pre-trained VGG-1699.7994.9694.45120.2303
VGG-16 with bridge connections99.8994.1793.89126343
VGG-16 with deformable convolutions99.5992.6392.38161635
VGG-1999.8993.893.57155352
Pre-trained VGG-1999.9195.295.01155352
VGG-19 with bridge connections99.9395.3394.73167392
Weakly DenseNet-1999.7994.0393.7156.6139
CBAMNet99.9791.4290.7949.5139
MSN-1999.894.594.2466.9142
BridgeNet-1999.9795.6195.4768.9140
Table 6. Effect of using a different number of bridge connections.
Table 6. Effect of using a different number of bridge connections.
NumberTest Error (%)Model Size (MB)
06.2956.6
24.5767.6
34.5368.9
44.6269.3

Share and Cite

MDPI and ACS Style

Xing, S.; Lee, M. Classification Accuracy Improvement for Small-Size Citrus Pests and Diseases Using Bridge Connections in Deep Neural Networks. Sensors 2020, 20, 4992. https://doi.org/10.3390/s20174992

AMA Style

Xing S, Lee M. Classification Accuracy Improvement for Small-Size Citrus Pests and Diseases Using Bridge Connections in Deep Neural Networks. Sensors. 2020; 20(17):4992. https://doi.org/10.3390/s20174992

Chicago/Turabian Style

Xing, Shuli, and Malrey Lee. 2020. "Classification Accuracy Improvement for Small-Size Citrus Pests and Diseases Using Bridge Connections in Deep Neural Networks" Sensors 20, no. 17: 4992. https://doi.org/10.3390/s20174992

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop