Next Article in Journal
Two New Species of Ripipterygidae (Orthoptera, Tridactyloidea) from Mid-Cretaceous of Myanmar with a Key to the Genera of Tridactyloidea in Amber
Previous Article in Journal
Efficacy Evaluation of Oregano Essential Oil Mixed with Bacillus thuringiensis israelensis and Diflubenzuron against Culex pipiens and Aedes albopictus in Road Drains of Italy
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Attention-Based Multiscale Feature Pyramid Network for Corn Pest Detection under Wild Environment

1
School of Information Engineering, Southwest University of Science and Technology, Mianyang 621010, China
2
Institute of Intelligent Machines, Hefei Institutes of Physical Science, Chinese Academy of Sciences, Hefei 230031, China
3
School of Internet, Anhui University, Hefei 230601, China
*
Author to whom correspondence should be addressed.
Insects 2022, 13(11), 978; https://doi.org/10.3390/insects13110978
Submission received: 25 September 2022 / Revised: 22 October 2022 / Accepted: 23 October 2022 / Published: 25 October 2022
(This article belongs to the Section Insect Pest and Vector Management)

Abstract

:

Simple Summary

Corn pest recognition and detection is an important step for Integrated Pest Management. Generally, traditional methods adopt manual observation and counting in wild field to monitor the occurrence degree of corn pests. However, this is time-consuming and labor-intensive. An accurate and automatic corn pest detection method based on a deep convolutional neural network has been proposed in this paper. Extensive experimental results on a large-scale corn pest dataset show that the proposed method has good performance and can achieve precise recognition and detection of corn pests.

Abstract

A serious outbreak of agricultural pests results in a great loss of corn production. Therefore, accurate and robust corn pest detection is important during the early warning, which can achieve the prevention of the damage caused by corn pests. To obtain an accurate detection of corn pests, a new method based on a convolutional neural network is introduced in this paper. Firstly, a large-scale corn pest dataset has been constructed which includes 7741 corn pest images with 10 classes. Secondly, a deep residual network with deformable convolution has been introduced to obtain the features of the corn pest images. To address the detection task of multi-scale corn pests, an attention-based multi-scale feature pyramid network has been developed. Finally, we combined the proposed modules with a two-stage detector into a single network, which achieves the identification and localization of corn pests in an image. Experimental results on the corn pest dataset demonstrate that the proposed method has good performance compared with other methods. Specifically, the proposed method achieves 70.1% mean Average Precision (mAP) and 74.3% Recall at the speed of 17.0 frames per second (FPS), which balances the accuracy and efficiency.

1. Introduction

Corn is a very important food crop in many countries. However, corn growth suffers from pests, resulting in great losses in corn production. Therefore, the early warning and forecast are the basis of the effective prevention and control of corn pests, which play important roles in agricultural management and decision-making for corn production. Until now, corn pests were detected by using the experience of agricultural technicians, which is time-consuming and uses up material resources. Additionally, the acquired information on corn pests is lagging in some remote areas. Thus, it needs to explore a fast, automatic, less expensive, and accurate method to address the detection of corn pests, which has great realistic significance.
In recent years, many machine-learning-based methods have been used for pest detection. For example, a scale-invariant feature transform (SIFT)-based feature learning method is used to recognize stone fly larvae images [1]; experimental results show that the recognition accuracy can achieve 82% for four classes of the stonefly. Based on the K-means cluster and correspondence filter, Faithpraise et al. developed a plant pest recognition system [2]. Wen et al. designed an image-based orchard insect automated identification and classification method using affine invariant local features, which achieves a classification rate of 86.6% [3]. Xie et al. proposed an automatic classification of field crop insects using multiple-task sparse representation and multiple-kernel learning [4]. However, the above machine-learning pest recognition methods need complex hand-crafted feature descriptors. Extracting accurate pest characteristics becomes more difficult when the backgrounds are complex in the wild environment.
Recently, because the convolutional neural network (CNN) has a powerful ability of feature extraction, the deep learning-based method has been widely applied to address the tasks of recognition and detection, which obtains great success. These object detection methods can be divided into proposal-based and proposal-free methods. For proposal-based methods, it can be divided into two parts, including region proposal generation and multi-classes detection. For example, Ross et al. proposed a region with CNN features (R-CNN) object detection method. It first generates about 2k region proposals for the input images, extracts a feature vector from each region proposal using CNN, and finally recognizes the class of each region proposal [5]. However, the detection efficiency is slow. Ross designed a fast region-based convolutional neural network method (Fast R-CNN) [6], which can improve the detection efficiency by using CNN to classify the object regions. However, the generation of object proposal is time-consuming. Ren et al. introduced Faster R-CNN [7] modules, which adopts a region proposal network (RPN) to produce proposals with few computations. To solve the multi-scale object detection problem, Lin et al. developed a feature pyramid network (FPN) [8], which can use multiple feature maps with different scales to recognize and localize objects. The feature maps of high level can be used to detect large objects and the feature maps of bottom layers with detailed information can be applied to detect small objects. To further improve the accuracy of object detection, multi-stage detectors are designed; for example, Cascade R-CNN detector [9]. Based on the two-stage detector, it adopts a three-stage detection method for multi-classes recognition and localization. However, the proposal-based methods will result in low efficiency. Thus, many region-free object detection methods have been proposed. These methods directly predict the localization and classes of objects. The traditional region-free detectors are the series of YOLO, including from YOLO v1 to YOLO v5 [10,11,12]. Different from YOLO, CornerNet predicts the top-left and right-bottom corners of the bounding box to achieve object detection [13]. FCOS adopts a new point-based prediction method, which further improves the detection speed and accuracy [14]. Based on the idea of points detection, a large number of point-based region-free detectors are proposed, including CenterNet [15], CentriptalNet [16], ExtremeNet [17], and so on, which perform well with object detection task.
Therefore, many researchers introduced CNN-based detection methods into agricultural pest detection. For example, a deep, learning-based, automatic, multi-class, wild pest recognition and localization method has been proposed using deep, hybrid, global, ad local activated features [18]. It could achieve 75.03% mAP on the built large-scale pest dataset, which outweighs other state-of-the-art methods. Wang et al. [19] have developed a deep convolutional neural network-based module to recognize pest images with 20 categories, which have a good practical significance for the intelligent identification of agriculture and forestry pests. Rahman et al. adopted state-of-the-art large-scale architectures such as VGG16 and InceptionV3 and fine-tuned them for detecting and recognizing rice diseases and pests [20]. Experimental results show the effectiveness of these models with real datasets. To address the precise detection of multi-class pests with small sizes, Jiao et al. [21] proposed an anchor-free, region-based, convolutional neural network (AF-RCNN) inspired by the working mechanism of the population visual system. Several experiments show that this method can obtain 56.4% mAP and 85.1% mRecall on a 24-classes pest dataset, which outperforms state-of-the-art methods at that time. Dong et al. introduced a multi-scale feature fusion network to detect multiply categories of pests, which has great improvements compared with other methods [22]. Additionally, to address the detection of aphids with tiny size and dense distribution, Teng et al. applied a transformer feature pyramid network and multi-resolution training method, which outperforms other state-of-the-art detectors [23]. Apart from the above pest detection methods, other CNN-based detection methods [24,25,26] are almost based on a two-stage object detector, Faster RCNN [7] and its modified versions [27,28], to identify and detect pests.
However, as we know that recent research rarely pays attention to the identification of corn pests; even if there are, only a small number of categories of corn pests are identified and detected. Thus, the lack of the corn pest dataset will hinder the precise recognition. Additionally, due to the influence of uncertain factors such as illumination and occlusion, the accuracy of current methods for corn pests with complex backgrounds is not high. Therefore, in this paper, we first built a large-scale corn pest dataset, including 7392 images with 10 types of corn pests. Then, to extract rich feature information from the corn pest image, we introduced deformable convolution into the deep residual network. Additionally, an attention-based, multi-scale, feature pyramid network is simultaneously designed, achieving accurate detection of corn with various scales. Finally, the proposed methods are combined with a Faster R-CNN detector into a unified network and achieve the detection of multiple classes of corn pests. The main contributions are listed as follows:
(1)
A deep residual network with deformable convolution is introduced to extract rich feature information of corn pests, which improves the expression ability of information of the network.
(2)
An attention-based multi-scale feature pyramid network is used to address the detection of corn pests of different sizes.
(3)
We have constructed a large-scale corn pest dataset, including 7392 corn pest images and 10 types of corn pests. By combining our method with the two-stage detector, the proposed method can achieve 70.1%mAP and 74.3% recall on the corn pest dataset.

2. Materials and Methods

2.1. Materials

2.1.1. Corn Pest Image Collection

The corn pest images were collected in the wild using a mobile phone and camera from 2018 to 2020 in the Anhui and Henan provinces. All images were saved as JPG format with various sizes. The dataset contains 7392 images with 10 common types of corn pests, as shown in Table 1. It shows the names of these corn pests, the number of corn pest instances, and the average relative scale size to the whole pest images. Additionally, Figure 1 shows examples of each corn pest.

2.1.2. Data Labeling

Data labeling is an important task in CNN-based object detectors. LabelImg software was used to annotate the label and location of each pest instance in an image by the professional plant protection expert. A pest in an image is usually labeled as ( x , y , w , h , κ ) , where ( x , y ) represents the coordinate of the center of bounding box, ( w , h ) denotes the width and height of the bounding box, and κ is the class of the corn pest. Pest location coordinates and classes are saved as an XML file. The number of annotated samples corresponds to the number of bounding boxes labeled in each image. Every image could contain more than one annotation depending on the number and classes of pests.

2.1.3. Data Splitting

To train and evaluate the performance of the CNN-based objector, all images and the corresponding annotations are randomly separated into the training set and testing set by a ratio of 9:1, including 6653 images for network training and 740 images for network testing.

2.1.4. Analysis of Corn Pest Dataset

As shown in the left of Figure 2, we can observe that the relative size of the corn pest tends to be small, which will bring a great challenge to accurate detection. Additionally, the distribution of the number of pest instances is imbalanced which hinders the recognition of corn pests with few samples, as shown in the right of Figure 2.

2.2. Methods

In this section, we report implementation details of our proposed detector, as shown in Figure 3. First, we introduced the deep residual network with deformable convolution. We revisit the network architecture of the feature pyramid network (FPN) and analyze its working principle. Then, attention-based multi-scale feature fusion pyramid networks (AMFFP-Net) were developed based on FPN. Finally, we merged the proposed modules with the Faster R-CNN detector, achieving the recognition and localization of multi-classes corn pests.

2.2.1. Deep Residual Network with Deformable Convolution Block

Deep residual network has a great ability to extract features, which performs well in various object detection tasks. However, for our corn pest detection task, there exist geometric transformations in pose, viewpoints, and part deformation of corn pest images, which bring great challenges for precise recognition and detection. Therefore, we introduce deformable convolution into deep residual network [29] which enhances the robust representation of corn pest.
The deformable convolution added an offset based on regular convolution; the parameters of the offset can be obtained by learning. Compared with regular convolution, the sampling space of deformable convolution is enlarged by adding the offsets; therefore, the area of receptive field is changed. To be specific, the process of deformable convolution can be defined by Equation (1):
y ( l 0 ) = p n Ω w ( l n ) x ( l 0 + l n + Δ l n )
where y ( l 0 ) represents the output of each location l0 in input x ; Ω represents the sampling space in the input feature map x ; w is the learnable weight; l n enumerates the location of sampling space Ω . Δ l n denotes the learnable offset, which can be obtained by the convolutional neural network.
Since the position after adding the offset is not an integer and does not correspond to the actual pixel points on the feature map, it is necessary to use interpolation to obtain the offset pixel values. Usually, the bilinear interpolation can be used. The formula is as follows:
x ( l ) = k G ( k , l ) x ( k ) = k g ( k x , l x ) g ( k y , l y ) x ( k ) = k max ( 0 , 1 k x l x ) max ( 0 , k y l y ) x ( k )
where G ( k , l ) denotes the two dimensional bilinear interpolation kernel and l represents the arbitrary location ( l = l 0 + l n + Δ l n ).
From Figure 3b, we can observe that the offset can be learned from a convolutional layer, and the kernel size is 3 × 3 with Dilation 1. Therefore, the size of the output feature map is the same with the input feature map. When training the network, we use a back-propagated (BP) algorithm to learn the parameter of the offsets.

2.2.2. Attention-Based Multi-Scale Feature Fusion Pyramid Network (AMFFP-Net)

To obtain the multi-scale features of corn pest images, we learn from feature pyramid network (FPN) [8], which merges the low-level and high-level feature maps. In this paper, we proposed an attention-based multi-scale feature fusion pyramid network to enhance the expression ability of FPN. In this section, we first describe the architecture of the FPN network, then introduce our proposed attention-based, multi-scale, feature fusion pyramid network in detail.
(1) Revisiting feature fusion in FPN
Two key elements, including the downsampling factor and the fusion proportion between adjacent layers, affect the performance of FPN. Previous works improve the performance by decreasing the downsampling factor; however, this will lead to an increase in the computation complexity.
In this section, we provide the background of the FPN [8]. Let B denote the 1 × 1 convolutional operation for changing channels, and F u p denotes upsampling operation for increasing solutions. Therefore, the aggregation of adjacent feature layers in the following manner:
P i = B i ( X i ) + α F u p ( P i + 1 )
where, α represents the fusion factor between two different adjacent layers, which is set to 1 in FPN.
(2) Attention-based feature fusion
We can observe that the fusion factor in FPN is the same regardless of the layers of feature maps. This will result in poor distinguishability during feature fusion between different layers. Therefore, in this study, we have added a learnable fusion factor to increase the distinguishability, which can benefit the recognition of different objects. Figure 3a shows the network architecture of our attentional FPN. In this figure, we only take three layers of a pyramid network as an example; however, we adopted feature maps from five residual blocks of ResNet [29] in the proposed AMFFP-Net module. Similar to FPN, all feature maps generated by each residual block were processed by a 1 × 1 convolutional layer for reducing the number of channels. Specifically, the feature map F3 was 2 × up-sampled by nearest interpolation and was thens fed into an attentional weight generator for producing the weights used in feature fusion. Then, the feature map F3 with attentional weights was fused with feature map F2. It is noted that, similar to FPN, our AMFFP-Net has five outputs; the top features P5 and P6 can be obtained by twice subsampling. Finally, we append a 3 × 3 convolutional layer to eliminate the aliasing effect.
The fusion process also can be represented as Equation (3). Different from FPN, the α in our AMFFP-Net is changeable. Feature maps from different levels have different α s . Thus, there are different α s in our AMFFP-Net module. In AMFFP-Net, the α s are developed by the attentional weights generator, as shown in Figure 3c. The attentional weights generator consists of a convolutional layer with a 1 × 1 kernel size, a ReLu activation function for non-linear transformation, a convolutional layer with 3 × 3 kernel size, and a sigmoid function used for generating weight maps.

2.2.3. Joint Detection

Following a two-stage detector, e.g., Faster R-CNN [7], we combined the proposed deep residual network with deformable convolution and AMFFP-Net with Fast R-CNN [6] detector. To be specific, a region proposal network (RPN) has been used to generate a set of pest proposals. Then, these corn pest proposals are input into the two fully connected layers, followed by the classification layer with (c + 1) outputs (c is the number of classes of corn pests) and the localization layer with 4c outputs. Finally, we obtained the name of the corn pest and its corresponding location in an image.

2.3. Evaluation Metrics

To evaluate the detection accuracy of our model and other compared methods, several standard metrics are applied, such as mean Average Precision (mAP) and Recall, which can be calculated as follows. First, the IoU is used to verify the overlap between the ground-truth and predicted bounding box, which can be defined:
I o U = a r e a ( G P ) a r e a ( G P )
where G and P denote the ground truth and predicted bounding box, respectively. The a r e a ( G P ) denotes the intersection of the ground truth and predicted bounding box and a r e a ( G P ) denotes the union of the ground truth and predicted bounding box.
Secondly, according to the value of IoU, we decided the true positive (TP) and false positive (FP). If the IoU of the predicted bounding box and ground truth is greater than 0.5, then the predicted bounding box is viewed as TP. Otherwise, it is FP. Thus, the precision and recall can be calculated using:
p r e c i s i o n = # T P # T P + # F P
r e c a l l = # T P G T
where # T P and # F P represent the number of detected and misdetected corn pests, respectively. Ground Truth (GT) denotes the total number of corn pests.
The precision and recall metrics are combined to fairly evaluate the performance of our method. Thus, the Average Precision (AP) was adopted to verify the models. The AP denotes the area under the precision/recall curve, which can be defined in Equation (7). Mean AP (mAP) averaged over all object classes is employed as the final measure to compare performance on all object classes, and it is defined as Equation (8):
A P = 0 1 P d R
m A P = 1 c j = 1 c A P j
where c is the number of classes, which is set to 10 in this work.

3. Results and Analysis

3.1. Experimental Platform and Parameters Setting

Experiment platform: All experiments of this work were run on a workstation equipped with one NVIDIA RTX 2080Ti GPU with 24 GB memory. The software environment is Ubuntu 18.02 from Canonical Ltd in London, UK, which is an open source software, and Python 3.8 designed by Guido van Rossum in Commonwealth of Virginia, American. All CNN-based models have been built using Pytorch, which is an open source framework designed by Facebook.
Parameters setting: The parameters of each comparison detection framework used in this paper are consistent with their default parameters without any adjustment.

3.2. Experimental Results and Analysis

Detection results of the proposed method and compared method are shown in Table 2. We can observe that our method can achieve 74.3% mean Recall and 70.1% mean AP, which obtains the improvements of 4.9%, 0.5%, and 2.8% AP compared with FPN [8], S-RPN [26], and Cascade R-CNN [9], respectively. This demonstrates that the deep residual network with deformable and attention-based multi-scale feature pyramid network contributes to the gain of the performance.
Additionally, we further explored the detection accuracy of each category of corn pest. In Table 2, we can see that the precision of class “DP” only has 44.0% AP and 46.3% recall, which is lower than other classes of corn pest. We found that serious occlusion will significantly decrease the recognition accuracy, as shown in Figure 4. In the next work, we will focus on this challenge.

3.3. Detection Efficiency

The detection efficiency is a metric that needs to be taken into consideration in the real application. In this paper, we use three different metrics to verify the efficiency of the proposed method. The result is reported in Table 3. It shows that our method can achieve 17.0 FPS, leading to the improvements of 2.5 and 3.9 FPS compared to S-RPN and Cascade R-CNN, respectively. However, its speed is slightly lower that of the FPN. From the view of GLOPs and the number of parameters, we can observe that it is slightly inferior to the FPN detector.

3.4. Ablation Experiment

We carried out a series of experiments to explore the effect of the deformable convolution and AMFFP-Net. The detection results on the corn pest dataset are reported in Table 4. Here, we take the Faster R-CNN detector with ResNet50 backbone as the baseline. When we introduced deformable convolution instead of conventional into a deep residual network, the mAP rose from 65.2% to 66.3%, implying that the deformable convolution has contributed to object detection. Additionally, when we adopted the proposed AMFFP-Net, the performance of our method could improve by 3.8% mAP, demonstrating that the attention mechanism is useful. From the evaluation metric of Recall, the trend is similar to that of mAP.

3.5. Visualized Detection Results and Analysis

To verify the effectiveness of our proposed method, we also visualized some detection results on the corn pest dataset, as shown in Figure 5. It demonstrates that the proposed method can achieve the accurate recognition and localization of corn pests. However, we also found that there exist some poor results; for example, some corn pests are undetected, as shown in Figure 6. On the left of this figure, the pest is not detected using our method because the pest instance is concealed in the background. This will hinder the precise detection of corn pests. Additionally, we also observed that the dense distribution of small corn pests also causes poor performance in the proposed detector, as shown in the middle and right figures of Figure 6. These problems need to be addressed in future work.

4. Conclusions and Future Work

Due to the complexity of the corn pest image, such as the complex background, the various size of corn pest instances, and so on, these problems bring great challenges for precise corn pest detection. In this paper, we first introduced deformable convolution into a deep residual network and then developed an attention-based multi-scale feature pyramid network for enhancing the feature representation of the corn pest image. Finally, these proposed modules were combined with a two-stage detector to achieve the detection of corn pests. Experimental results on large-scale corn pest datasets show that the detection results of our method outperform other approaches in terms of accuracy and speed. However, in Section 3.2, we see that dense distribution and serious overlapping of corn pests have great influence on the accuracy of detection. Therefore, in future work, we plan to adopt a coarse-to-fine strategy to address the problems of dense distribution and serious overlap.

Author Contributions

Conceptualization, C.K. and L.J.; methodology, C.K.; software, C.K. and L.J.; validation, C.K., L.J. and J.D.; formal analysis, C.K. and L.J.; investigation, C.K. and Z.L.; resources, H.H.; data curation, C.K., L.J. and H.H.; writing—original draft preparation, C.K. and L.J.; writing—review and editing, C.K., L.J. and R.W.; visualization, C.K.; supervision, R.W. and L.J.; project administration, L.J. and R.W.; funding acquisition, L.J. and R.W. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Natural Science Foundation of Anhui Higher Education Institutions of China receipted by L.J. (KJ2021A0025), Intergovernmental International Science and Technology Innovation Cooperation receipted by R.W. (2019YFE0125700), and National Key Research and Development Program receipted by R.W. (2021YFD2000205).

Data Availability Statement

Not applicable.

Acknowledgments

The authors would like to thank all the authors cited in this paper and anonymous referees for their helpful comments and suggestions.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Larios, N.; Deng, H.; Zhang, W.; Sarpola, M.; Yuen, J.; Paasch, R.; Moldenke, A.; Lytle, D.A.; Correa, S.R.; Mortensen, E.N.; et al. Automated insect identification through concatenated histograms of local appearance features: Feature vector generation and region detection for deformable objects. Mach. Vis. Appl. 2008, 19, 105–123. [Google Scholar] [CrossRef]
  2. Faithpraise, F.; Birch, P.; Young, R.; Obu, J.; Faithpraise, B.; Chatwin, C. Automatic plant detection and recogition using K-means clustering algorithm and correspondence filters. Int. J. Adv. Biotechnol. Res. 2013, 4, 189–199. [Google Scholar]
  3. Wen, C.; Guyer, D. Image-based orchard insect automated identification and classification method. Comput. Electron. Agric. 2012, 89, 110–115. [Google Scholar] [CrossRef]
  4. Xie, C.; Wang, R.; Zhang, J.; Chen, P.; Dong, W.; Li, R.; Chen, T.; Chen, H. Multi-level learning features for automatic classification of field crop pests. Comput. Electron. Agric. 2018, 152, 233–241. [Google Scholar] [CrossRef]
  5. Girshick, R.; Donahue, J.; Darrell, T.; Malik, J. Region-Based Convolutional Networks for Accurate Object Detection and Segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 2015, 38, 142–158. [Google Scholar] [CrossRef] [PubMed]
  6. Girshick, R. Fast r-cnn. In Proceedings of the International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; IEEE: Piscataway, NJ, USA, 2015; pp. 1440–1448. [Google Scholar]
  7. Ren, S.Q.; He, K.M.; Girshick, R.; Sun, J. Faster R-CNN: Towards real-time object detection with region proposal networks. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 1137–1149. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  8. Lin, T.-Y.; Dollár, P.; Girshick, R.; He, K.; Hariharan, B.; Belongie, S. Feature pyramid networks for object detection. In Proceedings of the Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; IEEE: Piscataway, NJ, USA, 2017; pp. 2117–2125. [Google Scholar]
  9. Cai, Z.; Vasconcelos, N. Cascade R-CNN: Delving Into High Quality Object Detection. In Proceedings of the Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; IEEE: Piscataway, NJ, USA, 2018; pp. 6154–6162. [Google Scholar]
  10. Redmon, J.; Divvala, S.K.; Girshick, R.; Farhadi, A. You Only Look Once: Unified, Real-Time Object Detection. In Proceedings of the Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016. [Google Scholar]
  11. Redmon, J.; Farhadi, A. YOLO9000: Better, faster, stronger. In Proceedings of the Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; IEEE: Piscataway, NJ, USA, 2017; pp. 6517–6525. [Google Scholar]
  12. Redmon, J.; Farhadi, A. Yolov3: An Incremental Improvement. arXiv 2018, arXiv:1804.02767. [Google Scholar]
  13. Law, H.; Deng, J. CornerNet: Detecting Objects as Paired Keypoints. In Proceedings of the European Conference on Computer Vision, Munich, Germany, 8–14 September 2018. [Google Scholar]
  14. Shelhamer, E.; Long, J.; Darrell, T. Fully Convolutional Networks for Semantic Segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 640–651. [Google Scholar] [CrossRef] [PubMed]
  15. Duan, K.; Bai, S.; Xie, L.; Qi, H.; Huang, Q.; Tian, Q. CenterNet: Keypoint Triplets for Object Detection. In Proceedings of the International Conference on Computer Vision, Seoul, Korea, 27 October–2 November 2019; IEEE: Piscataway, NJ, USA, 2019. [Google Scholar]
  16. Dong, Z.; Li, G.; Liao, Y.; Wang, F.; Ren, P.; Qian, C. CentripetalNet: Pursuing High-Quality Keypoint Pairs for Object Detection. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 10516–10525. [Google Scholar]
  17. Zhou, X.; Zhuo, J.; Krähenbühl, P. Bottom-Up Object Detection by Grouping Extreme and Center Points. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 850–859. [Google Scholar]
  18. Liu, L.; Xie, C.; Wang, R.; Yang, P.; Wang, F. Deep Learning based Automatic Multi-Class Wild Pest Monitoring Approach using Hybrid Global and Local Activated Features. IEEE Trans. Ind. Inform. 2020, 17, 7589–7598. [Google Scholar] [CrossRef]
  19. Wang, J.; Li, Y.; Feng, H.; Ren, L.; Du, X.; Wu, J. Common pests image recognition based on deep convolutional neural network. Comput. Electron. Agric. 2020, 179, 105834. [Google Scholar] [CrossRef]
  20. Rahman, R.; Arko, P.; Ali, M.E.; Khan, M.; Apon, S.H.; Nowrin, F.; Wasif, A. Identification and recognition of rice diseases and pests using convolutional neural networks. Biosyst. Eng. 2020, 194, 112–120. [Google Scholar] [CrossRef] [Green Version]
  21. Jiao, L.; Dong, S.; Zhang, S.; Xie, C.; Wang, H. AF-RCNN: An anchor-free convolutional neural network for multi-categories agricultural pest detection. Comput. Electron. Agric. 2020, 174, 105522. [Google Scholar] [CrossRef]
  22. Dong, S.; Du, J.; Jiao, L.; Wang, F.; Liu, K.; Teng, Y.; Wang, R. Automatic Crop Pest Detection Oriented Multiscale Feature Fusion Approach. Insects 2022, 13, 554. [Google Scholar] [CrossRef] [PubMed]
  23. Teng, Y.; Wang, R.; Du, J.; Huang, Z.; Zhou, Q.; Jiao, L. TD-Det: A Tiny Size Dense Aphid Detection Network under In-Field Environment. Insects 2022, 13, 501. [Google Scholar] [CrossRef] [PubMed]
  24. Shen, Y.; Zhou, H.; Li, J.; Jian, F.; Jayas, D. Detection of stored-grain insects using deep learning. Comput. Electron. Agric. 2018, 145, 319–325. [Google Scholar] [CrossRef]
  25. Selvaraj, M.G.; Vergara, A.; Ruiz, H.; Safari, N.; Elayabalan, S.; Ocimati, W.; Blomme, G. AI-powered banana diseases and pest detection. Plant Methods 2019, 15, 92. [Google Scholar] [CrossRef] [Green Version]
  26. Wang, R.; Jiao, L.; Xie, C.; Chen, P.; Du, J.; Li, R. S-RPN: Sampling-balanced region proposal network for small crop pest detection. Comput. Electron. Agric. 2021, 187, 106290. [Google Scholar] [CrossRef]
  27. Jiao, L.; Xie, C.; Chen, P.; Du, J.; Li, R.; Zhang, J. Adaptive feature fusion pyramid network for multi-classes agricultural pest detection. Comput. Electron. Agric. 2022, 195, 106827. [Google Scholar] [CrossRef]
  28. He, Y.; Zhou, Z.Y.; Tian, L.H.; Liu, Y.F.; Luo, X.W. Brown rice planthopper (Nilaparvata lugens Stal) detection based on deep learning. Precis. Agric. 2020, 21, 1385–1402. [Google Scholar] [CrossRef]
  29. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
Figure 1. Examples of 10-types of corn pest.
Figure 1. Examples of 10-types of corn pest.
Insects 13 00978 g001
Figure 2. Distribution of relative scale of corn pest instances.
Figure 2. Distribution of relative scale of corn pest instances.
Insects 13 00978 g002
Figure 3. (a) Overview of the pipeline proposed corn pest detection module. (b) Deep residual network with deformable convolution block. (c) Architecture of the proposed aFPN. The weights generator is used to produce a set of attentional weights, which is related to the upper layers.
Figure 3. (a) Overview of the pipeline proposed corn pest detection module. (b) Deep residual network with deformable convolution block. (c) Architecture of the proposed aFPN. The weights generator is used to produce a set of attentional weights, which is related to the upper layers.
Insects 13 00978 g003
Figure 4. Examples of serious occlusion problem. The red boxes denote the location of the corn pest.
Figure 4. Examples of serious occlusion problem. The red boxes denote the location of the corn pest.
Insects 13 00978 g004
Figure 5. Some examples of visualized detection results of the proposed method. The red boxes represent the predicted location of corn pest.
Figure 5. Some examples of visualized detection results of the proposed method. The red boxes represent the predicted location of corn pest.
Insects 13 00978 g005
Figure 6. Poor detection results of the proposed method. The red boxes represent the predicted location of corn pest.
Figure 6. Poor detection results of the proposed method. The red boxes represent the predicted location of corn pest.
Insects 13 00978 g006
Table 1. Description of corn pest dataset used in our work.
Table 1. Description of corn pest dataset used in our work.
IDScientific NameNumber of ImagesNumber of Corn Pest InstancesAverage Relative Size
1Leucania loreyi Duponchel (LLD)55550.192
2Ostrinia furnacalis (OF)6296500.042
3Agrotis ypsilon (AY)1461740.153
4Spodoptera litura Fabricius (SLF)266479760.306
5Dichocrocis punctiferalis (DP)7098490.038
6Helicoverpa armigera (HA)9169190.094
7Laodelphax striatellus (LS)1391400.061
8Spodoptera exigua Hiibner (SEH)1311410.048
9Rhopalosiphum padi (RP)24938750.007
10Spodoptera frugiperda (SF)175419700.057
Table 2. Detection results of the proposed method and compared detectors on corn pest dataset.
Table 2. Detection results of the proposed method and compared detectors on corn pest dataset.
ClassFPNS-RPNCascade R-CNNOur Method
RecallAPRecallAPRecallAPRecallAP
LLD83.381.8100100100100100100
OF60.656.467.659.159.251.669.060.0
AY88.274.583.180.188.981.883.381.8
SLF56.049.763.458.355.449.664.158.1
DP48.845.546.044.646.344.746.344.0
HA85.979.480.479.481.578.582.679.6
LS100100100100100100100100
SEH60.054.558.853.658.852.958.853.6
RP61.148.968.458.562.051.470.161.7
SF69.461.468.662.167.062.169.162.5
Mean71.365.273.669.671.967.374.370.1
Table 3. Detection efficiency of the proposed method and other compared methods on a single NVIDIA GPU.
Table 3. Detection efficiency of the proposed method and other compared methods on a single NVIDIA GPU.
MethodSpeed (FPS)GFLOPsNumber of Parameter (M)
FPN18.2216.3441.17
S-RPN14.5241.1246.23
Cascade R-CNN13.1244.1368.95
Ours17.0224.2241.82
Table 4. Ablation experimental results based on the baseline (Faster R-CNN detector).
Table 4. Ablation experimental results based on the baseline (Faster R-CNN detector).
Deformable ConvolutionAMFFP-NetmAPRecall
65.271.3
66.369.5
70.174.3
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Kang, C.; Jiao, L.; Wang, R.; Liu, Z.; Du, J.; Hu, H. Attention-Based Multiscale Feature Pyramid Network for Corn Pest Detection under Wild Environment. Insects 2022, 13, 978. https://doi.org/10.3390/insects13110978

AMA Style

Kang C, Jiao L, Wang R, Liu Z, Du J, Hu H. Attention-Based Multiscale Feature Pyramid Network for Corn Pest Detection under Wild Environment. Insects. 2022; 13(11):978. https://doi.org/10.3390/insects13110978

Chicago/Turabian Style

Kang, Chenrui, Lin Jiao, Rujing Wang, Zhigui Liu, Jianming Du, and Haiying Hu. 2022. "Attention-Based Multiscale Feature Pyramid Network for Corn Pest Detection under Wild Environment" Insects 13, no. 11: 978. https://doi.org/10.3390/insects13110978

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop