Next Article in Journal
Research on Soil Pore Segmentation of CT Images Based on MMLFR-UNet Hybrid Network
Previous Article in Journal
Exploring the Impact of Improved Maize Seeds on Productivity of Tanzanian Family Farms: A Maize Seed Stochastic Simulation (MaizeSim) Approach
Previous Article in Special Issue
Performance Analysis of Real-Time Detection Transformer and You Only Look Once Models for Weed Detection in Maize Cultivation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

LCFANet: A Novel Lightweight Cross-Level Feature Aggregation Network for Small Agricultural Pest Detection

1
Key Laboratory of Micro Nano Optoelectronic Devices and Intelligent Perception Systems, Yangtze Normal University, Chongqing 408100, China
2
CAS Engineering Laboratory for Intelligent Industrial Vision, Institute of Automation, Chinese Academy of Sciences, Beijing 100190, China
*
Author to whom correspondence should be addressed.
Agronomy 2025, 15(5), 1168; https://doi.org/10.3390/agronomy15051168
Submission received: 7 April 2025 / Revised: 1 May 2025 / Accepted: 9 May 2025 / Published: 11 May 2025

Abstract

:
In agricultural pest detection, the small size of pests poses a critical hurdle to detection accuracy. To mitigate this concern, we propose a Lightweight Cross-Level Feature Aggregation Network (LCFANet), which comprises three key components: a deep feature extraction network, a deep feature fusion network, and a multi-scale object detection head. Within the feature extraction and fusion networks, we introduce the Dual Temporal Feature Aggregation C3k2 (DTFA-C3k2) module, leveraging a spatiotemporal fusion mechanism to integrate multi-receptive field features while preserving fine-grained texture and structural details across scales. This significantly improves detection performance for objects with large scale variations. Additionally, we propose the Aggregated Downsampling Convolution (ADown-Conv) module, a dual-path compression unit that enhances feature representation while efficiently reducing spatial dimensions. For feature fusion, we design a Cross-Level Hierarchical Feature Pyramid (CLHFP), which employs bidirectional integration—backward pyramid construction for deep-to-shallow fusion and forward pyramid construction for feature refinement. The detection head incorporates a Multi-Scale Adaptive Spatial Fusion (MSASF) module, adaptively fusing features at specific scales to improve accuracy for varying-sized objects. Furthermore, we introduce the MPDINIoU loss function, combining InnerIoU and MPDIoU to optimize bounding box regression. The LCFANet-n model has 2.78 M parameters and a computational cost of 6.7 GFLOPs, enabling lightweight deployment. Extensive experiments on the public dataset demonstrate that the LCFANet-n model achieves a precision of 71.7 % , recall of 68.5 % , mAP50 of 70.4 % , and mAP50-95 of 45.1 % , reaching state-of-the-art (SOTA) performance in small-sized pest detection while maintaining a lightweight architecture.

1. Introduction

Agriculture serves as the foundation for humanity to sustain production and livelihoods and constitutes the fundamental guarantee for the normal functioning of human society. Ensuring the yield and quality of crops such as grains, fruits, and vegetables has always been an ongoing priority. Plant diseases and insect pests are one of the primary factors that hinder the normal growth of crops and lead to reduced yields. In severe cases, they can even negatively impact the economic development of nations and regions, as well as disrupt the normal livelihoods of people [1]. Therefore, the timely detection of plant diseases and insect pests, coupled with rapid response and intervention, is crucial. Due to the relatively underdeveloped agricultural infrastructure in most countries and regions, manual inspection remains the primary method for detecting plant diseases and insect pests during high-risk seasons. However, due to the high concealment, mobility, and morphological similarity of pests, this approach often leads to missed detection or misidentification, resulting in inappropriate control measures [2]. Such errors not only cause misuse and waste of pesticide resources but also fail to prevent the expansion of agricultural economic losses.
In response to the inherent limitations of traditional manual inspection methods for agricultural pest and disease control, an increasing number of studies are leveraging deep learning approaches to detect crop diseases and insect pests, driven by the widespread adoption of computer vision methods and artificial intelligence technologies across various domains [3]. Convolutional Neural Networks (CNNs), leveraging their profound capability to extract image pixel features and semantic features, have been extensively applied to classification and detection tasks for plant diseases and insect pests [4].
Deep learning-driven target identification and detection in insect imagery is a key focus in agricultural pest control research [3]. YOLO [5,6,7] redefines target detection by treating it as a regression task, offering a groundbreaking approach to real-time detection. Recent advancements in agricultural image analysis have witnessed a proliferation of detection frameworks derived from the YOLO architecture [8,9,10]. The CSF-YOLO framework, developed by Wang et al. [11], enhances YOLOv8-n [12] with a FasterNet backbone to achieve lightweight grape leafhopper damage assessment, supporting real-time vineyard surveillance. Huang et al. [13] proposed YOLO-YSTs, a wild pest detection method based on YOLOv10-n [14]. This approach was experimentally validated on a single-species pest dataset, achieving a mean average precision of 86.8 % . Feng et al. [15] proposed LCDDN-YOLO, a lightweight cotton disease detection model based on improved YOLOv8n. By introducing the BiFPN network and PDS-C2f module, the model enhances detection accuracy while achieving lightweight performance.
R-CNN [16,17,18] pioneered a two-stage framework by combining selective search region proposals with CNN-based feature extraction. Mu et al. [19] proposed a pest detection method called TFEMRNet based on the improved Mask R-CNN [20] model for edge devices. By incorporating a multi-attention mechanism, this method achieved a precision of 96.75 % in pest detection tasks. Li et al. [21] proposed an LLA-RCNN method, which achieved partial lightweighting of the Faster R-CNN [17] model by incorporating MobileNetV3 [22]. This approach enables accurate detection of pests on leaf surfaces. Guan et al. [23] proposed a GC-Faster RCNN agricultural pest detection model based on Faster R-CNN and a hybrid attention mechanism, achieving a 4.5 % improvement in average precision on the Insect25 dataset. Although the aforementioned agricultural pest detection models can achieve accurate insect identification, their reliance on the two-stage detection principle of R-CNN imposes significant barriers to breakthroughs in lightweight optimization.
Transformer-based detectors leverage self-attention mechanisms to model global contextual relationships, breaking convolutional inductive biases for improved scale adaptability [24,25,26]. Liu et al. [27] proposed an end-to-end pest object detection method based on the transformer, which achieved 95.3 % mAP in leaf-surface pest detection tasks by incorporating FRC and RPSA mechanisms. To achieve accurate pest detection in cistanche, Zhang et al. [28] proposed a transformer-based target identification module. Enhanced by a bridge attention mechanism and loss function, the network achieved 92 % average precision, demonstrating excellent results in complicated agricultural scenes. TinySegFormer [29] tackles edge computing limitations in agricultural pest detection by hybridizing the transformer and CNN architectures. It attained a 92 % segmentation precision at 65 FPS, demonstrating real-time applicability for leaf pest analysis. However, existing studies exclusively utilize near-field leaf-surface pest imagery, creating artificially magnified pest features that misrepresent actual field conditions characterized by smaller pest-to-image ratios.
Our analysis of the existing research indicates that current object detection methods face significant challenges in accurately identifying small-scale agricultural pests, primarily due to the following limitations:
(1)
Small pests often exhibit a low resolution and high morphological similarity, leading to a weak feature extraction and misclassification. Standard CNNs struggle to balance large receptive fields and fine-grained details required for small pests, resulting in poor localization [30].
(2)
To enhance small-scale object detection accuracy, current research predominantly adopts network parameter scaling strategies. However, increasing parameters significantly slows down inference speed. Real-time constraints in agricultural drones or edge devices make such models impractical for field deployment.
To overcome the above challenges in small pest imagery, we present the Lightweight Cross-Level Feature Aggregation Network (LCFANet). This framework integrates three core mechanisms: a deep feature extraction network that progressively converts RGB inputs from basic spatial details to advanced semantic information through hierarchical mapping, yielding four distinct feature maps; a hierarchical feature pyramid architecture employing top–down context propagation with lateral detail retention and adaptive feature refinement for cross-scale fusion; and a scale-aware detection head with trainable spatial weighting that achieves scale-invariant object recognition across all pyramid levels simultaneously. The integrated architecture maintains computational efficiency while addressing the characteristic challenges of small pest targets. The core contributions and innovative elements are detailed as follows:
(1)
LCFANet is proposed to achieve small pest object detection in insect trapping devices. Extensive evaluations on public Pest24 and our private datasets demonstrated its capability to accurately classify similar-feature pests and reliably detect objects with a small scale and debris, ultimately achieving SOTA detection performance.
(2)
We propose a Dual Temporal Feature Aggregation C3k2 (DTFA-C3k2) architecture, a Cross-Level Hierarchical Feature Pyramid (CLHFP), and a Multi-Scale Adaptive Spatial Fusion (MSASF) method in LCFANet. By integrating spatial-channel fusion, hierarchical feature interactions, and adaptive multi-scale fusion, these approaches maintain critical fine-grained details and structural hierarchies of small pests, significantly enhancing detection precision in agricultural scenarios.
(3)
To address the challenge in model lightweighting, we incorporate the Aggregated Downsampling Convolution (ADown-Conv) module into both the feature extraction and fusion stages of the LCFANet. ADown-Conv provides an effective downsampling solution for real-time object detection models through its lightweight design and structural flexibility.
The following sections are organized as follows: Section 2 details the pest dataset provenance and the architectural design of LCFANet. Section 3 evaluates the performance of LCFANet in small-scale insect object detection through generalization, comparative, and ablation experiments. Section 4 provides critical discussions on the proposed approach. Finally, the conclusion section summarizes the key findings and contributions of this paper.

2. Materials and Methods

2.1. Construction of Agricultural Pest Image Datasets

In this paper, to validate the effectiveness and generalization capability of our proposed pest detection model, we prepared a private dataset and a public dataset Pest24 [31]. Both datasets were collected using pest traps.
The image acquisition method and sample images of our private dataset are shown in Figure 1. We deployed self-powered pest traps equipped with image-capturing capabilities in an apple orchard located in Lingbao City, Henan Province, China. These traps utilized adhesive boards coated with different types of lures to capture three pest species: Grapholita molesta, Anarsia lineatella, and Cydia pomonella. Simultaneously, images of the adhesive boards were systematically captured. A total of 1246 images were obtained, of which 945 were validated as effective samples. Effective image samples contain pest images that need to be detected, while invalid images do not contain any target pest image information. Among the 945 valid images, there were 321 Grapholita molesta pest trap images, 307 Anarsia lineatella pest trap images, and 317 Cydia pomonella pest trap images. The average size of small-target pests is 48 pixels. The quantities of these three categories of pest trap images are essentially balanced. We used the LabelImg software to annotate these images in YOLO format, with each image corresponding to a TXT file containing the coordinates of the object detection bounding boxes.
The Pest24 dataset consists of 25 , 378 pest trap images with a resolution of 600   ×   800 , as illustrated in Figure 2. Among these images, the average size of small-target pests is 25 pixels. This dataset includes images of 24 pest categories, such as Rice planthopper, Meadow borer, and Holotrichia oblita, all annotated in YOLO format.
As shown in Figure 1 and Figure 2, the pests captured in both our private dataset and the Pest24 dataset exhibit notably small sizes, which aligns with the characteristics of pest images in field scenarios. Therefore, we utilized these two datasets as our research subjects and divided them into training, validation, and test sets at a ratio of 6:2:2.

2.2. Construction of the Lightweight Cross-Level Feature Aggregation Network

In practical agricultural scenarios, small-target pests pose significant challenges to object detection models due to their small pixel area and subtle inter-species feature differences. Meanwhile, since the timeliness of pest detection is crucial for implementing emergency measures and minimizing crop yield loss, the lightweight edge deployment of pest detection models is equally important. Therefore, the current research priority lies in improving the accuracy of small-sized, multi-target pest detection models while maintaining their lightweight design. To mitigate these difficulties, we developed the Lightweight Cross-Level Feature Aggregation Network (LCFANet), an enhanced framework derived from YOLOv11 [32] architecture, specifically designed for small-scale imagery analysis.
As depicted in Figure 3, the LCFANet framework integrates three core modules specifically designed for small-scale pest image target detection. The deep feature extraction network processes RGB inputs through convolutional operations, progressively transforming shallow spatial patterns (e.g., colors, edges, and textures) into high-level semantic representations, generating four hierarchical feature maps at resolutions of 80   ×   80   ×   64 , 40   ×   40   ×   128 , 30   ×   30   ×   192 , and 20   ×   20   ×   256 . These multi-scale features are subsequently fed into a hierarchical feature pyramid architecture that implements cross-resolution fusion through top–down contextual propagation, lateral detail preservation connections, and adaptive feature recalibration. Finally, the scale-aware detection head incorporates spatial fusion mechanisms with learnable weights, enabling simultaneous object detection across four pyramid levels while maintaining scale invariance. The architectural details and innovations of each module are detailed in the following sections.

2.2.1. Deep Feature Extraction Network

LCFANet generates four hierarchically structured feature maps through its image feature extraction backbone. These multi-level representations comprehensively integrate both shallow spatial patterns and deep semantic attributes of targets, establishing the foundation for robust small-scale target detection.
To counteract feature deterioration caused by network depth escalation while enhancing feature utilization efficiency, we propose the C3k2 module with Dual Temporal Feature Aggregation [33] capability (DTFA-C3k2). This architecture synergistically combines multi-receptive-field features through spatial-temporal fusion mechanisms, preserving critical fine-grained textures and structural details across scales, thereby improving detection accuracy for objects with significant scale variations.
Considering the increased computational complexity from deepened networks, we introduce an Aggregated Downsampling Convolution (ADown-Conv) [34] module with a lightweight architecture. ADown-Conv achieves an optimal balance between parameter efficiency and learning capacity, significantly enhancing operational performance while maintaining precision, which facilitates deployment in resource-constrained scenarios.
A. DTFA-C3k2 Module
In YOLOv11, the C3k2 mechanism serves as the core feature extraction methodology. This module integrates two cascaded bottleneck submodules that perform feature learning and propagation through streamlined convolutional operations and element-wise addition. However, the progressive feature degradation caused by low-level detail loss in bottleneck modules becomes inevitable with the deepened network layers of LCFANet, particularly constraining detection accuracy improvement for small-scale pest image targets.
To mitigate this drawback, we propose a DTFA-C3k2 architecture comprising two serially connected DTFA-C3k modules, which is depicted in Figure 4. Input features are processed by the DTFA-C3k module through a structured pipeline: the initial 1   ×   1 convolution adjusts the channel dimensions before splitting into dual branches. The primary branch undergoes two sequential 1   ×   1 convolutions to generate the refined feature f i , while the parallel branch preserves the split-originated feature F i . Both outputs maintain identical spatial-channel dimensions ( H   ×   W   ×   C ), forming bi-temporal feature layers that synergistically preserve spatial details while enabling multi-scale representation learning. The feature map transformation is as follows:
F i = S p l i t ( C o n v 1   ×   1 ( X ) ) ,
f i = C o n v 1   ×   1 ( C o n v 1   ×   1 ( F i ) ) ,
where X R H   ×   W   ×   C denotes input features. By concatenating the input feature pair ( F i , f i ) along the channel dimension, we construct a hierarchical perception feature extraction framework. The architecture employs four parallel dilated grouped convolution modules with 3   ×   3 kernels and differentiated dilation rates d { 1 , 2 , 3 , 4 } , where c denotes the input channel count. The gradient variation of dilation rates establishes multi-level receptive fields, enabling shallow branches to focus on local textural details while deeper branches capture wide-range contextual semantics, forming complementary feature representations. The incorporation of grouped convolution reduces computational complexity while preserving the spatial topological structure through channel-wise independence constraints, effectively avoiding information loss caused by conventional downsampling. The calculation is as follows:
f 3   ×   3 d , c = C o n v 3   ×   3 d , g ( C o n c a t ( F i , f i ) ,
f c o n c a t = C o n v 1   ×   1 ( C o n c a t ( f c 1 , f c 2 , f c 3 , f c 4 ) ) .
Subsequent processing utilizes a 1   ×   1 convolution for the channel dimension reduction of fused multi-scale features, coupled with the adaptive feature enhancement capability of the SimAM [35] attention mechanism. Leveraging its parameter-free attention computation based on the energy function E = 1 H W C i , j , k x i , j , k 2 , this mechanism dynamically amplifies responses in regions with significant variations. The entire pipeline maintains the original spatial resolution while balancing the multi-scale feature synergy and computational efficiency, establishing robust feature representation foundations for pest target object detection in complex scenarios. After extracting the commonality features from dual temporal features ( F i , f i ) , we perform multiplication between these commonality features and the original temporal features to quantify their similarity. This operation can be mathematically represented as follows:
F i ^ = S i m A M ( F i )   ×   S i m A M ( f c o n c a t ) ,
f i ^ = S i m A M ( f i )   ×   S i m A M ( f c o n c a t ) .
Finally, we integrate the components ( F i , f i , F ^ i , f ^ i , S i m A M ( f c o n c a t ) ) through element-wise summation, followed by convolutional operations coupled with the SimAM attention mechanism:
f D T F A C 3 k i = S i m A M ( C o n v 3   ×   3 ( ( F i + f i + F ^ i + f ^ i + S i m A M ( f c o n c a t ) ) .
The DTFA-C3k architecture optimizes feature utilization through dimension-consistent transformations, particularly crucial for pest target image targets requiring joint low-level texture and high-level semantic retention.
B. ADown-Conv Module
In LCFANet, the integration of the deep feature extraction network, Cross-Level Hierarchical Feature Pyramid (CLHFP), and scale-aware detection head results in increased model depth, parameter count, and computational complexity. To address this challenge in model lightweighting, we strategically incorporate ADown-Conv [34] modules into both the deep feature extraction network and CLHFP in LCFANet.
The ADown-Conv module is a dual-path compression unit designed to achieve efficient spatial reduction while enhancing feature representation. It comprises two complementary processing streams: the Context Preservation Path and the Salient Feature Path. The Context Preservation Path processes half channels through 3   ×   3 convolution. The Salient Feature Path extracts key features via max pooling and point-wise convolution. The final output combines both paths through channel concatenation. In Algorithm 1, ADown-Conv processes input X R C i n   ×   H   ×   W through parallel pooling-convolution paths, producing enriched output features X R C o u t   ×   H n   ×   W n .
Algorithm 1 ADown-Conv Forward Propagation
  • Require: Input feature map X R C i n   ×   H   ×   W
  • Ensure: Downsampled feature X R C o u t   ×   H n   ×   W n {here, n = 2}
      1:
    X AvgPool 2 d ( X , kernel _ size = 2 ) {Step 1: Average pooling}
      2:
    ( X 1 , X 2 ) chunk ( X , 2 , dim = 1 ) {Step 2: Channel splitting}
      3:
    X 1 Conv 3   ×   3 s t r i d e = 2 ( X 1 ) {The Context Preservation Path: 3   ×   3 convolution}
      4:
    X 2 MaxPool 2 d ( 3 , stride = 2 ) ( X 2 ) {The Salient Feature Path: Max pooling}
      5:
    X 2 Conv 1   ×   1 ( X 2 ) {Point-wise convolution}
      6:
    X concat ( X 1 , X 2 , dim = 1 ) {Feature fusion}
      7:
    return  X
ADown-Conv decomposes the standard convolution into parallel sub-operations, which reduces the parameters as follows:
P a r a m = 3 2 C i n C o u t 4 ( 3 2 ( C i n / 2 ) ( C o u t / 2 ) 4 + 1 2 ( C i n / 2 ) ( C o u t / 2 ) 4 ) ,
This yields parameter reduction compared to conventional approaches. The FLOPs complexity is optimized as follows:
O ADown - Conv = H W 4 9 C in C out 4 The Context Preservation Path + H W 4 C in C out 4 The Salient Feature Path ,
maintaining lower computations than standard strided convolutions.
In summary, ADown-Conv provides an effective downsampling solution for real-time object detection models through its lightweight design and structural flexibility. By integrating the ADown-Conv module into LCFANet, it achieves a significant reduction in parameter count while maintaining object detection accuracy.

2.2.2. Cross-Level Hierarchical Feature Pyramid

The deep feature extraction network hierarchically processes input RGB images through multi-scale learning, generating four-level feature map vectors (Level 1–4) that progressively evolve from shallow pixel patterns to deep semantic representations. Specifically, Level 1 features predominantly retain low-level visual cues like edge details and texture patterns, while Level 4 features emphasize high-level semantic understanding through contextual relationships and object part configurations.
We subsequently construct a Cross-Level Hierarchical Feature Pyramid (CLHFP). The architecture implements a bidirectional feature integration process through backward pyramid construction for deep-to-shallow fusion and forward pyramid construction for feature refinement, as depicted in Figure 3. The backward pyramid is initiated with Level 4’s 20   ×   20   ×   256 deep features, which undergo upsampling and concatenation with Level 3’s 30   ×   30   ×   192 features before DTFA-C3k2 processing to generate 30   ×   30   ×   192 outputs. This process sequentially repeats through Level 2 and Level 1, ultimately producing Scale 1’s 80   ×   80   ×   64 features through progressive upsampling and fusion. The forward pyramid is then activated by downscaling Scale 1 through ADown-Conv to 40   ×   40   ×   64 , which combines with the backward pyramid’s 80   ×   80   ×   128 residual features via concatenation and DTFA-C3k2 refinement to create Scale 2 of 40   ×   40   ×   64 . This pattern continues through subsequent scales: Scale 2’s downscaled 30   ×   30   ×   128 features merge with the backward pyramid’s 30   ×   30   ×   192 components to form Scale 3 of 30   ×   30   ×   128 , while Scale 3 integrates with the original Level 4 inputs to finalize Scale 4 of 20   ×   20   ×   256 . The complete multi-scale representation subsequently feeds into a scale-aware detection head for coordinated object recognition tasks.
The CLHFP architecture demonstrates substantial advancements in small-scale pest target image target detection through three principal contributions. Firstly, it implements bidirectional cross-resolution feature propagation, where the backward pyramid facilitates deep-to-shallow semantic enrichment through successive upsampling and concatenation operations, while the forward pyramid achieves feature consolidation via ADown-Conv-driven hierarchical downsampling. Secondly, the integration of DTFA-C3k2 modules at each fusion stage enables adaptive feature recalibration, effectively suppressing redundant activation while optimizing feature utilization. Thirdly, the architecture establishes a multi-granularity representation hierarchy through four geometrically progressive feature maps, which synergistically preserve high-frequency texture details and contextual semantics. The CLHFP permits the optimization of detection sensitivity for small-scale targets.

2.2.3. Multi-Scale Adaptive Spatial Fusion Detection Head

CLHFP addresses varying-scale feature fusion challenges. While feature pyramids remain fundamental in modern detection frameworks, they exhibit inherent limitations in one-stage detectors (e.g., YOLO, SSD) due to inter-scale feature conflicts. Conventional implementations employ heuristic-based feature assignment: high-level features (with a lower resolution but broader receptive fields) detect large objects, while low-level features (higher resolution with detailed textures) handle small targets. However, this rigid allocation creates conflicting gradient signals when processing images containing mixed-scale objects: regions designated as positive samples at one pyramid level are simultaneously treated as background at adjacent levels. The conflict becomes particularly pronounced in our architecture’s four-scale feature outputs, where overlapping receptive fields across pyramid levels amplify contradictory feature responses. This inter-layer discord disrupts gradient computation during backpropagation, ultimately compromising model convergence and detection accuracy.
To address multi-scale feature conflicts in feature pyramids within single-shot detectors, the ASFF method is proposed [36]. This approach enhances scale-invariant feature representation through learnable spatial weighting mechanisms, where features from different pyramid levels are adaptively combined based on their discriminative power at each spatial location. The ASFF framework dynamically suppresses conflicting signals while amplifying consensus patterns across scales, effectively resolving the inherent gradient interference caused by heuristic-based feature assignment in conventional pyramid architectures.
In this section, a Multi-Scale Adaptive Spatial Fusion (MSASF) module is presented, designed specifically for the four-scale feature maps generated by CLHFP, as illustrated in Figure 3. The MSASF module at level i ( i 1 , 2 , 3 , 4 ) performs varying-scale feature fusion through resolution-adaptive resampling and spatially weighted integration. This process combines the feature map F i with two adjacent-scale features via dual-path alignment: (1) for upsampling, the channel dimensions are first compressed by a 1   ×   1 convolutional layer to match F i ’s channel count C i , followed by bilinear interpolation to achieve resolution parity. (2) Downsampling employs scale-specific operators: a 3   ×   3 max-pooling layer (stride = 1) for 2 / 3 scaling ratios, a 2-stride 3   ×   3 convolution for 1 / 2 ratios, and a 4   ×   4 max-pooling layer (stride = 4) for 3 / 4 ratios, each simultaneously adjusting channel dimensions and spatial resolution. Crucially, the module learns spatial attention maps through depth-wise separable convolutions and sigmoid activation, enabling pixel-wise adaptive fusion weights across scales. This mechanism dynamically balances contributions from different pyramid levels ( 80   ×   80 to 20   ×   20 resolutions) while resolving channel-depth disparities (64 to 256 channels) inherent in LCFANet’s hierarchical features. The fused output maintains discriminative power for both high-frequency textures and contextual semantics through this resolution-aware blending paradigm.
Next, we introduce the adaptive fusion strategy, as illustrated in Figure 5. For MSASF-1 to MSASF-4, the fusion formulas are as follows:
M 1 = F 1 1 · α 1 + F 2 1 · β 1 + F 3 1 · γ 1 ,
M 2 = F 1 2 · α 2 + F 2 2 · β 2 + F 3 2 · γ 2 ,
M 3 = F 2 3 · α 3 + F 3 3 · β 3 + F 4 3 · γ 3 ,
M 4 = F 2 4 · α 4 + F 3 4 · β 4 + F 4 4 · γ 4 ,
where M i corresponds to the output of MSASF-i, and F j i represents the feature map transferred from Scale j to MSASF-i through resizing. α i ,   β i , and γ i denote the spatial adaptive weights for F j i at three different scales to M i . α i , β i , and γ i are obtained as follows:
α i + β i + γ i = 1 , α i , β i , γ i [ 0 , 1 ] ,
α i = e λ α i e λ α i + e λ β i + e λ γ i , β i = e λ β i e λ α i + e λ β i + e λ γ i , γ i = e λ γ i e λ α i + e λ β i + e λ γ i ,
where the parameters α i , β i , and γ i are derived through the softmax normalization of their corresponding control parameters λ α i , λ β i , and λ γ i . Specifically in the MSASF-1 module, these control parameters are generated via dedicated 1   ×   1 convolutional layers that process the input feature maps F 1 1 , F 2 1 , and F 3 1 , respectively, forming learnable weight scalars that undergo optimization through standard backpropagation during network training.

2.2.4. MPDINIoU Loss Function

The improvement of the localization accuracy for small objects remains a key objective in loss function design. Conventional localization loss functions, such as IoU loss and CIoU loss, lack specific considerations for tiny targets. To address this problem, we introduce the concepts of InnerIoU [37] and MPDIoU [38] and propose the MPDINIoU loss function. MPDINIoU is depicted in Figure 6 through a holistic integration of spatial overlap metrics, centroid displacement analysis, and dimensional variance assessment, with particular emphasis on inter-region intersection characteristics. The proposed approach significantly improves LCFANet’s capability for detecting diminutive objects in trap imagery.
The MPDINIoU dynamically scales the auxiliary detection region through an adaptive scaling γ , facilitating reliable small-scale target detection. The scaling coefficient γ is empirically constrained within the interval [ 0.5 , 1.5 ] . Values exceeding 1 trigger the generation of an expanded reference frame surpassing the ground-truth boundary, thereby extending the effective optimization space for low IoU instances. Conversely, sub-unity scaling induces a contracted detection window, which amplifies the gradient magnitude during backpropagation through intensified loss computation compared to conventional IoU metrics. The localization precision is further refined by incorporating spatial deviation metrics derived from coordinate offsets (denoted as d 1 , d 2 ) between corresponding vertices of the primary and auxiliary bounding boxes. The MPDINIoU is shown below:
b l g t = x c g t w g t   ×   γ 2 , b r g t = x c g t + w g t   ×   γ 2 , b t g t = y c g t h g t   ×   γ 2 , b b g t = y c g t + h g t   ×   γ 2 ,
b l = x c w   ×   γ 2 , b r = x c + w   ×   γ 2 , b t = y c h   ×   γ 2 , b b = y c + h   ×   γ 2 ,
I n n e r = ( m i n ( b r g t , b r ) m a x ( b l g t , b l ) )   ×   ( m i n ( b b g t , b b ) m a x ( b t g t , b t ) ) ,
I o U i n n e r = I n n e r ( w g t   ×   h g t + w   ×   h )   ×   γ 2 I n n e r ,
where { b l , b r , b t , b b , x c , y c } denote the boundary coordinates and the center point of the inner anchor box, while { b l g t , b r g t , b t g t , b b g t , x c g t , y c g t } correspond to the inner ground truth box.
The input image possesses dimensional parameters h i   ×   w i , where h i and w i denote its vertical and horizontal extents, respectively. The ground-truth annotation is defined by its top-left vertex coordinates ( x t l g t , y t l g t ) and bottom-right vertex ( x b r g t , y b r g t ) , while the predicted bounding region is characterized by the initial anchor point ( x t l p , y t l p ) and terminal anchor point ( x b r g t , y b r g t ) . The MPDINIoU loss function is as follows:
d 1 = ( x t l p x t l g t ) 2 + ( y t l p y t l g t ) 2 , d 2 = ( x b r p x b r g t ) 2 + ( y b r p y b r g t ) 2 ,
M P D I o U = I o U d 1 2 + d 2 2 h i 2 + w i 2 ,
L M P D I o U = 1 M P D I o U ,
L M P D I N I o U = L M P D I o U + I o U I o U i n n e r .

3. Results

3.1. Model Training Parameters and Evaluation Metrics

We conducted comparative experiments, ablation studies, and generalization validation experiments in this section. All experiments were implemented on an IW4210-8G server with 8 NVIDIA GeForce RTX 2080 Ti GPUs running Ubuntu 16.04 LTS.
To assess LCFANet’s effectiveness in small-scale pest image target detection, we performed extensive comparative analyses against leading real-time detection methods. The evaluation framework employed key metrics: precision, recall, mAP50 (mean Average Precision at a 50 % IoU threshold), and mAP50-95 (averaged across IoU thresholds from 50 % to 95 % in 5 % increments). The central objective was to systematically analyze the balance between detection accuracy and computational resource requirements, demonstrating LCFANet’s superior performance characteristics in practical agricultural pest detection applications. The calculation formulas are as follows:
P r e c i s i o n = T P T P + F P ,
where T P denotes that the model correctly predicts the positive class, and F P denotes that the model incorrectly predicts the positive class.
R e c a l l = T P T P + F N ,
where F N denotes that the model fails to detect an actual positive instance.
m A P = 1 n i = 1 n A P i ,
where n signifies the total number of object classes, and A P i stands for the average precision of the i-th category. Specifically, each A P i corresponds to the area beneath the precision–recall curve generated for its respective class, calculated by integrating the precision values across all recall levels for that category.
A P i = 0 1 P i d R i ,
where P i signifies the precision of class i corresponding to recall R i .
The params quantifies the model’s scale by computing total trainable parameters. The FLOPs measures the computational workload through floating-point operations per inference cycle, serving as an indicator of hardware resource demands.

3.2. Detection Performance and Generalization Experiments of LCFANet

The LCFANet model, an enhanced iteration of YOLOv11, includes 5 variants with escalating parameter sizes as well as computational demands (GFLOPs). Since model lightweighting is a primary focus of our study, we selected the LCFANet-n variant (the version with the smallest Params and GFLOPs) as our chosen architecture. We trained the LCFANet-n model separately on both our private dataset and the Pest24 dataset, with representative testing results illustrated in Figure 7 and Figure 8. The detection results of LCFANet-n are presented in Table 1. The detection performance of the three pest species on our private dataset is shown in Table 2.

3.3. Comparison Experiments

To validate the performance of LCFANet, we conducted comparative experiments divided into two components: one focusing on benchmarking against SOTA generic object detection models, and the other comparing with SOTA lightweight agricultural pest image detection models. In the comparative experiments with general object detection models, we established a benchmarking framework involving YOLOv5-v11, SSD [39], Faster R-CNN, as well as RT-DETR [40]. These models were evaluated against LCFANet across their respective variants categorized by ascending order of parameter counts. Comprehensive cross-version comparisons were conducted under equivalent computational constraints. The comparison experiments were validated on the Pest24 dataset. The results are presented in Figure 9 and Table 3. To further confirm the exceptional performance of LCFANet, we performed comparative experiments on the Pest24 dataset, evaluating it against 4 state-of-the-art lightweight pest detection models, as shown in Table 4. To visually illustrate the detection results of LCFANet-n, we provide selected detection result examples in Figure 10.

3.4. Ablation Experiments

To achieve accurate detection of small-sized pest image targets, we introduced the DTFA-C3k2 module, CLHFP, and MSASF module in the LCFANet detection model, while proposing an MPDINIoU loss function. To realize model lightweighting, the ADown-Conv module was incorporated. Ablation studies were performed to verify the efficacy of these functional modules, with outcomes detailed in Table 5.
Ablation experiments were carried out on the LCFANet-n model, separated into two parts. The initial phase examines the effect of each innovation on model performance, including model parameters, computational costs, and average precision. In Table 5, the performance parameters of each innovation are compared with those of YOLOv11-n (without any innovations), with improvements highlighted in red and degradations marked in green. The second part analyzes the combined effects of innovations by incrementally adding components and comparing results with/without each added module. Performance improvements are highlighted in red, while degradations are marked in green.

4. Discussion

Timely detection and intervention of pests are prerequisites for effective pest control and agricultural loss reduction. To achieve this, an increasing number of studies employ deep learning methods to process images captured by pest traps or mobile devices, thereby identifying pest species and quantities to guide control measures. Wang et al. [11] enhanced the YOLOv8 network, proposing the CSF-YOLO detection model. Feng et al. [15] developed the LCDDN-YOLO model based on YOLOv8. Huang et al. [13] introduced YOLO-YSTs using YOLOv10, improving detection accuracy while achieving lightweight deployment via a modified YOLOv10n. Our analysis reveals that customized adaptations of high-performing generic detection models are viable for addressing challenges in pest imagery, such as small target sizes, debris clutter, overlapping pests, and inter-species similarity.
Therefore, we selected the currently best-performing YOLOv11 model and proposed an LCFANet model based on YOLOv11 to address challenges in pest trap images such as abundant debris, overlapping pests, small pest targets, multiple pest species, and similar image features among different species. The model improves pest detection accuracy through the proposed DTFA-C3k2 module, CLHFP, MSASF module, and MPDIoU loss function, while achieving lightweight deployment via the introduction of the ADown-Conv module.

4.1. Experimental Results’ Discussion

We describe the conducted detection performance and generalization experiments, comparison experiments, and ablation experiments in Section 3.
In the detection performance and generalization experiments, as illustrated in Figure 7 and Table 2, the LCFANet-n model demonstrates a precise detection of three small-target pest species in trap images. On a test set of 189 images containing 2985 detection instances, LCFANet-n achieves a precision of 92.4%, a recall of 91.7%, an mAP50 of 86.8%, and an mAP50-95 of 54.7%. On the Pest24 dataset containing 24 pest categories, the LCFANet-n model also demonstrates robust detection performance, as shown in Figure 8. Evaluated on a test set of 5075 images with 38,666 detection instances, LCFANet-n achieves a precision of 71.7 % , a recall of 68.5 % , an mAP50 of 70.4 % , and an mAP50-95 of 45.1 % . From the detection results on both public and private datasets, we observe that the LCFANet-n model demonstrates strong performance in small-sized, multi-category agricultural pest detection tasks, exhibiting robust generalization capability.
In the comparison experiments, we conducted experiments on both private datasets and the public Pest24 dataset, demonstrating LCFANet’s excellent generalization performance across different datasets. As illustrated in Figure 9, where model Params and GFLOPs are plotted on the horizontal axes against detection precision metrics (mAP50 and mAP50-95) on the vertical axes, LCFANet consistently outperforms baseline models across all model scales. Table 3 provides a detailed comparison of params, GFLOPs, precision, recall, mAP50, and mAP50-95 for all evaluated models. The lightweight LCFANet-n variant demonstrates a competitive efficiency with 2.78 M parameters ( 0.79 M higher than YOLOv9-t, the most parameter-efficient baseline) and 6.7 GFLOPs ( 0.8 higher than YOLOv5-n, the lowest-computation baseline), while achieving significant accuracy improvements of 8.2 % mAP50 and 8.8 % mAP50-95 over YOLOv11-n, the best-performing lightweight baseline. Across all model variants, LCFANet maintains consistent performance superiority, particularly with the LCFANet-m attaining SOTA detection precision at a 74.4 % mAP50 and a 48.8 % mAP50-95, effectively balancing computational demands and detection precision in small-pest detection scenarios.
To further confirm the exceptional performance of LCFANet, we performed comparative experiments on the Pest24 dataset, evaluating it against 4 state-of-the-art lightweight pest detection models, as shown in Table 4. As evidenced by Table 4, the LCFANet-n model achieves optimal lightweight design while outperforming all 4 SOTA models in detection accuracy. Specifically, it demonstrates a 3.6 % improvement in mAP50 over the top-performing model LLA-RCNN and a 5.7 % improvement in mAP50-95 over CSF-YOLO, establishing its effectiveness in balancing computational efficiency and precision for small-pest detection applications. The LCFANet-n model achieves SOTA performance in agricultural small-pest image target detection, demonstrating the highest precision, recall, and mAP among comparative models.
To visually illustrate the detection results of LCFANet-n, we provide selected detection result examples in Figure 10. It can be observed that in small-pest image target detection tasks, the LCFANet-n model performs the best. For small-sized target detection issues, the LCFANet-n model can accurately detect targets, while other comparison models exhibit missed detections. In detecting targets with similar image features, the LCFANet-n models can accurately distinguish between different pest species, whereas other comparison models show misclassifications. For partially occluded targets, LCFANet-n achieves accurate detection, while other models fail to detect it. These results further confirm that the LCFANet-n model demonstrates superiority in small-pest image target detection tasks; achieves precise classification of targets with similar image features, such as the shape, size, color, and texture of pest images; as well as effectively handles complex scenarios such as occlusions.
In ablation experiments, by analyzing Table 5, it can be concluded that the DTFA-C3k2 module, CLHFP, and MSASF module all contribute to improved model accuracy, thereby demonstrating their positive effects in integrating shallow pixel features with deep semantic features and enhancing feature utilization. Among these, CLHFP demonstrated the most significant improvement, increasing the mAP50 by 4.2 % and mAP50-95 by 3.6 % . However, their introduction also led to increased model parameters and computational costs. To address this, we incorporated the lightweight ADown-Conv module, which reduced the parameters by approximately 23 % and computational costs by approximately 15 % . While lightweighting initially caused a slight accuracy decline, the detection performance saw further improvement with the MPDINIoU loss function. These ablation studies confirm the proposed modules’ effectiveness in improving both precision and model efficiency.

4.2. Limitations and Future Works

Although the proposed method demonstrates good accuracy and lightweight performance for small-target pest detection, some limitations remain. Firstly, the growth rate and trend of pest populations are key indicators for assessing infestation severity. However, our current method can only detect pest categories and counts in single images, without the capability to analyze population dynamics. In future work, we will investigate temporal patterns in pest trap images to establish statistical models, enabling precise monitoring and management of pest outbreaks by integrating ecological patterns. Secondly, current small-target pest detection relies on supervised deep learning requiring substantial training samples for accuracy. However, in practical agricultural scenarios, some pest species occur infrequently, creating significant challenges for sample collection. Nevertheless, monitoring these rare pests remains essential. Therefore, we will explore semi-supervised and unsupervised deep learning approaches to achieve accurate identification and detection for pest species with limited training samples.

5. Conclusions

The development of high-precision and computationally efficient models for agricultural small-target pest detection remains a critical research focus. This paper presents LCFANet, an innovative deep learning framework specifically designed for small agricultural pest detection, which addresses three key challenges: the accurate identification of small-sized targets under complex imaging conditions, precise classification of visually similar species, and robust detection of partially occluded specimens. To address these challenges, this paper proposes LCFANet, a lightweight small-target agricultural pest detection model, with the following key contributions:
(1)
The proposed DTFA-C3k2 module effectively handles scale variations through the spatiotemporal fusion of multi-receptive-field features, preserving cross-scale fine-grained textures and structural details to enhance detection accuracy.
(2)
The CLHFP architecture systematically reorganizes multi-scale feature representations from the backbone network, overcoming challenges posed by target scale diversity and complex background interference.
(3)
The MSASF module implements learnable adaptive fusion of four-scale features from CLHFP, significantly improving detection performance.
(4)
The ADown-Conv modules in both feature extraction and fusion networks achieve model compression, reducing parameters by 23 % and computational costs by 15 % while maintaining detection accuracy.
Extensive experiments on both our proprietary dataset and the public Pest24 benchmark demonstrate LCFANet’s superior performance. Compared with existing pest detectors, our model achieved state-of-the-art results with 3.6 % and 5.7 % improvements in mAP50 and mAP50-95, respectively, while maintaining an efficient architecture ( 2.78 M parameters, 6.7 GFLOPs). These advancements position LCFANet as an optimal solution for real-time agricultural pest monitoring applications.

Author Contributions

Conceptualization, S.H. and Y.T. (Yunong Tian); methodology, S.H. and Y.T. (Yunong Tian); validation, Z.L.; investigation, Y.T. (Yong Tan); resources, S.H. and Y.T. (Yunong Tian); data curation, Z.L.; writing—original draft preparation, S.H. and Y.T. (Yong Tan); writing—review and editing, Z.L.; supervision, Z.L. and S.H. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded in part by the National Natural Science Foundation of China (Grant No. 62206275), the Science and Technology Research Program of Chongqing Municipal Education Commission (Grant Nos. KJZD-K202401406, KJQN202101446, KJQN20240143), and Natural Science Foundation of Chongqing (Grant No. cstc2021jcyj-msxmX0510).

Data Availability Statement

The data used in this study are available upon request from the corresponding author via email.

Acknowledgments

Acknowledgments for the data support from Chinese Academy of Sciences.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Chen, Z.; Luo, Y.; Wang, L.; Sun, D.; Wang, Y.; Zhou, J.; Luo, B.; Liu, H.; Yan, R.; Wang, L. Advancements in Life Tables Applied to Integrated Pest Management with an Emphasis on Two-Sex Life Tables. Insects 2025, 16, 261. [Google Scholar] [CrossRef]
  2. Guo, B.; Wang, J.; Guo, M.; Chen, M.; Chen, Y.; Miao, Y. Overview of Pest Detection and Recognition Algorithms. Electronics 2024, 13, 3008. [Google Scholar] [CrossRef]
  3. Wang, S.; Xu, D.; Liang, H.; Bai, Y.; Li, X.; Zhou, J.; Su, C.; Wei, W. Advances in Deep Learning Applications for Plant Disease and Pest Detection: A Review. Remote Sens. 2025, 17, 698. [Google Scholar] [CrossRef]
  4. Paymode, A.S.; Malode, V.B. Transfer Learning for Multi-Crop Leaf Disease Image Classification Using Convolutional Neural Network VGG. Artif. Intell. Agric. 2022, 6, 23–33. [Google Scholar] [CrossRef]
  5. Khan, D.; Waqas, M.; Tahir, M.; Islam, S.U.; Amin, M.; Ishtiaq, A.; Jan, L. Revolutionizing Real-Time Object Detection: YOLO and MobileNet SSD Integration. J. Comput. Biomed. Inform. 2023, 6, 41–49. [Google Scholar]
  6. Hussain, M.; Khanam, R. In-Depth Review of YOLOv1 to YOLOv10 Variants for Enhanced Photovoltaic Defect Detection. Solar 2024, 4, 351–386. [Google Scholar] [CrossRef]
  7. Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You Only Look Once: Unified, Real-Time Object Detection. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 779–788. [Google Scholar]
  8. Badgujar, C.M.; Poulose, A.; Gan, H. Agricultural object detection with You Only Look Once (YOLO) Algorithm: A bibliometric and systematic literature review. Comput. Electron. Agric. 2024, 223, 109090. [Google Scholar] [CrossRef]
  9. Tian, Y.; Wang, S.; Li, E.; Yang, G.; Liang, Z.; Tan, M. MD-YOLO: Multi-scale Dense YOLO for small target pest detection. Comput. Electron. Agric. 2023, 213, 108233. [Google Scholar] [CrossRef]
  10. Xu, W.; Xu, T.; Thomasson, J.A.; Chen, W.; Karthikeyan, R.; Tian, G.; Shi, Y.; Ji, C.; Su, Q. A lightweight SSV2-YOLO based model for detection of sugarcane aphids in unstructured natural environments. Comput. Electron. Agric. 2023, 211, 107961. [Google Scholar] [CrossRef]
  11. Wang, C.; Wang, L.; Ma, G.; Zhu, L. CSF-YOLO: A Lightweight Model for Detecting Grape Leafhopper Damage Levels. Agronomy 2025, 15, 741. [Google Scholar] [CrossRef]
  12. Jocher, G.; Chaurasia, A.; Qiu, J. Ultralytics YOLOv8 2023. Available online: https://github.com/ultralytics/ultralytics (accessed on 24 October 2024).
  13. Huang, Y.; Liu, Z.; Zhao, H.; Tang, C.; Liu, B.; Li, Z.; Wan, F.; Qian, W.; Qiao, X. YOLO-YSTs: An Improved YOLOv10n-Based Method for Real-Time Field Pest Detection. Agronomy 2025, 15, 575. [Google Scholar] [CrossRef]
  14. Wang, A.; Chen, H.; Liu, L. Yolov10: Real-time end-to-end object detection. arXiv 2024, arXiv:2405.14458. [Google Scholar]
  15. Feng, H.; Chen, X.; Duan, Z. LCDDN-YOLO: Lightweight Cotton Disease Detection in Natural Environment, Based on Improved YOLOv8. Agriculture 2025, 15, 421. [Google Scholar] [CrossRef]
  16. Girshick, R. Fast r-cnn. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015. [Google Scholar]
  17. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster r-cnn: Towards real-time object detection with region proposal networks. arXiv 2015, arXiv:1506.01497. [Google Scholar] [CrossRef]
  18. Girshick, R.; Donahue, J.; Darrell, T.; Malik, J. Rich feature hierarchies for accurate object detection and semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 580–587. [Google Scholar]
  19. Mu, J.; Sun, L.; Ma, B.; Liu, R.; Liu, S.; Hu, X.; Zhang, H.; Wang, J. TFEMRNet: A Two-Stage Multi-Feature Fusion Model for Efficient Small Pest Detection on Edge Platforms. AgriEngineering 2024, 6, 4688–4703. [Google Scholar] [CrossRef]
  20. He, K.; Gkioxari, G.; Dollár, P.; Girshick, R. Mask r-cnn. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017. [Google Scholar]
  21. Li, K.-R.; Duan, L.-J.; Deng, Y.-J.; Liu, J.-L.; Long, C.-F.; Zhu, X.-H. Pest Detection Based on Lightweight Locality-Aware Faster R-CNN. Agronomy 2024, 14, 2303. [Google Scholar] [CrossRef]
  22. Howard, A.; Sandler, M.; Chu, G.; Chen, L.C.; Chen, B.; Tan, M.; Wang, W.; Zhu, Y.; Pang, R.; Vasudevan, V.; et al. Searching for mobilenetv3. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 1314–1324. [Google Scholar]
  23. Guan, B.; Wu, Y.; Zhu, J.; Kong, J.; Dong, W. GC-Faster RCNN: The Object Detection Algorithm for Agricultural Pests Based on Improved Hybrid Attention Mechanism. Plants 2025, 14, 1106. [Google Scholar] [CrossRef]
  24. Koay, H.V.; Chuah, J.H.; Chow, C.; Chang, Y. Detecting and recognizing driver distraction through various data modality using machine learning: A review, recent advances, simplified framework and open challenges (2014–2021). Eng. Appl. Artif. Intell. 2022, 115, 105309. [Google Scholar] [CrossRef]
  25. Carion, N.; Massa, F.; Synnaeve, G.; Usunier, N.; Kirillov, A.; Zagoruyko, S. End-to-end object detection with transformers. In Proceedings of the European Conference on Computer Vision, Glasgow, UK, 23–28 August 2020; pp. 1, 2, 4, 6. [Google Scholar]
  26. Zhu, X.; Su, W.; Lu, L.; Li, B.; Wang, X.; Dai, J. Deformable DETR: Deformable Transformers for End-to-End Object Detection. arXiv 2021, arXiv:2010.04159. [Google Scholar]
  27. Liu, H.; Zhan, Y.; Sun, J.; Mao, Q.; Wu, T. A transformer-based model with feature compensation and local information enhancement for end-to-end pest detection. Comput. Electron. Agric. 2017, 138, 200–209. [Google Scholar] [CrossRef]
  28. Zhang, H.; Gong, Z.; Hu, C.; Chen, C.; Wang, Z.; Yu, B.; Suo, J.; Jiang, C.; Lv, C. A Transformer-Based Detection Network for Precision Cistanche Pest and Disease Management in Smart Agriculture. Plants 2025, 14, 499. [Google Scholar] [CrossRef] [PubMed]
  29. Zhang, Y.; Lv, C. TinySegformer: A lightweight visual segmentation model for real-time agricultural pest detection. Comput. Electron. Agric. 2024, 218, 180740. [Google Scholar] [CrossRef]
  30. Bottou, L.; Curtis, F.E.; Fang, L.; Nocedal, J. Optimization Methods for Large-Scale Machine Learning. SIAM Rev. 2018, 60, 223–311. [Google Scholar] [CrossRef]
  31. Wang, Q.; Zhang, S.; Dong, S.; Zhang, G.; Yang, J.; Li, R.; Wang, H. Pest24: A large-scale very small object data set of agricultural pests for multi-target detection. Comput. Electron. Agric. 2020, 175, 105585. [Google Scholar] [CrossRef]
  32. Khanam, R.; Hussain, M. YOLOV11: An Overview of the Key Architectural Enhancements. arXiv 2024, arXiv:2410.17725. [Google Scholar]
  33. Zhang, Z.; Bao, L.; Xiang, S.; Xie, G.; Gao, R. B2CNet: A Progressive Change Boundary-to-Center Refinement Network for Multitemporal Remote Sensing Images Change Detection. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2024, 17, 11322–11338. [Google Scholar] [CrossRef]
  34. Wang, C.Y.; Yeh, I.H.; Liao, H.Y.M. Yolov9: Learning what you want to learn using programmable gradient information. arXiv 2024, arXiv:2402.13616. [Google Scholar]
  35. Yang, L.; Zhang, R.Y.; Li, L.; Xie, X. SimAM: A Simple, Parameter-Free Attention Module for Convolutional Neural Networks. In Proceedings of the International Conference on Machine Learning, Virtual, 18–24 July 2021. [Google Scholar]
  36. Liu, S.; Huang, D.; Wang, Y. Learning spatial fusion for single-shot object detection. arXiv 2019, arXiv:1911.09516. [Google Scholar]
  37. Zhang, H.; Xu, C.; Zhang, S. Inner-IoU: More Effective Intersection over Union Loss with Auxiliary Bounding Box. arXiv 2023, arXiv:2311.02877. [Google Scholar]
  38. Ma, S.; Xu, Y. MPDIoU: A Loss for Efficient and Accurate Bounding Box Regression. arXiv 2023, arXiv:2307.07662. [Google Scholar]
  39. Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.Y.; Berg, A.C. SSD: Single Shot MultiBox Detector. arXiv 2015, arXiv:1512.02325. [Google Scholar]
  40. Lv, W.; Zhao, Y.; Xu, S.; Wei, J.; Wang, G.; Cui, C.; Du, Y.; Dang, Q.; Liu, Y. DETRs Beat YOLOs on Real-time Object Detection. arXiv 2023, arXiv:2304.08069. [Google Scholar]
Figure 1. The proprietary pest trap for dataset acquisition.
Figure 1. The proprietary pest trap for dataset acquisition.
Agronomy 15 01168 g001
Figure 2. Examples of the Pest24 [31] dataset.
Figure 2. Examples of the Pest24 [31] dataset.
Agronomy 15 01168 g002
Figure 3. The structure of LCFANet.
Figure 3. The structure of LCFANet.
Agronomy 15 01168 g003
Figure 4. Structural diagram of DTFA-C3k.
Figure 4. Structural diagram of DTFA-C3k.
Agronomy 15 01168 g004
Figure 5. Adaptive fusion strategy of MSASF module, as exemplified by MSASF-1.
Figure 5. Adaptive fusion strategy of MSASF module, as exemplified by MSASF-1.
Agronomy 15 01168 g005
Figure 6. The diagram of MPDINIoU.
Figure 6. The diagram of MPDINIoU.
Agronomy 15 01168 g006
Figure 7. Test examples on our private dataset.
Figure 7. Test examples on our private dataset.
Agronomy 15 01168 g007
Figure 8. Test examples on the Pest24 [31] dataset.
Figure 8. Test examples on the Pest24 [31] dataset.
Agronomy 15 01168 g008
Figure 9. Detection performance comparison between LCFANet and other SOTA baseline algorithms: (a) the mAP50 curves corresponding to models with different parameters. (b) The mAP50-95 curves corresponding to models with different parameters. (c) The mAP50 curves corresponding to models with different GFLOPs. (d) The mAP50-95 curves corresponding to models with different GFLOPs.
Figure 9. Detection performance comparison between LCFANet and other SOTA baseline algorithms: (a) the mAP50 curves corresponding to models with different parameters. (b) The mAP50-95 curves corresponding to models with different parameters. (c) The mAP50 curves corresponding to models with different GFLOPs. (d) The mAP50-95 curves corresponding to models with different GFLOPs.
Agronomy 15 01168 g009
Figure 10. Detection result examples of LCFANet and other SOTA lightweight pest image detection algorithms.
Figure 10. Detection result examples of LCFANet and other SOTA lightweight pest image detection algorithms.
Agronomy 15 01168 g010aAgronomy 15 01168 g010b
Table 1. Detection performance of LCFANet-n on the test datasets.
Table 1. Detection performance of LCFANet-n on the test datasets.
DatasetTest ImagesTest InstancesPrecision (%)Recall (%)mAP50(%)mAP50-95 (%)
Pest24 [31]507538,66671.768.570.445.1
Our Dataset189298592.491.786.854.7
Table 2. Detection performance of three pest species on our private dataset.
Table 2. Detection performance of three pest species on our private dataset.
CategoryTest ImagesTest InstancesPrecision (%)Recall (%)mAP50 (%)mAP50-95 (%)
Grapholita molesta65108491.390.584.952.1
Anarsia lineatella6098792.591.387.255.0
Cydia pomonella6491493.192.788.457.6
Overall189298592.491.786.854.7
Table 3. Performance comparison between LCFANet and other SOTA baseline algorithms.
Table 3. Performance comparison between LCFANet and other SOTA baseline algorithms.
ModelParams (M)GFLOPsPrecision (%)Recall (%)mAP50 (%)mAP50-95 (%)
YOLOv5-n2.195.952.662.558.033.6
YOLOv7-tiny5.3913.062.952.658.734.5
YOLOv8-n2.697.063.656.061.936.2
YOLOv9-t1.996.757.957.158.433.5
YOLOv10-n2.718.455.656.455.331.5
YOLOv11-n2.596.558.264.362.236.3
LCFANet-n (Ours)2.786.771.768.570.445.1
YOLOv5-s7.8319.076.063.168.341.5
YOLOv8-s9.8423.663.568.867.041.6
YOLOv9-s6.3222.871.767.466.238.8
YOLOv10-s8.0824.873.061.466.339.9
YOLOv11-s9.4321.663.974.168.342.3
LCFANet-s (Ours)9.9122.178.070.072.946.9
YOLOv5-m22.1452.964.870.370.345.3
YOLOv8-m23.2267.972.363.770.345.7
YOLOv9-m16.7361.166.068.569.044.2
YOLOv10-m16.5064.161.966.767.143.1
YOLOv11-m20.0668.266.571.873.047.0
LCFANet-m (Ours)22.1270.473.971.574.448.8
YOLOv5-l47.96114.968.169.771.846.6
YOLOv734.24104.765.256.064.439.6
YOLOv8-l39.47145.970.767.771.744.6
YOLOv9-c21.3784.266.369.370.844.7
YOLOv10-b20.4798.868.163.169.243.5
YOLOv10-l25.79127.365.163.167.143.5
YOLOv11-l25.3287.369.665.472.046.3
RT-DETR-l33.53109.463.365.264.138.5
SSD23.3790.364.661.061.637.3
Faster R-CNN32.4590.267.169.668.742.1
LCFANet-l (Ours)28.3591.372.770.473.249.0
YOLOv5-x89.06215.069.965.571.145.5
YOLOv7-x73.20189.967.057.565.640.9
YOLOv8-x61.64227.668.768.371.243.9
YOLOv9-e53.98173.167.570.568.844.8
YOLOv10-x31.68171.158.462.060.538.4
YOLOv11-x56.89195.566.565.968.845.0
RT-DETR-x67.83233.564.266.665.741.2
LCFANet-x (Ours)59.73199.866.473.973.048.1
Table 4. Performance comparison between LCFANet and other SOTA lightweight pest image detection algorithms.
Table 4. Performance comparison between LCFANet and other SOTA lightweight pest image detection algorithms.
ModelParams (M)GFLOPsPrecision (%)Recall (%)mAP50 (%)mAP50-95 (%)
YOLO-YSTs [13]6.513.862.564.464.837.9
CSF-YOLO [11]3.78.868.762.965.239.4
SSV2-YOLO [10]13.716.363.962.463.334.7
LLA-RCNN [21]6.474.565.766.566.838.1
LCFANet-n (Ours)2.786.771.768.570.445.1
Table 5. Results of the ablation experiments.
Table 5. Results of the ablation experiments.
DTFA
-C3k2
ADown
-Conv
CLHFPMSASF
Module
MPDINIoU
Loss
Params (M)GFLOPsmAP50 (%)mAP50-95 (%)
2.596.562.236.3
2.61 (↑ 0.02)6.563.8 (↑ 1.6)38.4 (↑ 2.1)
1.98 (↓ 0.61)5.5 (↓ 1.0)61.8 (↓ 0.4)36.0 (↓ 0.3)
3.44 (↑ 0.85)7.6 (↑ 1.1)66.4 (↑ 4.2)39.9 (↑ 3.6)
2.84 (↑ 0.25)6.9 (↑ 0.4)65.6 (↑ 3.4)39.1 (↑ 2.8)
2.596.563.4 (↑ 1.2)37.8 (↑ 1.5)
2.596.562.236.3
2.84 (↑ 0.25)6.9 (↑ 0.4)65.6 (↑ 3.4)39.1 (↑ 2.8)
3.57 (↑ 0.73)7.8 (↑ 0.9)67.9 (↑ 2.3)41.7 (↑ 2.6)
3.58 (↑ 0.01)7.9 (↑ 0.1)68.7 (↑ 0.8)43.3 (↑ 1.6)
2.78 (↓ 0.8)6.7 (↓ 1.2)68.2 (↓ 0.5)42.9 (↓ 0.4)
2.786.770.4 (↑ 2.2)45.1 (↑ 2.2)
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Huang, S.; Tian, Y.; Tan, Y.; Liang, Z. LCFANet: A Novel Lightweight Cross-Level Feature Aggregation Network for Small Agricultural Pest Detection. Agronomy 2025, 15, 1168. https://doi.org/10.3390/agronomy15051168

AMA Style

Huang S, Tian Y, Tan Y, Liang Z. LCFANet: A Novel Lightweight Cross-Level Feature Aggregation Network for Small Agricultural Pest Detection. Agronomy. 2025; 15(5):1168. https://doi.org/10.3390/agronomy15051168

Chicago/Turabian Style

Huang, Shijian, Yunong Tian, Yong Tan, and Zize Liang. 2025. "LCFANet: A Novel Lightweight Cross-Level Feature Aggregation Network for Small Agricultural Pest Detection" Agronomy 15, no. 5: 1168. https://doi.org/10.3390/agronomy15051168

APA Style

Huang, S., Tian, Y., Tan, Y., & Liang, Z. (2025). LCFANet: A Novel Lightweight Cross-Level Feature Aggregation Network for Small Agricultural Pest Detection. Agronomy, 15(5), 1168. https://doi.org/10.3390/agronomy15051168

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop