Next Article in Journal
Descriptions of Two New Species of Encarsia (Hymenoptera: Aphelinidae) and Mitochondrial Genome Analysis of Three Species of the Genus
Previous Article in Journal
Aquatic Diptera in Phytotelmata of Bromeliaceae and Zingerberales
Previous Article in Special Issue
Functional Characterization of CaSpr2 in Jasmonate-Dependent Induced Defense Against Western Flower Thrips in Capsicum annuum
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Multi-Scale Vision–Sensor Collaborative Framework for Small-Target Insect Pest Management

1
China Agricultural University, Beijing 100083, China
2
National School of Development, Peking University, Beijing 100871, China
*
Author to whom correspondence should be addressed.
Insects 2026, 17(3), 281; https://doi.org/10.3390/insects17030281
Submission received: 28 January 2026 / Revised: 27 February 2026 / Accepted: 1 March 2026 / Published: 4 March 2026

Simple Summary

Small-target insect pests pose a major challenge to intelligent agricultural monitoring due to their tiny size, complex backgrounds, and strong dependence on environmental conditions. To address these issues, this study proposes a multi-scale vision–sensor collaborative framework that integrates visual imagery with environmental sensing data for accurate pest recognition. The model first captures fine-grained pest features through multi-scale visual representation learning, and then introduces environmental factors—such as temperature, humidity, and illumination—as prior information to guide feature discrimination. A collaborative fusion mechanism is further designed to enhance cross-modal consistency and improve classification robustness. Experiments conducted on a real multimodal pest dataset collected from farmland and greenhouse environments demonstrate that the proposed method achieves 93.1% accuracy, 92.0% precision, 91.2% recall, and a 91.6% F1-score, outperforming conventional machine learning and single-modality deep learning approaches. These results indicate that integrating ecological sensing with computer vision can provide a reliable technical pathway for early pest detection and precision agricultural management.

Abstract

In complex agricultural production environments, small-target pests—characterized by tiny scales, strong background confusion, and close dependence on environmental conditions—pose major challenges to precise monitoring and green pest control. To facilitate the transition from experience-driven to data-driven pest management, a multi-scale vision–sensor collaborative recognition method is proposed for field and protected agriculture scenarios to improve the accuracy and stability of small-target pest recognition under complex conditions. The method jointly models multi-scale visual representations and pest ecological mechanisms: a multi-scale visual feature module enhances fine-grained texture and morphological cues of small targets in deep networks, alleviating feature sparsity and scale mismatch, while environmental sensor data, including temperature, humidity, and illumination, are introduced as priors to modulate visual features and explicitly incorporate ecological constraints into the discrimination process. Stable multimodal fusion and pest category prediction are then achieved through a vision–sensor collaborative discrimination module. Experiments on a multimodal dataset collected from real farmland and greenhouse environments in Linhe District, Bayannur City, Inner Mongolia, demonstrate that the proposed method achieves approximately 93.1% accuracy, 92.0% precision, 91.2% recall, and a 91.6% F 1 -score on the test set, significantly outperforming traditional machine learning approaches, single-scale deep learning models, and multi-scale vision baselines without environmental priors. Category-level evaluations show balanced performance across multiple small-target pests, including aphids, thrips, whiteflies, leafhoppers, spider mites, and leaf beetles, while ablation studies confirm the critical contributions of multi-scale visual modeling, environmental prior modulation, and vision–sensor collaborative discrimination.

1. Introduction

Accurate and early identification of agricultural pests is a core part of crop protection and green pest management, which is critical to protecting crop yield and quality, cutting chemical pesticide use and preserving agricultural ecological security [1]. In the early crop growth stage, pest populations are small and damage symptoms are inconspicuous, so timely monitoring is essential for precise intervention without large-scale pesticide application, thus reducing control costs and ecological risks [2]. Small-scale pests like aphids and thrips are particularly challenging to identify due to their tiny size, rapid reproduction, dense aggregation and cryptic habitats on leaf undersides or in plant gaps, which cause their populations to expand quickly and threaten crop growth persistently, making them a key difficulty in intelligent field monitoring and automated recognition [3]. With the advancement of high-resolution imaging, intelligent trapping systems and agricultural IoT technologies, computer vision-based pest recognition has become a vital research direction in smart agriculture, enabling all-weather, low-cost and large-scale monitoring [4].
Before deep learning was widely adopted, pest identification mainly depended on traditional machine vision combined with expert knowledge [5], using handcrafted color, texture or morphological features with threshold segmentation, edge detection or shallow classifiers [6]. However, the complex leaf textures, drastic illumination changes and subtle morphological differences of small pests in real field environments make these methods highly sensitive to imaging conditions and poor in generalization [7]. Additionally, handcrafted features fail to capture the diverse morphology and spatial distribution of pests, leading to frequent false and missed detections in complex backgrounds [8], which greatly restricts the application of traditional methods in small-target pest recognition [7]. With the rapid development of deep learning [9], convolutional neural networks (CNNs) and their variants have shown strong feature learning capabilities in agricultural pest recognition [10]. End-to-end training enables the automatic learning of discriminative features, partially overcoming the limitations of manual feature design [11], and attention mechanisms and Transformer architectures further boost global contextual modeling [12]. Nevertheless, most existing deep learning-based methods rely on a single visual modality, and their performance in small-target pest recognition is still limited under complex field conditions [13]. Small targets occupy only a tiny proportion of image pixels, and their features are easily weakened or lost during repeated downsampling [14]; meanwhile, illumination variation, occlusion, background noise and imaging blur further intensify target–background confusion [15]. More importantly, the intrinsic ecological relationships between pest occurrence and environmental conditions are often neglected, and environmental factors such as temperature, humidity and illumination are rarely used as prior constraints [16].
Recent studies have explored advanced visual modeling strategies: Lyu et al. [17] proposed a CNN–Transformer hybrid framework for multi-label natural enemy–pest detection, achieving remarkable improvements in F1-score and mAP; Zhang et al. [18] introduced the DFRCP framework to enhance pest damage detection in motion-blurred images, with a performance gain of about 10.4% when combined with YOLOv11; Vhatkar et al. [19] proposed an optimization-assisted SSD integrated with graph attention networks to improve pest detection in complex backgrounds; and Seol et al. [20] enhanced spray perception accuracy in precision aerial spraying via spatial–temporal filtering and 3D deep learning. Despite these progress, automatic recognition of small-target pests still faces three key challenges: the scale mismatch between small targets and deep network receptive fields, high target–background similarity under complex conditions, and inadequate modeling and fusion of environmental prior information. To tackle these challenges, a multi-scale vision–sensor collaborative pest recognition framework is proposed. Multi-scale visual feature modeling strengthens the fine-grained representation of small targets. From a biological perspective, instantaneous microclimate conditions—such as immediate temperature, humidity, and illumination at the exact time of photography—heavily influence the acute behavioral responses of small arthropods, such as their tendency to seek shade on leaf undersides or within plant crevices during intense midday heat and light. By capturing these acute behavioral contexts, instantaneous environmental sensor data is introduced as prior modulation signals to guide the model to generate ecologically plausible visual responses, thereby improving recognition performance in complex agricultural scenarios. The main contributions of this study are formulated as follows.
  • A multi-scale visual feature modeling mechanism for small-target pests is proposed, where cross-scale information fusion effectively mitigates feature sparsity and scale mismatch problems;
  • A vision–sensor collaborative prior modulation strategy is designed to deeply integrate instantaneous environmental sensor information into the pest recognition process, enhancing the model’s robustness under complex local conditions;
  • Systematic experiments on real field-collected datasets verify that the proposed method significantly outperforms multiple mainstream approaches in small-target pest recognition tasks;
  • The proposed framework enables early and accurate identification of small-target pests to support precision pest management, which is potentially beneficial for reducing pesticide input, lowering production costs, and improving agricultural productivity.

2. Related Work

2.1. Deep Learning-Based Pest Recognition Methods

With the growing adoption of computer vision and artificial intelligence in agriculture, automatic pest recognition has evolved from traditional machine vision techniques to deep learning-based models [21]. Early approaches relied on handcrafted color, texture, and morphological features combined with shallow classifiers such as support vector machines and random forests [22]. However, these methods are highly dependent on feature engineering and imaging conditions, making them sensitive to illumination variation, background complexity, and individual morphological differences [23]. In real field environments, complex leaf textures and background noise, together with the high visual similarity between small pests and surrounding backgrounds, significantly limit the generalization ability and stability of traditional methods [24]. With the emergence of convolutional neural networks, deep learning-based pest recognition has become the dominant research paradigm [25]. CNNs can automatically learn hierarchical feature representations through multi-layer convolution and nonlinear mappings, partially overcoming the limitations of manual feature design [26]. Classical classification networks and detection frameworks have been introduced into pest recognition and have achieved promising results under controlled conditions or relatively simple backgrounds [27]. Attention mechanisms and Transformer architectures have further been explored to enhance global contextual modeling and mitigate background interference [28]. Nevertheless, most existing models are originally designed for medium- and large-scale objects, and their downsampling and feature extraction strategies are not tailored to small-target pests [29]. When targets occupy only a small fraction of image pixels, fine-grained details are easily weakened or lost during deep feature extraction, resulting in limited discriminative capability for pests such as aphids and thrips [30]. This limitation becomes more pronounced in complex field environments and remains a key bottleneck in current deep learning-based pest recognition research [31].

2.2. Small-Target Visual Recognition and Multi-Scale Modeling

To address the challenges of small-target recognition, multi-scale modeling has become an important research direction in computer vision [32]. The core idea is to jointly model target features across multiple scale spaces to compensate for information loss at a single scale [33]. Representative approaches include feature pyramid networks, multi-level feature fusion architectures, and cross-scale attention mechanisms [34]. By integrating high-resolution shallow features with low-resolution semantic features, these methods have substantially improved small-target detection performance in general object detection tasks [35]. Theoretically, multi-scale modeling helps alleviate scale mismatch by preserving spatial details while introducing semantic abstraction [36]. However, most existing multi-scale methods are designed for structured scenarios such as urban environments or remote sensing imagery, where background patterns are relatively regular [37]. Agricultural scenes are inherently unstructured, characterized by complex crop morphologies, repetitive textures, and significant appearance variations of pests across growth stages and environmental conditions [38]. When directly applied to pest recognition tasks, existing multi-scale methods often fail to fully exploit their advantages, and problems such as target–background confusion and high false-detection rates persist [39]. Consequently, designing multi-scale visual modeling strategies that are better adapted to small-target pests in agricultural scenes remains an open research problem [40].

2.3. Multimodal and Sensor-Assisted Agricultural Perception Methods

Pest occurrence processes are closely related to environmental conditions, providing a theoretical basis for multimodal and sensor-assisted agricultural perception [41]. Numerous studies have shown that factors such as temperature, humidity, and illumination significantly influence pest growth, activity patterns, and population dynamics. Accordingly, environmental sensors have been widely deployed for pest outbreak analysis and early risk warning [42]. Most existing studies model sensor data as time-series signals using statistical or machine learning approaches, offering macro-level decision support for agricultural management. However, sensor information is typically decoupled from visual recognition, limiting its contribution to fine-grained perception tasks [43]. From a deep learning perspective, current multimodal agricultural perception methods remain limited in collaborative modeling between visual and sensor information [44]. Some studies simply concatenate environmental variables with visual features as additional inputs [45], but such shallow fusion strategies fail to fully exploit the guiding role of environmental information in visual discrimination [46]. This limitation is particularly evident in small-target pest recognition, where performance gains are often unstable [47]. Moreover, environmental conditions are rarely treated as ecological priors to modulate visual feature learning, causing models to rely predominantly on visual cues and making robust recognition under complex field environments difficult to achieve [48].

3. Materials and Method

3.1. Data Collection

The dataset used in this study was collected from representative open-field and facility-based agricultural production environments in Linhe District, Bayannur City, Inner Mongolia. It should be noted that all field data originate from the same geographic region. Although this ensures environmental consistency and controlled ecological variation, it may limit the direct generalization of the model to regions with substantially different climatic conditions, cropping systems, or pest population structures. This region is located in the Hetao Irrigation Area, where agricultural cultivation is intensive and pest occurrence exhibits strong seasonality and representativeness. In addition to field-collected data, a small portion of supplementary data was obtained from publicly available online sources. The data collection period spanned from April 2023 to November 2023, covering multiple critical crop growth stages, including the seedling stage, vegetative growth stage, and early fruiting stage, while also encompassing typical climatic conditions such as spring–summer transition, high-temperature drought, and large diurnal temperature variations. Pest image data were mainly acquired from open-field vegetable plots and solar greenhouse environments, with crop types including maize, tomato, and pepper. These settings effectively reflect the real-world distribution characteristics of pests in local agricultural production. The dataset consists of both image data and environmental sensor data, as shown in Table 1.
Image data were acquired using a combination of fixed agricultural monitoring cameras and handheld high-definition cameras. Fixed cameras were deployed between crop rows or at the top of greenhouses to continuously record pest occurrence and activity under natural conditions, whereas handheld cameras were used to supplement multi-view and local high-resolution images in order to enhance the visibility of small-target pests. During acquisition, original image resolutions were preserved as much as possible, with resolutions mainly ranging from 1920 × 1080 to 3840 × 2160 , so as to avoid the loss of fine-grained details of small targets caused by excessive compression. All images were collected under natural illumination without artificial intervention in the field environment, thereby ensuring the authenticity and complexity of the data. In terms of pest categories, the dataset focuses on small-target pests that are common in Linhe District and difficult to control. Specifically, the monitored arthropods and their corresponding host crops include the green peach aphid (Myzus persicae) and leafhoppers (Empoasca spp.) infesting peppers, western flower thrips (Frankliniella occidentalis) and sweetpotato whiteflies (Bemisia tabaci) on tomatoes, as well as two-spotted spider mites (Tetranychus urticae) and leaf beetles (Monolepta hieroglyphica) on maize, as shown in Figure 1. The density levels of these arthropods varied significantly across the captured data, ranging from low-density occurrences with a few scattered individuals to high-density colonies characterized by severe spatial overlap and aggregation. These pests are generally small in size and tend to aggregate on the undersides of leaves, young shoots, or near leaf veins. Under conditions of complex leaf textures and varying illumination, they are easily confused with background regions, exhibiting typical characteristics of small-target recognition tasks. All image samples were annotated frame by frame by professionals with plant protection expertise, followed by cross-validation procedures to ensure the accuracy of pest categories and annotated regions.
In addition to visual images, environmental sensor data temporally aligned with the image data were synchronously collected to characterize the ecological background of pest occurrence. Sensor nodes were deployed within the data collection areas to continuously record key environmental variables, including air temperature, relative humidity, and illumination intensity. The sampling frequency was set to once every 10 min, and precise temporal alignment with image data was achieved using timestamps. During data acquisition, anomaly monitoring was conducted in real time, and missing values were subsequently corrected and normalized during post-processing.
Image data were collected using fixed agricultural monitoring cameras (Hikvision DS-2CD3T47G2-L, Hikvision, Hangzhou, China) and handheld high-definition cameras (Canon EOS 90D, Canon Inc., Tokyo, Japan) equipped with EF-S 18–135 mm lenses. During acquisition, exposure time was adaptively set within the range of 1/60–1/200 s under natural illumination, and white balance was configured in automatic mode to accommodate dynamic field lighting conditions. Environmental variables were synchronously recorded using integrated agricultural sensor nodes (Model: RS-WS-N01, Renke Instruments, Yantai, Shandong, China). Air temperature was measured in degrees Celsius (°C), humidity was defined as relative humidity (%RH), and illumination intensity was measured in lux. All sensors were calibrated prior to deployment to ensure measurement consistency across different acquisition sites.

3.2. Data Pre-Processing and Augmentation Strategy

In complex agricultural scenarios, small-target pest recognition tasks are often affected by issues such as inconsistent image resolutions, significant background noise, extremely small-target scales, and temporal asynchrony among multimodal data sources. If these factors are not properly addressed, the convergence stability and generalization performance of deep models are directly degraded. Therefore, systematic pre-processing and augmentation of both image data and environmental sensor data before model training constitute a critical prerequisite for ensuring the effectiveness of the proposed multi-scale vision–sensor collaborative framework.
For visual data, differences in acquisition devices and capture distances typically lead to large variations in image resolution, whereas deep learning models generally require fixed input sizes to enable batch training and feature alignment. Based on this principle, resolution normalization is first applied to all raw images. Let a raw image be represented as I R H × W × C , where H, W, and C denote the image height, width, and number of channels, respectively. Through bilinear interpolation, the image is mapped to a unified resolution H 0 × W 0 , which can be expressed as
I = R ( I ; H 0 , W 0 ) ,
where R ( · ) denotes the resolution resampling operator. This process preserves the overall structural information while ensuring spatial-scale comparability across images from different sources, thereby providing a unified input basis for subsequent multi-scale feature modeling. However, simple global resizing may further reduce the pixel proportion of small-target pests, which weakens their discriminative features. To alleviate this issue, a cropping-based local enhancement strategy is introduced on top of resolution normalization. Local regions are randomly selected from each image for cropping and then rescaled to the original input size, so that small targets occupy a larger relative proportion in the cropped images. Let the top-left coordinate of the cropping region be ( x , y ) , with cropping width w c and height h c . The cropping operation is formulated as
I c = I ( x : x + w c , ; y : y + h c ) ,
after which a scale transformation is applied to map I c back to the unified input resolution. In this manner, the relative scale of small targets is increased without modifying the network architecture, thereby mitigating the progressive weakening of small-target features in deep networks from the data perspective. On this basis, a data augmentation mechanism combining random scale perturbation and local magnification is further designed for small-target pests. The core idea is to impose random perturbations on target scales during training to enhance the model’s adaptability to scale variations. Specifically, a random scaling factor α is applied to the cropped image I c , where α follows a uniform distribution over the interval [ α min , α max ] , i.e.,
α U ( α min , α max ) ,
and the scaled image is represented as
I s = S ( I c ; α ) ,
where S ( · ) denotes the scale transformation operator. By continuously altering the relative size of small targets during training, more robust multi-scale representations are learned, effectively reducing the model’s dependence on fixed scale distributions. In addition to visual data, environmental sensor data, including temperature, humidity, and illumination, are synchronously collected, as these variables possess important ecological significance in pest occurrence and activity patterns. However, sensor data in real-world acquisition processes are often accompanied by noise, missing values, and inconsistent sampling frequencies, which can negatively affect vision–sensor collaborative modeling if directly fed into the model. Therefore, systematic cleaning and normalization of sensor data are required prior to data fusion. Let the sensor data vector collected at time t be denoted as s ( t ) = [ s 1 ( t ) , s 2 ( t ) , , s K ( t ) ] , where K represents the number of sensor variables. For outlier detection, a statistical distribution-based removal strategy is adopted. For each variable s k ( t ) , its mean μ k and standard deviation σ k are computed, and values outside a reasonable range are treated as anomalies, i.e.,
| s k ( t ) μ k | > λ σ k ,
where λ is an empirical threshold. After anomaly removal, missing values are filled using linear interpolation to ensure the continuity of the time series. Since the sampling frequencies of image data and sensor data are generally different, multimodal temporal alignment is further performed. Let the image acquisition timestamp be t i and the sensor sampling timestamp be t s . Synchronization between image and sensor data is achieved by searching for the nearest neighbor time point that satisfies | t i t s | < Δ t , which can be formulated as
s ( t i ) = s ( arg min t s | t i t s | ) ,
where Δ t denotes the maximum allowable temporal deviation. This strategy ensures that each image is associated with a set of ecologically meaningful environmental variables, thereby providing reliable inputs for subsequent vision–sensor collaborative modeling. After cleaning and alignment, sensor data are further standardized to avoid training interference caused by differences in numerical scales among variables. Zero-mean unit-variance normalization is applied to each variable:
s ^ k ( t ) = s k ( t ) μ k σ k ,
where s ^ k ( t ) denotes the normalized sensor variable. This process facilitates faster model convergence and improves the stability of multimodal feature fusion. Through the above procedures of image pre-processing, small-target enhancement, and sensor data cleaning and synchronization, a multimodal input system with spatial-scale and temporal consistency is constructed at the data level. This pre-processing and augmentation strategy not only enhances the perceptibility of small-target pests in the visual domain but also ensures the semantic and temporal alignment of environmental sensor information, thereby establishing a solid foundation for subsequent multi-scale visual feature modeling and environmental prior modulation modules and providing the necessary conditions for ecologically rational vision–sensor collaborative recognition.

3.3. Proposed Method

3.3.1. Overall

After data processing and temporal alignment, a single input image is denoted as X R H 0 × W 0 × C , and its corresponding environmental sensor vector is denoted as s R K . The overall model follows a pipeline-style architecture consisting of multi-scale visual representation learning, environmental prior modulation, and vision–sensor collaborative discrimination. First, the image X is fed into the multi-scale visual feature modeling module, where a hierarchical backbone network extracts a feature pyramid { F 1 , F 2 , , F L } from shallow to deep layers. The shallow feature F 1 preserves higher spatial resolution to capture texture and edge details of small targets, while the deep feature F L provides stronger semantic representations to suppress background interference. Subsequently, a cross-scale fusion unit aligns { F l } into a unified feature space and aggregates information to obtain an enhanced multi-scale visual representation F m s , ensuring that small targets remain sufficiently distinguishable in deep semantic space. Next, the environmental sensor prior modulation module takes s as input and generates a modulation vector g through nonlinear mapping, which is applied to F m s for dynamic weighting and guidance, yielding the environment-modulated visual feature F ˜ . This process can be interpreted as an ecological-consistency-based recalibration of visual responses along channel or spatial dimensions, strengthening visual evidence under environment conditions consistent with pest occurrence while suppressing noise activations under mismatched conditions. Finally, the vision–sensor collaborative discrimination and classification module receives F ˜ and constructs a joint discriminative representation z, which is processed by a lightweight classification head to perform feature compression and category separation, outputting pest category probabilities p = softmax ( W z + b ) . The representation z jointly encodes multi-scale fine-grained appearance cues and stable semantics constrained by environmental priors, enabling robust recognition of small-target pests such as aphids and thrips under complex backgrounds and varying conditions.

3.3.2. Multi-Scale Small-Target Visual Feature Modeling Module

The multi-scale small-target visual feature modeling module takes the processed input image X R H 0 × W 0 × C as its starting point. The overall structure follows the design philosophy of multi-resolution parallel modeling, cross-scale relational fusion, and channel-adaptive enhancement, aiming to mitigate the progressive attenuation of small-target features in deep networks.
As shown in Figure 2, image features are first mapped layer by layer through a shared-parameter convolutional backbone network to obtain a set of hierarchical feature representations { F ( 1 ) , F ( 2 ) , F ( 3 ) , F ( g ) } . Here, F ( 1 ) R H 0 × W 0 × C 1 represents high-resolution shallow features, F ( 2 ) and F ( 3 ) correspond to medium- and low-resolution features, respectively, and F ( g ) denotes a global-scale feature obtained via global receptive field modeling. Features at each scale are constructed using multiple 3 × 3 convolutional layers followed by batch normalization and nonlinear activation, ensuring consistent representational capacity across different spatial scales. To address the issue that small targets are easily overwhelmed in deep semantic layers, a channel attention mechanism is introduced at each scale to adaptively recalibrate the importance of feature channels. Specifically, global average pooling is applied to each scale feature F ( l ) to obtain a channel descriptor vector z ( l ) R C l , which is then transformed through nonlinear mappings to generate channel weights w ( l ) = σ ( W 2 δ ( W 1 z ( l ) ) ) , where δ ( · ) and σ ( · ) denote the activation function and normalization function, respectively. The weighted feature F ^ ( l ) = w ( l ) F ( l ) explicitly enhances fine-grained channel responses associated with small targets, thereby reducing interference from background textures.
On this basis, cross-scale fusion is performed to enable information interaction among features at different resolutions. Specifically, features at each scale are first mapped to a unified spatial resolution via scale alignment operators, after which learnable weights { α l } are introduced to combine features from different scales, yielding the fused feature
F m s = l α l · F ^ ( l ) , l α l = 1 .
This weighted fusion strategy avoids redundant representations caused by naive concatenation and allows the model to automatically balance high-resolution detail information and low-resolution semantic information according to task requirements. In addition, to enhance local consistency across scales, a cross-scale relational unit further models feature correlations within spatial neighborhoods, explicitly capturing correspondences of target regions across scales and improving continuous responses in densely distributed small-target areas. From a mathematical perspective, the introduction of multi-scale parallel mappings and channel-adaptive weighting in feature space is equivalent to imposing a set of scale-dependent reconstruction constraints on the original feature representation. As a result, feature responses corresponding to small targets are simultaneously amplified and kept consistent across multiple scales. This design effectively reduces discriminative instability caused by sparsity at a single scale, enabling small-target pests to maintain separable feature structures even in deep semantic space and providing a more robust and sufficient visual representation for subsequent environmental prior modulation and vision–sensor collaborative discrimination modules.

3.3.3. Environmental Sensor Prior Modulation Module

The environmental sensor prior modulation module operates on the output of the multi-scale small-target visual feature modeling module. Its core objective is to introduce ecological constraints and perform learnable and interpretable dynamic modulation of visual features without altering the main discriminative structure of the visual backbone.
As shown in Figure 3, the fused multi-scale visual feature is denoted as F m s R H × W × C , where H and W represent spatial resolution and C denotes the number of channels. In this implementation, H = W = 64 and C = 256 , indicating that the feature already integrates high-resolution details and low-resolution semantic information. Correspondingly, the environmental sensor input vector is denoted as s R K with K = 3 , representing air temperature, relative humidity, and illumination intensity, respectively, and is standardized to ensure comparable numerical scales. To map low-dimensional environmental variables to modulation signals applicable to high-dimensional visual features, a lightweight environmental encoding subnetwork is constructed. This subnetwork adopts a three-layer fully connected structure, with input dimension K = 3 , first hidden layer width d 1 = 32 , second hidden layer width d 2 = 128 , and output layer width equal to the visual feature channel number C = 256 . Through this design, environmental variables are progressively embedded into a modulation space aligned with visual channels. The mapping process is formulated as
e 1 = ϕ ( W 1 s + b 1 ) , e 2 = ϕ ( W 2 e 1 + b 2 ) ,
g = σ ( W 3 e 2 + b 3 ) ,
where W 1 R 32 × 3 , W 2 R 128 × 32 , and W 3 R 256 × 128 . Here, ϕ ( · ) denotes a nonlinear activation function and σ ( · ) denotes a normalization mapping to stabilize the value range of the modulation vector g R 256 . Each element of g can be interpreted as an importance weight for the corresponding visual channel under the current environmental condition.
The modulation vector g is then broadcast along spatial dimensions and applied to the visual feature F m s via channel-wise modulation, yielding the environment-constrained visual representation F ˜ :
F ˜ h , w , c = g c · F h , w , c m s , h , w , c .
This operation is mathematically equivalent to introducing a diagonal modulation operator in feature space, which imposes environment-dependent scaling constraints on different semantic channels. Compared with direct concatenation of sensor features, this multiplicative modulation does not introduce additional spatial noise but instead reshapes the energy distribution of feature responses, systematically amplifying or suppressing visual activations in an ecologically consistent manner.
From a theoretical standpoint, if visual features are treated as random variables F m s , their expected responses under different environmental conditions can be written as E [ F m s s ] . After introducing the modulation vector, the model effectively learns the conditional distribution E [ F ˜ s ] = E [ g ( s ) F m s ] , which is equivalent, under a first-order approximation, to introducing environment-dependent prior weights in feature space, thereby reducing the variance of features inconsistent with environmental conditions. This property is particularly important for small-target pest recognition, where background noise often exhibits high instability across environments, while pest-related features tend to show stronger ecological consistency.
When combined with the multi-scale visual module, this modulation mechanism does not alter the spatial structure or scale relationships of F m s but operates along the channel dimension, reordering semantic responses of fused multi-scale features under environmental constraints. This serial design of visual modeling followed by prior modulation ensures that geometric and textural information of small targets is fully captured first, and then semantically constrained by environmental information, avoiding premature interference of environmental variables that could limit visual expressiveness. In the context of the task addressed in this study, the module significantly enhances discriminative channels for pest categories under high-risk environmental conditions such as high temperature and humidity, while suppressing spurious activations under unfavorable conditions, thereby improving recognition stability and generalization in complex agricultural scenarios while maintaining model simplicity and interpretability.

3.3.4. Vision–Sensor Collaborative Discrimination and Classification Module

The vision–sensor collaborative discrimination and classification module takes the environment-modulated multi-scale visual representations as its primary input and uses sensor embeddings as conditional semantics to achieve consistent alignment and discriminative enhancement across scales.
As shown in Figure 4, the four feature levels after multi-scale visual modeling and prior modulation are denoted as { F ˜ 1 , F ˜ 2 , F ˜ 3 , F ˜ 4 } , with spatial sizes and channel numbers set as F ˜ 1 R 64 × 64 × 64 , F ˜ 2 R 32 × 32 × 128 , F ˜ 3 R 16 × 16 × 192 , and F ˜ 4 R 8 × 8 × 256 . The standardized sensor vector is denoted as s R 3 . To provide controllable discriminative constraints across scales, a two-layer sensor encoder first maps s into a shared semantic space with dimension d = 64 :
h = ψ ( W a s + b a ) , u = ψ ( W b h + b b ) ,
where W a R 32 × 3 and W b R 64 × 32 , and ψ ( · ) denotes a nonlinear activation function. The vector u is then projected through four scale-specific linear transformations to generate conditional vectors { c i } i = 1 4 :
c i = P i u + q i , P i R C i × 64 ,
where ( C 1 , C 2 , C 3 , C 4 ) = ( 64 , 128 , 192 , 256 ) . For each scale, a conditional gating and residual injection mechanism is applied to fuse sensor semantics with local visual responses without disrupting multi-scale geometric structure. Specifically, a gating map is generated from c i and broadcast spatially, and the fused feature M i is obtained as
G i = reshape ( c i , 1 , 1 , C i ) 1 H i × W i , M i = F ˜ i + φ F ˜ i G i ,
where ( H 1 , H 2 , H 3 , H 4 ) = ( 64 , 32 , 16 , 8 ) , φ ( · ) denotes a 1 × 1 convolution followed by ReLU, and ⊗ indicates broadcasting. The fused features at all scales are then upsampled to 64 × 64 and aggregated:
M ¯ 1 = M 1 , M ¯ 2 = Up ( M 2 , 2 ) , M ¯ 3 = Up ( M 3 , 4 ) , M ¯ 4 = Up ( M 4 , 8 ) ,
F c s = ρ Conv 1 × 1 M ¯ 1 M ¯ 2 M ¯ 3 M ¯ 4 , F c s R 64 × 64 × 256 ,
where Up ( · ) denotes bilinear upsampling, ∥ denotes channel concatenation, Conv 1 × 1 compresses the concatenated 640 channels to 256, and ρ ( · ) denotes ReLU. The final classification head applies global average pooling followed by a two-layer perceptron to output probabilities for six pest categories. Let the pooled vector be z R 256 :
z = GAP ( F c s ) , o = W 2 ψ ( W 1 z + b 1 ) + b 2 , p = Softmax ( o ) ,
where W 1 R 128 × 256 and W 2 R 6 × 128 . The key advantage of this design lies in the fact that sensor information is not simply concatenated at the decision layer but participates in feature formation through scale-aligned conditional gating, enabling consistent reordering and stable activation of multi-scale visual responses under different environmental states. From a risk minimization perspective, let the true label be Y, visual features be V, and sensor variables be S. For any loss function and hypothesis class F , the optimal conditional predictor satisfies
f = arg min f F E ( Y , f ( V , S ) ) , E ( Y , f ( V , S ) ) E ( Y , f ( V , S ¯ ) ) ,
where S ¯ denotes a degenerate constant condition independent of V. By conditioning visual representations on S and forming M i across scales, the realizable set of discriminative functions is effectively expanded, thereby reducing expected empirical risk under fixed model capacity and improving separability of small targets under complex backgrounds and scale variations.

4. Results and Discussion

4.1. Experimental Configuration

4.1.1. Hardware and Software Platform

All experiments were conducted on a dedicated deep learning server equipped with an NVIDIA RTX 3090 GPU with 24 GB of memory, which was used to accelerate both model training and inference. The GPU provided sufficient parallel computing capability to support multi-scale feature extraction and batch-based processing during training. The server was further equipped with an Intel Xeon Gold 6226R CPU running at 2.90 GHz with 16 physical cores, along with 128 GB of system memory, ensuring efficient data loading, pre-processing, and caching of multimodal inputs. A high-speed solid-state drive (SSD) with a storage capacity of 2 TB was used to store raw image data, environmental sensor records, and intermediate experimental outputs, thereby minimizing input/output latency and improving overall training efficiency. This hardware configuration enabled stable execution of repeated experiments and cross-validation procedures on real-world agricultural datasets.
On the software side, all experiments were implemented under a Linux operating system (Ubuntu 20.04 LTS). Model development, training, and inference were carried out using the PyTorch deep learning framework (version 1.13.1), with CUDA (version 11.7) and cuDNN (version 8.5) providing GPU acceleration for numerical computation and tensor operations. Image pre-processing and data augmentation were performed using OpenCV (version 4.6.0) and the torchvision library (version 0.14.1), while environmental sensor data cleaning and normalization were implemented using NumPy (version 1.23.5) and SciPy (version 1.9.3). To ensure reproducibility, random seeds were fixed for the Python 3.10.12 environment, NumPy, and PyTorch throughout all experiments, and identical software configurations were maintained across all experimental runs.
During model training, the constructed dataset was divided into training, validation, and test sets, accounting for 70%, 15%, and 15% of the total data, respectively. To prevent potential spatial or temporal data leakage, samples collected from the same farmland plot and the same acquisition session were assigned to the same subset, ensuring strict independence across training, validation, and test partitions. This grouping strategy guarantees that no images captured from identical field locations or time periods appear in multiple subsets. Model optimization was performed using a gradient-based optimization algorithm, with the initial learning rate set to α and dynamically adjusted based on validation performance during training to mitigate overfitting. The batch size was set to B to balance GPU memory usage and training stability, and the number of training epochs was set to E until convergence was observed on the validation set. To further enhance the reliability and robustness of evaluation, a 5-fold cross-validation strategy was introduced on the training set. In each fold, non-overlapping grouped subsets were maintained, thereby reducing partition randomness while preserving spatial–temporal independence. The training set was partitioned into five non-overlapping subsets, with one subset selected as validation in each round and the remaining subsets used for training. The final performance was reported as the average over the five runs. This configuration effectively reduced the influence of data partition randomness and yielded more robust and reliable performance evaluation.

4.1.2. Baseline Models and Evaluation Metrics

Multiple representative traditional methods and deep learning models were selected as baseline approaches to comprehensively evaluate the performance advantages of the proposed method in pest recognition tasks. The support vector machine model based on handcrafted features [49] constructs color, texture, and morphological descriptors and adopts the maximum margin classification criterion, thereby exhibiting stable discriminative capability under limited sample sizes. The random forest model [50] employs an ensemble learning mechanism based on multiple decision trees, enabling effective integration of diverse handcrafted features and improving adaptability to complex data distributions. Among deep learning baselines, the single-scale CNN [51] relies on an end-to-end feature learning framework to automatically extract discriminative representations from raw images, allowing effective characterization of local visual patterns of pests. ResNet [52] introduces residual connections to alleviate training difficulties in deep networks, thereby enhancing feature representation capacity while maintaining a deep architecture. Vision Transformer [53] adopts a self-attention mechanism to model global contextual relationships, which facilitates the capture of long-range semantic dependencies within images. The FPN-based multi-scale vision model [54] fuses features at different resolutions through a feature pyramid structure and demonstrates strong multi-scale modeling capability in small-target detection and recognition tasks. These baseline methods reflect the mainstream technical paradigms in current pest recognition research from different modeling perspectives and provide comprehensive and representative references for performance comparison with the proposed approach.
In the pest recognition experiments, multiple evaluation metrics were adopted to comprehensively assess both the overall classification performance and the recognition capability for small target pests. Accuracy was used to measure overall prediction correctness, Precision was employed to characterize the reliability of predictions for a specific pest category, Recall was used to reflect the detection capability for true pest samples, and the F 1 score was adopted to balance Precision and Recall, with particular emphasis placed on its performance for small-target pest categories. The mathematical definitions of the evaluation metrics are given as follows:
A c c u r a c y = T P + T N T P + T N + F P + F N ,
P r e c i s i o n = T P T P + F P ,
R e c a l l = T P T P + F N ,
F 1 = 2 × P r e c i s i o n × R e c a l l P r e c i s i o n + R e c a l l .
Here, T P denotes the number of samples correctly identified as the target pest, T N denotes the number of samples correctly identified as non-target pests, F P denotes the number of samples incorrectly identified as the target pest, and F N denotes the number of pest samples that are present but not detected by the model. These variables are derived from the confusion matrix and constitute the fundamental components for evaluating classification model performance.

4.2. Overall Performance Comparison with Baseline Methods

This experiment was designed to systematically evaluate the pest recognition performance of the proposed method in real agricultural scenarios from an overall perspective and to compare it with multiple representative traditional approaches and deep learning baselines. To ensure statistical rigor, performance differences among all methods were analyzed using one-way analysis of variance (ANOVA). When a statistically significant overall effect was observed, Tukey’s honestly significant difference (HSD) post hoc test was conducted for pairwise comparisons. The significance level was set to p < 0.05 , and statistically significant differences are explicitly indicated in Table 2 using superscript symbols.
As shown in Table 2 and Figure 5, the SVM and random forest methods based on handcrafted features achieve the lowest overall performance. Statistical analysis confirms that their results differ significantly from those of the proposed method ( p < 0.05 ). This limitation primarily arises from the fact that their feature spaces are constructed using fixed human-designed priors, which makes it difficult to characterize the high-dimensional nonlinear distributions exhibited by pests under complex backgrounds, especially in the presence of illumination variations, repetitive background textures, and extremely small-target scales, where decision boundaries between classes are prone to overlap. With the transition from traditional machine learning models to end-to-end deep learning approaches, the performance of single-scale CNN, ResNet, and Vision Transformer models gradually improves. Nevertheless, ANOVA results indicate that these improvements remain statistically inferior to the proposed multi-scale vision–sensor collaborative framework ( p < 0.05 ). After introducing multi-scale visual modeling, both the FPN-based method and the constructed multi-scale vision model achieve substantial improvements across all metrics. However, pairwise post hoc comparisons further demonstrate that incorporating environmental sensor priors yields statistically significant gains over purely vision-based multi-scale models. Consequently, the proposed method consistently achieves the highest Accuracy, Recall, and F 1 score, and the superiority over all baseline approaches is statistically validated under the defined significance criterion ( p < 0.05 ), thereby reinforcing both the empirical and statistical reliability of the results.

4.3. Recognition Performance Across Different Pest Categories

This experiment evaluates the model performance at the category level, aiming to analyze the stability and consistency of the proposed method across different pest types, and to examine whether the model exhibits significant bias toward specific sample distributions, morphological variations, or ecological habits. Unlike the overall performance comparison, this experiment emphasizes independent discriminative capability for each pest category rather than aggregated average results, thereby assessing whether the learned representations possess genuine generalization ability.
As shown in Table 3 and Figure 6, the model achieves relatively balanced performance across all six pest categories. High-frequency, small-scale pests with relatively stable morphology, such as aphids and spider mites, exhibit the highest Accuracy and F 1 scores, indicating that the model can effectively exploit multi-scale visual features to capture their characteristic textures and shapes. For pests such as thrips, whiteflies, and leafhoppers, which are more prone to confusion with background textures and whose appearances vary significantly across growth stages, performance shows a moderate decrease but remains at a high level overall, demonstrating robust behavior under complex backgrounds. The leaf beetle category yields the lowest metrics, which is closely related to its smaller sample size, larger pose variation, and frequent local occlusions. Nevertheless, no severe performance imbalance is observed, suggesting that the model maintains a reasonable level of adaptability to long-tail categories.
It should be noted that the last row of this table reports macro-averaged results obtained by assigning equal weights to each pest category, whereas the corresponding results in Table 2 are computed by aggregating all test samples directly. Due to inherent imbalances in sample size and recognition difficulty across pest categories, macro-averaged metrics are more sensitive to rare or difficult classes and therefore tend to be lower than sample-level overall metrics. From a theoretical standpoint, performance differences across pest categories are fundamentally determined by their separability in feature space. Aphids and spider mites tend to form compact and stable clusters under multi-scale feature representations, resulting in clearer decision boundaries, whereas pests such as thrips, whiteflies, and leaf beetles exhibit higher similarity to leaf textures and background noise, leading to greater overlap in high-dimensional feature space. By enhancing multi-scale discriminability and imposing ecological constraints through environmental prior modulation, the proposed method suppresses feature responses inconsistent with environmental conditions and reduces inter-class performance disparity, demonstrating stable generalization across multiple small-target pest categories.

4.4. Ablation Study on Key Modules

The ablation study was designed to systematically verify the role and necessity of each key component in the proposed framework, and to analyze how different architectural choices affect the discriminative capability of the model, thereby ensuring that performance gains are not solely attributed to increased model complexity or parameter count.
As shown in Table 4 and Figure 7, the complete model consistently achieves the best performance across all evaluation metrics, indicating strong complementarity among multi-scale visual modeling, environmental sensor prior modulation, and vision–sensor collaborative discrimination. Removing the multi-scale visual modeling module leads to the most pronounced performance degradation, demonstrating that single-scale representations are insufficient to capture the spatial complexity and scale variation of small pests, and that deep features alone lack adequate fine-grained expressiveness. In contrast, removing the sensor prior modulation or collaborative discrimination module results in relatively smaller but still substantial performance drops, highlighting that environmental information and cross-modal collaboration are not redundant but instead play a critical role in stabilizing decision boundaries in complex agricultural environments.
From a theoretical perspective, differences observed in the ablation results reflect fundamental changes in feature space construction. Without multi-scale modeling, the model performs discrimination within a single feature subspace, where feature distributions are more susceptible to overlap under scale variation and background interference, leading to unstable classification boundaries. The sensor prior modulation module constrains visual features through conditional environmental information, suppressing responses that are inconsistent with ecological conditions. Removing this module increases the degrees of freedom of the discriminative function in irrelevant feature directions, thereby introducing noisy activations. Replacing the vision–sensor collaborative structure with simple concatenation further reduces the model to linear multimodal fusion at the feature level, which lacks conditional alignment capability and prevents effective semantic coordination across modalities. Overall, the ablation study demonstrates that the proposed modules function synergistically to form an integrated framework, which is the fundamental reason why the complete model achieves superior performance in small-target pest recognition tasks.

4.5. Recognition Performance Under Crowded Colony Conditions

In practical agricultural environments, small-target pests frequently form dense colonies, where multiple individuals overlap within limited spatial regions. Such spatial aggregation and occlusion significantly increase intra-class texture redundancy and background confusion, posing additional challenges for recognition models. To explicitly evaluate model robustness under different colony densities, the original test set was further divided into two subsets according to manually annotated pest counts and spatial overlap levels: (1) a sparse subset, containing images with limited individual counts and weak spatial overlap; and (2) a dense subset, consisting of images characterized by severe aggregation and frequent occlusion. This partition was performed only on the test set and did not affect the training process. Four representative methods—including Vision Transformer (single-scale), FPN-based multi-scale vision (without sensor information), multi-scale vision without sensor prior modulation, and the proposed multi-scale vision–sensor collaborative framework—were evaluated separately on both subsets. The quantitative results are summarized in Table 5.
As shown in Table 5, all methods exhibit performance degradation under dense colony conditions compared to sparse scenarios. However, the magnitude of degradation differs substantially. Single-scale Vision Transformer shows a noticeable drop in both Accuracy and F 1 score, indicating sensitivity to severe overlap and texture crowding. Multi-scale visual modeling partially alleviates this issue by preserving high-resolution shallow features and cross-scale fusion, yet performance still declines when environmental priors are not incorporated. In contrast, the proposed method demonstrates significantly improved stability under dense conditions. Although a moderate performance decrease is observed compared to sparse samples, the degradation margin is smaller than that of all baseline methods. This improvement can be attributed to three factors: (1) multi-scale texture density modeling enhances discriminative representation of colony morphology; (2) cross-scale relational fusion captures spatial distribution structures across receptive fields; and (3) environmental prior modulation suppresses ecologically inconsistent background activations, thereby reducing noise amplification in crowded scenes. These results confirm that while the framework is designed for image-level recognition rather than instance-level segmentation, it maintains robust category discrimination capability even under severe aggregation and occlusion conditions.

4.6. Discussion

4.6.1. Practical Implications for Intelligent Pest Monitoring

The proposed method demonstrates clear practical value in real agricultural production by effectively addressing the challenges of early-stage small-target pest detection under complex field conditions. In open-field farmland and greenhouse facilities in Bayannur City, Inner Mongolia, pests such as aphids, thrips, and whiteflies often emerge with low visual saliency and small population sizes, making timely manual inspection difficult. By integrating fixed visual monitoring with environmental sensors, the proposed multi-scale vision–sensor framework enables continuous and non-intrusive surveillance, allowing subtle visual signs of pest occurrence to be detected reliably. Moreover, the environmental prior modulation mechanism dynamically adjusts visual attention in response to changing temperature, humidity, and illumination conditions, reducing false alarms caused by background variation and short-term disturbances, and thereby improving the stability of pest recognition in real production environments.
Although the proposed framework demonstrates stable performance within the collected dataset, it is important to acknowledge that the current data are obtained from a single geographic region. Therefore, the generalization capability of the model across heterogeneous climate zones, different crop systems, and varying pest ecological dynamics requires further validation. Future research will incorporate multi-regional data collection and cross-region evaluation protocols to systematically assess transferability and improve model robustness.

4.6.2. Implications for Agricultural Economics and Sustainable Management

Beyond technical performance, the proposed approach also exhibits important implications from an agricultural economics perspective. Accurate and early identification of pest occurrence provides a data-driven basis for precision pest management, enabling localized and timely control measures instead of broad, preventive pesticide application. This targeted intervention strategy can substantially reduce unnecessary pesticide input, lower chemical and labor costs, and mitigate negative environmental externalities, which collectively contribute to improved production efficiency and resource allocation in agricultural systems. In greenhouse production, early risk alerts allow growers to intervene before pest populations expand, reducing yield losses and avoiding costly large-scale treatments. In open-field agriculture, improved discrimination between true pests and background elements helps prevent unnecessary control actions, further reducing operational costs. By stabilizing crop yields and decreasing input variability, the proposed method supports income stability for farmers and enhances the economic resilience of regional agricultural production. From a long-term perspective, such intelligent monitoring technologies align with sustainable pest management strategies, facilitating the transition toward greener, more economically efficient plant protection practices.

4.7. Limitation and Future Work

Although the proposed multi-scale vision–sensor collaborative method achieves notable performance improvements in small-target pest recognition under complex agricultural scenarios, several aspects remain to be further explored. Crucially, from a biological standpoint, the current model leverages instantaneous weather and light conditions to correct recognition based on the acute behavioral responses of pests within a specific local microclimate. Consequently, the effect of this recognition correction will inevitably and radically change depending on the macro-weather conditions of different geographical areas and seasons. The environmental priors learned in this study are thus geographically and seasonally bounded. Furthermore, the current framework relies on coordinated data acquisition from fixed visual devices and environmental sensors, and its performance may be influenced by sensor deployment density and acquisition stability. Under extreme weather conditions or sensor occlusion, missing environmental information may affect the effectiveness of prior modulation. In addition, the proposed model focuses primarily on multi-scale feature modeling at the static image level, while pest migration, aggregation, and diffusion behaviors along the temporal dimension are not explicitly modeled. Pest population dynamics are fundamentally driven by cumulative and temporal climatic effects, such as temperature accumulation, long-term humidity fluctuations, and seasonal transitions, rather than just instantaneous conditions. The absence of explicit temporal environmental modeling constitutes a primary limitation of the present study. Future work will extend the proposed framework by incorporating time-series environmental data and continuous image sequences through temporal modeling architectures, such as Long Short-Term Memory networks or temporal Transformer models. This extension is expected to capture the dynamic evolution of pest occurrence and improve early warning timeliness. It should be noted that the environmental prior modeling in this study is based on instantaneous microclimate variables at the moment of image acquisition, without incorporating temporal-scale environmental dynamics. Although immediate temperature, humidity, and illumination conditions can reflect short-term behavioral responses of pests, diurnal variations, periodic climate fluctuations, and seasonal transitions may exert cumulative effects on pest population occurrence. Therefore, the absence of temporal modeling limits the framework’s ability to fully characterize ecological dynamics. Future work will integrate continuous time-series sensor data and temporal deep learning architectures to further enhance the ecological interpretability and long-term predictive capability of the multimodal collaborative framework.
It is also important to note that although the dataset spans April to November 2023 and covers major crops across multiple growth stages, the geographic scope remains concentrated within the Hetao Irrigation Area, a typical temperate arid and semi-arid agro-ecological zone characterized by large diurnal temperature variations and high summer temperatures with low precipitation. Pest behavioral responses and crop–environment interactions in this specific climatic context may differ substantially from those in the North China Plain, the humid middle–lower Yangtze River Basin, or southern protected agricultural systems. Therefore, the generalization capability of the current model under heterogeneous climate zones and different pest subspecies must be further validated. Future research will incorporate multi-site data collection across various climatic zones, combined with cross-regional validation and domain adaptation strategies to systematically evaluate the robustness and transferability of the proposed framework. In addition, integrating lightweight model design with edge computing techniques would facilitate deployment on field terminals, enabling low-latency online monitoring. Expanding the range of pest categories and exploring joint modeling with disease monitoring and crop growth assessment tasks would further contribute to the development of more comprehensive agricultural sensing systems. Additionally, acoustic sensing technologies have demonstrated considerable potential in detecting collective insect activity through the analysis of feeding sounds, wingbeat signals, and movement-induced vibrations [55]. By extracting temporal–frequency characteristics from acoustic signals, these approaches enable continuous monitoring of pest activity intensity and behavioral patterns under field conditions. In particular, spectral decomposition and energy distribution analysis allow for the identification of characteristic frequency bands associated with specific insect groups, thereby supporting early infestation detection and dynamic risk assessment [56]. Beyond purely acoustic approaches, distributed fiber-optic vibration monitoring systems have emerged as another promising sensing paradigm for large-scale agricultural environments [57]. Such systems leverage distributed acoustic sensing (DAS) principles to capture micro-vibration signals along extended fiber paths deployed across crop rows. Through spatiotemporal signal reconstruction, these technologies can achieve spatial localization of insect-induced disturbances, providing structural information about colony aggregation patterns and movement trajectories. Furthermore, vibration intensity profiling and spectral response modeling enable quantitative characterization of collective insect dynamics at the population level [58]. While these alternative sensing technologies offer valuable macro-level insights into pest activity distribution and colony-level behavior, they are typically optimized for activity detection, anomaly monitoring, or spatial localization rather than fine-grained species-level discrimination. In contrast, the proposed vision–sensor collaborative framework focuses on image-level morphological recognition under ecological constraints, explicitly modeling multi-scale visual texture representations and environmental priors to enhance discriminative robustness for small targets. Therefore, rather than functioning as competing methodologies, acoustic and distributed vibration sensing approaches can be regarded as complementary to visual perception systems. Acoustic and fiber-optic monitoring provide continuous, large-area structural awareness and dynamic colony information, whereas vision-based modeling enables precise categorical identification at the individual or image level. Future research will explore the integration of fiber-optic vibration signals, acoustic features, visual imagery, and environmental sensor data within a unified multimodal learning architecture, aiming to construct a spatially aware, hierarchically coordinated pest monitoring system that bridges colony-level dynamics and fine-grained visual recognition.

5. Conclusions

In complex agricultural production environments, achieving accurate and stable recognition of small-target pests constitutes a fundamental prerequisite for advancing precision plant protection management and green pest control. Owing to the small physical scale of pest individuals, strong background interference, and highly variable ecological conditions, traditional approaches based on handcrafted features or single-scale visual representations have difficulty achieving reliable performance in real-world scenarios. To address this challenge, a multi-scale vision–sensor collaborative pest recognition framework oriented toward real field and protected agriculture environments is constructed from the perspective of pest occurrence mechanisms and multimodal perceptual synergy, providing a new technical pathway for intelligent monitoring of small-target pests in complex scenes. The primary innovation of this study lies in the organic integration of multi-scale small-target visual modeling with environmental sensor priors, together with the realization of conditional representation learning through a vision–sensor collaborative discrimination mechanism. On the one hand, multi-scale visual feature modeling effectively alleviates the progressive attenuation of small-target representations in deep networks, thereby enhancing the perceptibility of fine-grained texture and morphological cues. On the other hand, environmental variables such as temperature, humidity, and illumination are introduced to modulate visual features as priors, enabling ecological constraints related to pest occurrence to be explicitly considered during the discrimination process and consequently improving recognition stability under complex backgrounds and varying environmental conditions. Experimental results demonstrate that the proposed method achieves approximately 93.1 % Accuracy, 92.0 % Precision, 91.2 % Recall, and 91.6 %   F 1 score on the test set, significantly outperforming traditional machine learning approaches and multiple deep learning baseline models. Furthermore, category-level evaluations and ablation studies confirm the balanced performance of the model across different pest types as well as the effectiveness of each key module, indicating that the proposed approach not only exhibits superior overall performance but also maintains strong robustness in scenarios involving small sample sizes or high inter-class confusion.

Author Contributions

Conceptualization, C.W., Y.C., S.C. and Y.S.; Data curation, R.H. and Z.X.; Formal analysis, R.C.; Funding acquisition, Y.S.; Investigation, R.C.; Methodology, C.W., Y.C. and S.C.; Project administration, Y.S.; Resources, R.H. and Z.X.; Software, C.W., Y.C. and S.C.; Supervision, Y.S.; Validation, R.C. and Z.X.; Visualization, R.H.; Writing—original draft, C.W., Y.C., S.C., R.C., R.H., Z.X. and Y.S.; C.W., Y.C., and S.C. contributed equally to this work. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Modern Agricultural Industrial Technology System Beijing Innovation Team (BAIC08-2025-YJ03).

Data Availability Statement

The data and source code presented in this study are openly available in a public GitHub repository https://github.com/Aurelius-04/MVCF.git, accessed on 26 February 2026.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Swarnkar, S.K.; Rathore, Y.K.; Swarnkar, V.K. Machine learning models for early detection of pest infestation in crops: A comparative study. In Smart Agriculture; CRC Press: Boca Raton, FL, USA, 2024; pp. 147–162. [Google Scholar]
  2. Ye, K.; Hu, G.; Tong, Z.; Xu, Y.; Zheng, J. Key intelligent pesticide prescription spraying technologies for the control of pests, diseases, and weeds: A review. Agriculture 2025, 15, 81. [Google Scholar] [CrossRef]
  3. Aziz, D.; Rafiq, S.; Saini, P.; Ahad, I.; Gonal, B.; Rehman, S.A.; Rashid, S.; Saini, P.; Rohela, G.K.; Aalum, K.; et al. Remote sensing and artificial intelligence: Revolutionizing pest management in agriculture. Front. Sustain. Food Syst. 2025, 9, 1551460. [Google Scholar] [CrossRef]
  4. Sharma, A.; Patel, R.K.; Pranjal, P.; Panchal, B.; Chouhan, S.S. Computer vision-based smart monitoring and control system for crop. In Applications of Computer Vision and Drone Technology in Agriculture 4.0; Springer: Berlin/Heidelberg, Germany, 2024; pp. 65–82. [Google Scholar]
  5. Batz, P.; Will, T.; Thiel, S.; Ziesche, T.M.; Joachim, C. From identification to forecasting: The potential of image recognition and artificial intelligence for aphid pest monitoring. Front. Plant Sci. 2023, 14, 1150748. [Google Scholar] [CrossRef] [PubMed]
  6. Obasekore, H.; Fanni, M.; Ahmed, S.M.; Parque, V.; Kang, B.Y. Agricultural robot-centered recognition of early-developmental pest stage based on deep learning: A case study on fall armyworm (Spodoptera frugiperda). Sensors 2023, 23, 3147. [Google Scholar] [CrossRef] [PubMed]
  7. Wang, S.; Xu, D.; Liang, H.; Bai, Y.; Li, X.; Zhou, J.; Su, C.; Wei, W. Advances in deep learning applications for plant disease and pest detection: A review. Remote Sens. 2025, 17, 698. [Google Scholar] [CrossRef]
  8. Rahman, W.; Hossain, M.M.; Hasan, M.M.; Iqbal, M.S.; Rahman, M.M.; Hasan, K.F.; Moni, M.A. Automated detection of harmful insects in agriculture: A smart framework leveraging IoT, machine learning, and blockchain. IEEE Trans. Artif. Intell. 2024, 5, 4787–4798. [Google Scholar] [CrossRef]
  9. Zhang, L.; Zhang, Y.; Ma, X. A new strategy for tuning ReLUs: Self-adaptive linear units (SALUs). In Proceedings of the ICMLCA 2021; 2nd International Conference on Machine Learning and Computer Application; VDE: Berlin, Germany, 2021; pp. 1–8. [Google Scholar]
  10. Ullah, N.; Khan, J.A.; Alharbi, L.A.; Raza, A.; Khan, W.; Ahmad, I. An efficient approach for crops pests recognition and classification based on novel DeepPestNet deep learning model. IEEE Access 2022, 10, 73019–73032. [Google Scholar] [CrossRef]
  11. Zhao, Y.; Xie, Q. Review of deep learning applications for detecting special components in agricultural products. Computers 2025, 14, 309. [Google Scholar] [CrossRef]
  12. Tang, Z.; Lu, J.; Chen, Z.; Qi, F.; Zhang, L. Improved Pest-YOLO: Real-time pest detection based on efficient channel attention mechanism and transformer encoder. Ecol. Inform. 2023, 78, 102340. [Google Scholar] [CrossRef]
  13. Yang, Z.; Ma, W.; Lu, J.; Tian, Z.; Peng, K. The application status and trends of machine vision in tea production. Appl. Sci. 2023, 13, 10744. [Google Scholar] [CrossRef]
  14. Wang, K.; Chen, K.; Du, H.; Liu, S.; Xu, J.; Zhao, J.; Chen, H.; Liu, Y.; Liu, Y. New image dataset and new negative sample judgment method for crop pest recognition based on deep learning models. Ecol. Inform. 2022, 69, 101620. [Google Scholar] [CrossRef]
  15. Zhang, L.; Sun, S.; Zhao, H.; Li, Z.; Li, D. Multi-species insect recognition method based on computer visions: Sustainable agricultural development. Ecol. Inform. 2025, 88, 103125. [Google Scholar] [CrossRef]
  16. Aarif, K.O.M.; Alam, A.; Hotak, Y. Smart sensor technologies shaping the future of precision agriculture: Recent advances and future outlooks. J. Sens. 2025, 2025, 2460098. [Google Scholar] [CrossRef]
  17. Lyu, Y.; Lu, F.; Wang, X.; Wang, Y.; Wang, Z.; Zhu, Y.; Wang, Z.; Dong, M. A CNN-Transformer Hybrid Framework for Multi-Label Predator–Prey Detection in Agricultural Fields. Sensors 2025, 25, 4719. [Google Scholar] [CrossRef] [PubMed]
  18. Zhang, H.; Wang, Y.; Li, F.; Wang, H. Motion Blur Robust Wheat Pest Damage Detection with Dynamic Fuzzy Feature Fusion. arXiv 2026, arXiv:2601.03046. [Google Scholar] [CrossRef]
  19. Vhatkar, K.N. An intellectual model of pest detection and classification using enhanced optimization-assisted single shot detector and graph attention network. Evol. Intell. 2025, 18, 3. [Google Scholar] [CrossRef]
  20. Seol, J.; Kim, C.; Ju, E.; Son, H.I. STPAS: Spatial-Temporal Filtering-based Perception and Analysis System for Precision Aerial Spraying. IEEE Access 2024, 12, 145997–146008. [Google Scholar] [CrossRef]
  21. Zhang, M.; Liu, C.; Li, Z.; Yin, B. From Convolutional Networks to Vision Transformers: Evolution of Deep Learning in Agricultural Pest and Disease Identification. Agronomy 2025, 15, 1079. [Google Scholar] [CrossRef]
  22. Domingues, T.; Brandão, T.; Ferreira, J.C. Machine learning for detection and prediction of crop diseases and pests: A comprehensive survey. Agriculture 2022, 12, 1350. [Google Scholar] [CrossRef]
  23. Qasim, M.; Shah, S.M.A.; Safi, Q.G.K.; Mahmood, D.; Iqbal, A.; Nauman, A.; Kim, S.W. An Adaptive Features Fusion Convolutional Neural Network for Multi-Class Agriculture Pest Detection. Comput. Mater. Contin. 2025, 83, 4429–4445. [Google Scholar] [CrossRef]
  24. Xu, X.; Li, H.; Gao, Q.; Zhou, M.; Meng, T.; Yin, L.; Chai, X. Visual attention focusing on fine-grained foreground and eliminating background bias for pest image identification. IEEE Access 2024, 83, 161732–161741. [Google Scholar] [CrossRef]
  25. Xu, W.; Sun, L.; Zhen, C.; Liu, B.; Yang, Z.; Yang, W. Deep learning-based image recognition of agricultural pests. Appl. Sci. 2022, 12, 12896. [Google Scholar] [CrossRef]
  26. Lin, J.; Chen, X.; Cai, J.; Pan, R.; Cernava, T.; Migheli, Q.; Zhang, X.; Qin, Y. Looking from shallow to deep: Hierarchical complementary networks for large scale pest identification. Comput. Electron. Agric. 2023, 214, 108342. [Google Scholar] [CrossRef]
  27. Ojo, M.O.; Zahid, A. Deep learning in controlled environment agriculture: A review of recent advancements, challenges and prospects. Sensors 2022, 22, 7965. [Google Scholar] [CrossRef] [PubMed]
  28. Karthik, R.; Ajay, A.; Bisht, A.S.; Illakiya, T.; Suganthi, K. A deep learning approach for crop disease and pest classification using Swin transformer and Dual-Attention Multi-scale fusion network. IEEE Access 2024, 12, 152639–152655. [Google Scholar] [CrossRef]
  29. Bonato, M.; Martin, E.A.; Cord, A.F.; Seppelt, R.; Beckmann, M.; Strauch, M. Applying generic landscape-scale models of natural pest control to real data: Associations between crops, pests and biocontrol agents make the difference. Agric. Ecosyst. Environ. 2023, 342, 108215. [Google Scholar] [CrossRef]
  30. Xiong, H.; Li, J.; Wang, T.; Zhang, F.; Wang, Z. EResNet-SVM: An overfitting-relieved deep learning model for recognition of plant diseases and pests. J. Sci. Food Agric. 2024, 104, 6018–6034. [Google Scholar] [CrossRef]
  31. Zhang, Y.; Lv, C. TinySegformer: A lightweight visual segmentation model for real-time agricultural pest detection. Comput. Electron. Agric. 2024, 218, 108740. [Google Scholar] [CrossRef]
  32. Zhang, Y.; Wa, S.; Liu, Y.; Zhou, X.; Sun, P.; Ma, Q. High-accuracy detection of maize leaf diseases CNN based on multi-pathway activation function module. Remote Sens. 2021, 13, 4218. [Google Scholar] [CrossRef]
  33. Wang, D.; Cao, W.; Zhang, F.; Li, Z.; Xu, S.; Wu, X. A review of deep learning in multiscale agricultural sensing. Remote Sens. 2022, 14, 559. [Google Scholar] [CrossRef]
  34. Li, Y.; Hu, X.; Guo, J.; Wang, Z.; Yu, J. Honeysuckle Pest Detection with a Pyramid Attention Network for Multi-Dimensional Feature Fusion. Symmetry 2026, 18, 118. [Google Scholar] [CrossRef]
  35. Dong, S.; Du, J.; Jiao, L.; Wang, F.; Liu, K.; Teng, Y.; Wang, R. Automatic crop pest detection oriented multiscale feature fusion approach. Insects 2022, 13, 554. [Google Scholar] [CrossRef]
  36. Wang, L.; Liu, Y.; Liu, J.; Wang, Y.; Xu, S. Based on the multi-scale information sharing network of fine-grained attention for agricultural pest detection. PLoS ONE 2023, 18, e0286732. [Google Scholar]
  37. Yang, H.; Sun, H.; Wang, K.; Yang, J.; Hasan Ali Baig, M. Enhanced Farmland Extraction from Gaofen-2: Multi-Scale Segmentation, SVM Integration, and Multi-Temporal Analysis. Agriculture 2025, 15, 1073. [Google Scholar] [CrossRef]
  38. MacLeod, N.; Canty, R.J.; Polaszek, A. Morphology-based identification of Bemisia tabaci cryptic species puparia via embedded group-contrast convolution neural network analysis. Syst. Biol. 2022, 71, 1095–1109. [Google Scholar] [CrossRef] [PubMed]
  39. Xu, C.; Yu, C.; Zhang, S.; Wang, X. Multi-scale convolution-capsule network for crop insect pest recognition. Electronics 2022, 11, 1630. [Google Scholar] [CrossRef]
  40. Li, W.; Luo, Y.; Jiang, P.; Dong, X.; Tang, K.; Liang, Z.; Shi, Y. A sustainable crop protection through integrated technologies: UAV-based detection, real-time pesticide mixing, and adaptive spraying. Sci. Rep. 2025, 15, 35748. [Google Scholar] [CrossRef]
  41. Courson, E.; Petit, S.; Poggi, S.; Ricci, B. Weather and landscape drivers of the regional level of pest occurrence in arable agriculture: A multi-pest analysis at the French national scale. Agric. Ecosyst. Environ. 2022, 338, 108105. [Google Scholar] [CrossRef]
  42. Wu, Q.; Zeng, J.; Wu, K. Research and application of crop pest monitoring and early warning technology in China. Front. Agric. Sci. Eng. 2022, 9, 19–36. [Google Scholar] [CrossRef]
  43. Khan, A.; Malebary, S.J.; Dang, L.M.; Binzagr, F.; Song, H.K.; Moon, H. AI-enabled crop management framework for pest detection using visual sensor data. Plants 2024, 13, 653. [Google Scholar] [CrossRef]
  44. Liu, Z.; Li, S.; Yang, Y.; Jiang, X.; Wang, M.; Chen, D.; Jiang, T.; Dong, M. High-Precision Pest Management Based on Multimodal Fusion and Attention-Guided Lightweight Networks. Insects 2025, 16, 850. [Google Scholar] [CrossRef] [PubMed]
  45. Lin, X.; Wa, S.; Zhang, Y.; Ma, Q. A dilated segmentation network with the morphological correction method in farming area image Series. Remote Sens. 2022, 14, 1771. [Google Scholar] [CrossRef]
  46. Zhou, X.; Chen, S.; Ren, Y.; Zhang, Y.; Fu, J.; Fan, D.; Lin, J.; Wang, Q. Atrous Pyramid GAN Segmentation Network for Fish Images with High Performance. Electronics 2022, 11, 911. [Google Scholar] [CrossRef]
  47. Wang, R.; Jiao, L.; Liu, K. Deep Learning for Agricultural Visual Perception: Crop Pest and Disease Detection; Springer: Singapore, 2023. [Google Scholar] [CrossRef]
  48. Amoah-Nuamah, J.; Child, B.; Okyere, E.Y.; Adams, O.; Danquah, J.A. Applications of artificial intelligence in forest health surveillance and management. Discov. For. 2025, 1, 56. [Google Scholar] [CrossRef]
  49. Chandra, M.A.; Bedi, S. Survey on SVM and their application in image classification. Int. J. Inf. Technol. 2021, 13, 1–11. [Google Scholar] [CrossRef]
  50. Biau, G. Analysis of a random forests model. J. Mach. Learn. Res. 2012, 13, 1063–1095. [Google Scholar]
  51. Chua, L.O. CNN: A vision of complexity. Int. J. Bifurc. Chaos 1997, 7, 2219–2425. [Google Scholar] [CrossRef]
  52. Liang, J. Image classification based on RESNET. J. Phys. Conf. Ser. 2020, 1634, 012110. [Google Scholar] [CrossRef]
  53. Han, K.; Wang, Y.; Chen, H.; Chen, X.; Guo, J.; Liu, Z.; Tang, Y.; Xiao, A.; Xu, C.; Xu, Y.; et al. A survey on vision transformer. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 45, 87–110. [Google Scholar] [CrossRef]
  54. Li, Z.; Li, E.; Xu, T.; Samat, A.; Liu, W. Feature alignment FPN for oriented object detection in remote sensing images. IEEE Geosci. Remote Sens. Lett. 2023, 20, 6001705. [Google Scholar] [CrossRef]
  55. Turov, A.T.; Konstantinov, Y.A.; Totmina, E.E.; Votinova, A.G.; Masich, G.F.; Korobko, D.A.; Fotiadi, A.A. Registration of Sounds Emitted by the Madagascar Hissing Cockroach Using a Distributed Acoustic Sensor. Sensors 2025, 25, 2101. [Google Scholar] [CrossRef]
  56. Ashry, I.; Mao, Y.; Al-Fehaid, Y.; Al-Shawaf, A.; Al-Bagshi, M.; Al-Brahim, S.; Ng, T.K.; Ooi, B.S. Early detection of red palm weevil using distributed optical sensor. Sci. Rep. 2020, 10, 3155. [Google Scholar] [CrossRef]
  57. Wang, B.; Mao, Y.; Ashry, I.; Al-Fehaid, Y.; Al-Shawaf, A.; Ng, T.K.; Yu, C.; Ooi, B.S. Towards Detecting Red Palm Weevil Using Machine Learning and Fiber Optic Distributed Acoustic Sensing. Sensors 2021, 21, 1592. [Google Scholar] [CrossRef]
  58. Ashry, I.; Wang, B.; Mao, Y.; Sait, M.; Guo, Y.; Al-Fehaid, Y.; Al-Shawaf, A.; Ng, T.K.; Ooi, B.S. CNN–Aided Optical Fiber Distributed Acoustic Sensing for Early Detection of Red Palm Weevil: A Field Experiment. Sensors 2022, 22, 6491. [Google Scholar] [CrossRef]
Figure 1. Instances of crops and common insect pests encompassed within the dataset.
Figure 1. Instances of crops and common insect pests encompassed within the dataset.
Insects 17 00281 g001
Figure 2. Illustration of the multi-scale small-target visual feature modeling module.
Figure 2. Illustration of the multi-scale small-target visual feature modeling module.
Insects 17 00281 g002
Figure 3. Illustration of the environmental sensor prior modulation module.
Figure 3. Illustration of the environmental sensor prior modulation module.
Insects 17 00281 g003
Figure 4. Illustration of the vision–sensor collaborative discrimination and classification module.
Figure 4. Illustration of the vision–sensor collaborative discrimination and classification module.
Insects 17 00281 g004
Figure 5. Overall performance comparison of different methods of pest recognition in terms of Accuracy, Precision, Recall, and F1 score.
Figure 5. Overall performance comparison of different methods of pest recognition in terms of Accuracy, Precision, Recall, and F1 score.
Insects 17 00281 g005
Figure 6. Performance distribution of different pest categories in terms of Accuracy, Precision, Recall, and F1 score.
Figure 6. Performance distribution of different pest categories in terms of Accuracy, Precision, Recall, and F1 score.
Insects 17 00281 g006
Figure 7. Performance trends of different model variants in the ablation study across Accuracy, Precision, Recall, and F1 score.
Figure 7. Performance trends of different model variants in the ablation study across Accuracy, Precision, Recall, and F1 score.
Insects 17 00281 g007
Table 1. Statistics of the multimodal pest dataset.
Table 1. Statistics of the multimodal pest dataset.
Data Type/Pest CategoryNumber of SamplesAcquisition ScenarioAcquisition Period
Aphid3180Field and greenhouse2023.04–2023.11
Thrips2460Greenhouse2023.04–2023.11
Whitefly1985Greenhouse2023.05–2023.10
Leafhopper1540Field2023.05–2023.09
Spider mite1870Field and greenhouse2023.06–2023.10
Leaf beetle1095Field2023.04–2023.08
Total pest image data12,130Field and greenhouse2023.04–2023.11
Air temperature data34,560Field and greenhouse2023.04–2023.11
Relative humidity data34,560Field and greenhouse2023.04–2023.11
Illumination intensity data25,920Field and greenhouse2023.05–2023.10
Table 2. Overall performance comparison of different methods on the test set (average results ± standard deviation of 5-fold cross-validation; statistical significance evaluated by one-way ANOVA followed by Tukey’s HSD post hoc test, significance level p < 0.05 ).
Table 2. Overall performance comparison of different methods on the test set (average results ± standard deviation of 5-fold cross-validation; statistical significance evaluated by one-way ANOVA followed by Tukey’s HSD post hoc test, significance level p < 0.05 ).
MethodAccuracy (%)Precision (%)Recall (%) F 1 (%)
SVM + handcrafted features78.4 ± 1.5 *76.9 ± 1.6 *74.2 ± 1.8 *75.5 ± 1.6 *
Random forest + handcrafted features80.1 ± 1.4 *78.6 ± 1.5 *76.8 ± 1.5 *77.7 ± 1.4 *
CNN (single-scale)84.7 ± 1.2 *83.2 ± 1.3 *81.5 ± 1.4 *82.3 ± 1.3 *
ResNet (single-scale)86.9 ± 1.0 *85.6 ± 1.1 *83.8 ± 1.2 *84.7 ± 1.1 *
Vision Transformer (single-scale)87.4 ± 1.1 *86.1 ± 1.0 *84.2 ± 1.1 *85.1 ± 1.0 *
FPN-based multi-scale vision (no sensor)89.6 ± 0.9 *88.3 ± 0.8 *87.1 ± 1.0 *87.7 ± 0.9 *
Multi-scale vision (w/o sensor prior)90.4 ± 0.8 *89.1 ± 0.9 *88.2 ± 0.8 *88.6 ± 0.8 *
Proposed method93.1 ± 0.592.0 ± 0.691.2 ± 0.591.6 ± 0.5
* Indicates a statistically significant difference compared to the proposed method ( p < 0.05 , one-way ANOVA).
Table 3. Classification performance of six pest categories (test set, average results of 5-fold cross-validation).
Table 3. Classification performance of six pest categories (test set, average results of 5-fold cross-validation).
Pest CategoryAccuracy (%)Precision (%)Recall (%) F 1 (%)
Aphid94.293.592.893.1
Thrips92.691.390.590.9
Whitefly91.890.689.790.1
Leafhopper92.191.090.290.6
Spider mite93.492.291.591.8
Leaf beetle90.789.488.689.0
Macro-average92.591.390.690.9
Table 4. Ablation study results of key modules (test set, average results ± standard deviation of 5-fold cross-validation; statistical significance assessed by one-way ANOVA followed by Tukey’s HSD post hoc test).
Table 4. Ablation study results of key modules (test set, average results ± standard deviation of 5-fold cross-validation; statistical significance assessed by one-way ANOVA followed by Tukey’s HSD post hoc test).
VariantAccuracy (%)Precision (%)Recall (%) F 1 (%)
Full model (proposed)93.1 ± 0.592.0 ± 0.691.2 ± 0.591.6 ± 0.5
w/o multi-scale visual modeling88.3 ± 1.0 *86.9 ± 1.1 *85.7 ± 1.2 *86.3 ± 1.1 *
w/o sensor prior modulation90.4 ± 0.8 *89.1 ± 0.9 *88.2 ± 0.8 *88.6 ± 0.8 *
w/o vision–sensor collaborative head89.6 ± 0.9 *88.2 ± 1.0 *87.1 ± 0.9 *87.6 ± 0.9 *
Replace conditional gating with concatenation89.1 ± 0.9 *87.8 ± 0.8 *86.9 ± 1.0 *87.3 ± 0.9 *
* Indicates a statistically significant difference compared to the full model ( p < 0.05 , one-way ANOVA).
Table 5. Recognition performance under different pest colony density conditions (sparse vs. dense subsets of the test set).
Table 5. Recognition performance under different pest colony density conditions (sparse vs. dense subsets of the test set).
MethodDensityAccuracy (%)Recall (%) F 1 (%)
Vision Transformer (single-scale)Sparse88.685.986.8
Vision Transformer (single-scale)Dense84.981.783.1
FPN-based multi-scale (no sensor)Sparse90.888.389.2
FPN-based multi-scale (no sensor)Dense87.484.685.9
Multi-scale vision (w/o sensor prior)Sparse91.689.790.2
Multi-scale vision (w/o sensor prior)Dense88.886.187.3
Proposed methodSparse93.892.392.5
Proposed methodDense92.290.891.0
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wang, C.; Chen, Y.; Chen, S.; Chen, R.; Xia, Z.; Hu, R.; Song, Y. A Multi-Scale Vision–Sensor Collaborative Framework for Small-Target Insect Pest Management. Insects 2026, 17, 281. https://doi.org/10.3390/insects17030281

AMA Style

Wang C, Chen Y, Chen S, Chen R, Xia Z, Hu R, Song Y. A Multi-Scale Vision–Sensor Collaborative Framework for Small-Target Insect Pest Management. Insects. 2026; 17(3):281. https://doi.org/10.3390/insects17030281

Chicago/Turabian Style

Wang, Chongyu, Yicheng Chen, Shangshan Chen, Ranran Chen, Ziqi Xia, Ruoyu Hu, and Yihong Song. 2026. "A Multi-Scale Vision–Sensor Collaborative Framework for Small-Target Insect Pest Management" Insects 17, no. 3: 281. https://doi.org/10.3390/insects17030281

APA Style

Wang, C., Chen, Y., Chen, S., Chen, R., Xia, Z., Hu, R., & Song, Y. (2026). A Multi-Scale Vision–Sensor Collaborative Framework for Small-Target Insect Pest Management. Insects, 17(3), 281. https://doi.org/10.3390/insects17030281

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop