Next Article in Journal
MicroRNA319-TCP19-IAA3.2 Module Mediates Lateral Root Growth in Populus tomentosa
Previous Article in Journal
Spatial and Temporal Inconsistency of Forest Resilience and Forest Vegetation Greening in Southwest China Under Climate Change
Previous Article in Special Issue
KBNet: A Language and Vision Fusion Multi-Modal Framework for Rice Disease Segmentation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Detection of Pine Wilt Disease in UAV Remote Sensing Images Based on SLMW-Net

1
Institute of Artificial Intelligence Application, Central South University of Forestry and Technology, Changsha 410004, China
2
Hunan Academy of Forestry, Changsha 410018, China
*
Authors to whom correspondence should be addressed.
Plants 2025, 14(16), 2490; https://doi.org/10.3390/plants14162490
Submission received: 29 July 2025 / Revised: 6 August 2025 / Accepted: 8 August 2025 / Published: 11 August 2025
(This article belongs to the Special Issue Advances in Artificial Intelligence for Plant Research)

Abstract

The pine wood nematode is responsible for pine wilt disease, which poses a significant threat to forest ecosystems worldwide. If not quickly detected and removed, the disease spreads rapidly. Advancements in UAV and image detection technologies are crucial for disease monitoring, enabling efficient and automated identification of pine wilt disease. However, challenges persist in the detection of pine wilt disease, including complex UAV imagery backgrounds, difficulty extracting subtle features, and prediction frame bias. In this study, we develop a specialized UAV remote sensing pine forest ARen dataset and introduce a novel pine wilt disease detection model, SLMW-Net. Firstly, the Self-Learning Feature Extraction Module (SFEM) is proposed, combining a convolutional operation and a learnable normalization layer, which effectively solves the problem of difficult feature extraction from pine trees in complex backgrounds and reduces the interference of irrelevant regions. Secondly, the MicroFeature Attention Mechanism (MFAM) is designed to enhance the capture of tiny features of pine trees infected by initial nematode diseases by combining Grouped Attention and Gated Feed-Forward. Then, Weighted and Linearly Scaled IoU Loss (WLIoU Loss) is introduced, which combines weight adjustment and linear stretch truncation to improve the learning strategy, enhance the model performance and generalization ability. SLMW-Net is trained on the self-built ARen dataset and compared with seven existing methods. The experimental results show that SLMW-Net outperforms all other methods, achieving an mAP@0.5 of 86.7% and an mAP@0.5:0.95 of 40.1%. Compared to the backbone model, the mAP@0.5 increased from 83.9% to 86.7%. Therefore, the proposed SLMW-Net has demonstrated strong capabilities to address three major challenges related to pine wilt disease detection, helping to protect forest health and maintain ecological balance.

1. Introduction

Pine forests are an extremely important component of forest ecosystems [1], playing a vital role in promoting biodiversity, soil and water conservation, and climate regulation [2]. Additionally, pine trees provide various resources such as timber and resin, possessing significant economic value and being widely utilized in industries such as construction, papermaking, and chemicals [3]. However, pine wilt disease is commonly referred to as cancer in pine forests. It is caused by a pathogen pine wood nematode and is capable of killing trees quickly and rapidly [4]. Vectors such as the infected Serragon genus of insects, like Monochamus alternatus, are responsible for transmitting the disease. Pine wilt disease is highly contagious [5]. If a pine tree is infected by a nematode, it quickly spreads throughout in the trunk of the pine tree and kills it by destroying the movement of water and nutrients. Pine wilt disease is spreading at explosive speed [6], and tends to devastate pine forests within three to five years [7]. Pine wilt disease represents a critical threat to the sustainable management of pine forests worldwide, posing particularly significant concerns across regions such as Asia, Europe, and North America [8]. Given its high mortality rate and strong transmissibility, early detection and diagnosis are essential. Only with accurate and efficient detection technologies can timely control measures be implemented, thereby reducing the ecological and economic impact of the disease [9].
Traditional detection of pine wilt disease has primarily relied on manual field surveys conducted by forestry professionals through on-site inspections [10]. Although such approaches provide direct and intuitive observations, they are constrained by several inherent limitations, including high labor intensity, sensitivity to complex terrain, and low operational efficiency. These drawbacks severely restrict their scalability for large-scale forest monitoring and hinder timely and accurate diagnosis [11]. To overcome these limitations, advancements in machine learning have spurred increasing interest in integrating UAV remote sensing imagery with classical machine learning algorithms, aiming to enable automated detection and assessment of pine wilt disease caused by the pine wood nematode [12,13,14,15]. While these approaches have enhanced automation to a certain extent, they remain dependent on manually engineered features and the selection of appropriate classifiers, limiting their adaptability to heterogeneous and dynamic forest environments and thus constraining their practical application [16]. In recent years, the rapid evolution of deep learning techniques has led to their growing application in pine wilt disease detection. Compared with traditional methodologies, deep learning models possess the capability to automatically extract high-level features from large-scale raw datasets, rendering them more robust in handling complex backgrounds and diverse forest scenarios [17]. This capacity offers substantial technical support for the intelligent and automated detection of forest diseases, greatly promoting their implementation in real-world forest management practices.
At present, object detection approaches based on deep learning are generally divided into two main categories: two-stage and one-stage models. Two-stage models, represented by the R-CNN series [18], usually begin by generating candidate regions, followed by separate classification and bounding box regression. These models offer high detection accuracy but are limited by inference speed and computational efficiency. In contrast, one-stage detection models, such as the SSD series proposed by Liu, W et al. [19] and the YOLO series by Redmon, J et al. [20] are more widely applicable in practical applications due to their advantages of simple structure and fast speed, and are particularly suitable for tasks with high real-time requirements. As a representative one-stage detection framework, the YOLO series has undergone continuous optimization across versions and has become a mainstream general-purpose object detection tool. As early as 2017, YOLOv2 [21] attracted widespread attention for its superior feature extraction capabilities and high detection efficiency. Subsequently, Wu, K et al. [22] improved YOLOv3 for pine wilt disease detection by introducing the Darknet backbone, a logical classifier, and the BCE loss function, significantly enhancing the accuracy of infected tree identification. Building on this, YOLOv4 [23] further optimized its architecture by incorporating a BoF and BoS, including data augmentation techniques such as CutMix, CutOut, and Mixup. These enhancements improved detection accuracy without significantly increasing inference cost. The model also integrated non-linear activation functions and skip connections, boosting its feature representation capabilities. Based on this framework, Sun, Z et al. [24] expanded the backbone network by integrating the Inceptionv2 structure, achieving improvements in both detection efficiency and accuracy while reducing the number of parameters. With continued research, Wang, L et al. [25] integrated the SE-Net into the YOLOv5 framework, significantly enhancing its feature extraction capability. Additionally, they introduced a BiFPN to improve perception and adaptability to multi-scale objects in complex scenes. Building on this, Ye, X et al. [26] combined YOLOv5 with the StrongSORT algorithm to realize dynamic visual tracking and counting of infected pine trees, expanding the model’s practical application in intelligent forestry. The subsequent YOLOv7 [27,28] introduced the E-ELAN, which significantly improved feature learning efficiency and training performance while maintaining model lightweight characteristics. Following this, YOLOv8 [29] further innovated by incorporating anchor-free detection mechanisms, the C2f module, an improved decoupled head structure, and a more efficient loss function. These improvements led to a notable increase in detection accuracy while preserving the model’s strength in real-time inference performance. As a result, research on optimizing YOLOv8 continues to emerge, aiming to further improve its performance for specific tasks. For example, Xu, S et al. [30] enhanced the convolutional network’s expressive power by introducing ODConv, designed a DyHead module for precise PWD feature extraction, and adopted the MPDioU regression mechanism to improve the accuracy and efficiency of bounding box prediction. Additionally, Yao, J et al. [31] developed the Pine-YOLO model based on YOLOv8, integrating DSConv and MCA, which significantly improved the robustness in complex forestry scenarios. Building on this series of advancements, YOLOv11 [32] achieved further breakthroughs in architecture design by incorporating the C3k2 module, the SPPF structure, and the C2PSA component. These enhancements substantially improved feature extraction capabilities and overall performance, providing stronger technical support for object detection tasks.
Overall, despite significant progress in detecting pine wilt disease, several challenges remain. First, remote sensing images captured by UAV often have complex backgrounds: Changes in lighting, the presence of image noise, and visual similarities between unrelated objects such as buildings, water bodies, and affected trees pose challenges to accurately identifying pine trees in the scene. These factors can increase the possibility of detection errors and hinder the overall effectiveness of the model. Second, it remains difficult to extract the subtle features of infected pine trees. The early symptoms of pine wilt disease are often mild, and the pathological changes on the bark and branches are hard to recognize. Due to the lack of effective attention mechanisms, existing models struggle to accurately capture and identify these subtle disease features. Third, prediction box deviations are common. When infected pine trees are densely distributed, conventional detection algorithms often generate a large number of redundant prediction boxes, and the loss functions fail to account for directional mismatches between predicted and ground-truth boxes. This may lead to inaccurate localization of infected trees, thereby reducing training efficiency and diminishing detection accuracy.
In light of these challenges, our primary research objective is to address three key issues in pine wilt disease detection: severe background interference, difficulty in extracting subtle features, and deviations in predicted bounding boxes. To achieve this objective, we selected the YOLOv11 model, which is known for its strong accuracy and adaptability, as the baseline network. Building upon this foundation, we developed a new detection model called SLMW-Net for identifying pine wilt disease. The main contributions of this study are summarized as follows:
(a)
A high-resolution ARen dataset is developed to include trees infected by pine wood nematodes. The dataset consists of 750 annotated UAV images for training and testing purposes. This dataset captures a wide range of forest conditions, such as bare soil, red broadleaf species, dead trees, and various background disturbances, ensuring a comprehensive representation of complex environmental factors.
(b)
A Self-Learning Feature Extraction Module (SFEM) is proposed that contains a convolution block and a learnable normalization layer. This design improves the discriminatory representation of pine wilt disease features through normalization while effectively retaining the original local details of the input. It is very efficient at extracting characteristics of nematode-infected pine trees, even in scenarios with complex ground conditions and significant vegetation interference.
(c)
A MicroFeature Attention Mechanism (MFAM) is introduced, combining Grouped Attention with a Gated Feed-Forward network. This method significantly improves the ability to capture pine wood nematode characteristics and also enhances the overall accuracy of feature representation. By overcoming the limitations of traditional attention mechanisms in detecting microscopic disease characteristics, this method greatly improves the accuracy of pine wood nematode detection.
(d)
A Weighted and Linearly Scaled IoU Loss (WLIoU Loss) is designed to further enhance the training process. It is specifically tailored for pine trees infected with pine wilt disease. By regulating multifaceted factors, the WLIoU Loss function surpasses category imbalance and hard sample detection challenges and enhances positive sample weighting by stretching and truncation mechanisms, thereby effectively addressing the issue of biased prediction boxes.

2. Datasets and Methods

2.1. Data Acquisition

The ARen dataset for pine wilt disease used in this study was obtained from the Hunan Academy of Forestry, China. The dataset collection area is located in Anren County, Chenzhou City, Hunan Province, with geographic coordinates ranging from 113°05′ to 113°36′ east longitude and 26°17′ to 26°51′ north latitude. This region, located in the southeast of Hunan Province, has a typical subtropical monsoon climate. The main forest types include slow-growing broadleaf forests, horsetail pine, wetland pine, and others, with pine forests being the predominant type. The infection severity of pine trees is notably high, making this area a typical representative region for pine wilt disease. The pine trees within the study area differ in age and are unevenly distributed both vertically and horizontally. In terms of vertical structure, the forest is stratified into three distinct layers: the top layer features tall, mature pines forming the canopy; the middle layer includes moderately sized pines along with various other tree species; and the bottom layer is primarily occupied by shrubs and herbaceous vegetation. In terms of horizontal structure, the trees in the forest are unevenly distributed, with alternating dense clusters and open spaces, highlighting the complexity and diversity of the forest structure. Data collection took place between September and October 2024. During this period, most pine trees infected with pine wilt disease exhibited typical reddish-brown symptoms, while other broadleaf trees showed no significant changes. As such, this period provided the best opportunity for capturing the most visible symptoms of pine wilt disease. We used the CW-15 UAV as the flight platform, with a flight altitude of approximately 500 m and a ground sampling distance (GSD) of 10 cm. It was equipped with a full-frame 61-megapixel sensor and a 35 mm focal length lens. Figure 1 illustrates the data acquisition.

2.2. Data Processing

Due to the wide coverage of the UAV-captured imagery, each original image was cropped to 800 × 800 pixel tiles. Under the professional guidance of forestry experts, 750 image patches were selected after filtering, retaining those containing areas of affected pine trees and excluding images in which vegetation features could not be accurately identified due to factors such as cloud cover, strong shadows, or image blurring. These images were then randomly divided into training and validation sets in a ratio of 8:2, with the training set containing 600 images and the validation set containing 150 images. For image annotation, forestry experts manually labeled the data using the LabelImg software 1.8.6. Specifically, images of pine wilt disease were imported into the LabelImg platform, where experienced foresters delineated diseased regions through manual annotation. The resulting labels were subsequently saved in txt format.

2.3. SLMW-Net

In this research, we introduce SLMW-Net, a target detection method based on YOLOv11, designed to efficiently and accurately detect pine wilt disease in natural settings. The model integrates the SFEM and MFAM modules and employs the WLIoU Loss to compute model loss, guiding the optimization of model parameters and thereby enhancing training effectiveness. These improvements collectively boost the efficiency of pine wilt disease detection, ensuring accurate recognition even under challenging conditions such as complex backgrounds or subtle disease features. As depicted in Figure 2, its architecture is composed of three principal parts: the backbone, the neck, and multiple head detection modules. Inside the backbone, we add the SFEM block after three C3k2 blocks to enhance feature extraction, aiding in the capture of characteristic information specific to pine wilt disease and addressing the challenges posed by complex backgrounds. Following processing by the SPPF block, the feature map is passed into the MFAM module to further refine the representation of input features. By applying Grouped Attention, Gated Feed-Forward, and depthwise separated convolution, the model can capture subtle disease features more effectively, focusing on regions with minor disease manifestations, which significantly improves detection accuracy. In the neck, prior to the output of each detection head, we incorporate the Gated Transformer Module to further enhance feature fusion, enabling the network to learn increasingly complex representations of disease features.

2.3.1. Self-Learning Feature Extraction Module (SFEM)

In the detection of pine wilt disease, the convolution operation serves as the core computational module in neural networks. It selects the most significant local features by filtering within the receptive field [33], which are critical for disease identification [34]. However, in UAV remote sensing images, the complex and variable background makes it challenging for traditional single-scale convolutions to capture features across different spatial scales, resulting in incomplete information extraction. Additionally, commonly used normalization techniques often lack sufficient adaptability to image noise and illumination variations, further limiting model performance. To address it, we introduce a self-learning feature extraction module (SFEM), designed to enhance feature representation and improve the efficiency of subsequent deployment. SFEM comprises a convolutional layer and a learnable normalization layer, which together progressively strengthen the ability to extract pine wilt disease-related features across multi-layer feature maps. This approach not only preserves the original local details of the inputs but also reconstructs a more discriminative feature representation through normalization. The design is particularly effective in environments with complex ground conditions and significant vegetation disturbance.
To capture feature information from the input image, we apply SFEM prior to the feature output at each scale within the backbone of YOLOv11. The SFEM structure is depicted in Figure 2b, and its specific processing flow is as follows. The feature extraction of the base convolution is performed first, where X denotes the input feature map, denotes the 2D convolution, W denotes the convolution kernel parameter, and Y denotes the convolution output.
Y = C o n v X = W X
Subsequently, a learnable reparameterization batch normalization L e a B N is introduced in the normalization process, where μ c and σ c 2 denote the mean and variance of each channel. γ c and β c denote the scaling and translation parameters of each channel. α is a self-learning parameter, initially set to 1, and is optimized during training to control the proportion of direct inputs and retain feature information. Specifically, α will be added to the model’s list of trainable parameters, incorporated into the computation graph during forward propagation, participate in backpropagation, and subsequently be updated automatically based on the gradients.
L e a B N Y C , H , W = γ c × Y C , H , W μ c σ c 2 + ε + β c + α × Y C , H , W
Z = L e a B N Y C , H , W
Finally, activation is performed using the SiLU function to obtain nonlinear relationships of the input features, so that our model can obtain more details about the features.
O u t p u t = S i L U Z = Z × σ Z
In conclusion, this module enhances the local features of the diseased spots during feature extraction, effectively suppresses the random noise in the background, improves the ability to focus on the pine wood nematode-infested area, and improves the model performance.

2.3.2. MicroFeature Attention Mechanism (MFAM)

Some infected pine trees initially exhibit subtle signs of nematode infection, which are often challenging for traditional models to detect due to the difficulty in extracting these minute features. Consequently, accurately identifying and extracting these subtle traits is essential for effectively detecting pine wilt disease. Drawing inspiration from human visual processing mechanisms, attention mechanisms enable neural networks to more accurately extract key features by assigning weighted coefficients to them [35]. Attention mechanisms are now commonly integrated into convolutional neural networks (CNNs) [36], graph convolutional networks (GCNs) [37], and other neural network architectures. As attention mechanisms continue to develop, traditional methods have shown significant limitations. These methods often focus on wider areas and tend to ignore key details, especially when identifying subtle signs of pine wilt disease. To solve these problems, we introduced the MicroFeature Attention Mechanism (MFAM). This module has significant advantages in identifying complex and diverse features and can accurately detect weak symptoms of pine wilt disease while maintaining high adaptability and robustness.
Figure 2c depicts the architecture of MFAM. Clearly visible is that MFAM employs a number of important such as the Gated Transformer Module (GTM), Grouped Attention (GA), Gated Feed-Forward (GFF), and Depthwise Separable Convolution (DWConv). By leveraging the collaborative operation of these modules, MFAM effectively responded to the challenge of detecting subtle features in the early stages of pine wilt disease. It significantly improves the characterization of these minor disease features in UAV remote sensing images, allowing the model to capture disease-related information more accurately and in detail. Specifically, the GTM combines the benefits of Grouped Attention and the Gated Feed-Forward network. The GA module strengthens attention to subtle features by dynamically adjusting attention weights, thereby highlighting key disease indicators and filtering out irrelevant information. The GFF module helps to better capture the spatial details of pine wilt disease characteristics, gradually extracts edge, texture, and shape information through hidden channels, while abstracting more complex features at a deeper level. This contributes to enhanced detection accuracy and robustness, particularly for early stage disease features. Furthermore, DWConv reduces information loss and computational complexity while enhancing feature representation, allowing the network to efficiently process large-scale remote sensing images. The mish activation function mitigates issues of gradient vanishing and dead neurons, bolstering feature learning capability. Its smooth gradient and nonlinear transformation enhance the ability to capture early symptoms of pine wilt disease. The processing flow of MFAM is outlined below.
The MFAM uses a staged partial structure that allows only a portion of the channels to pass through the complex GTM, which reduces the amount of computation while retaining a portion of the original feature channels, which helps information fusion and the flow of the gradient. In the MFAM, the input is X B × C × H × W , the input is first processed with ascending convolution, followed by dimensional splitting, and then one part is left unprocessed, and the other part goes into the GTM after processing. The two parts are then spliced, and finally, dimensionality reduction is performed to return to the original feature map size. Where D X is the result of dimension upgrading, A and B are the two parts after splitting according to the channel, with the same dimension of X . The results of splicing and downscaling is C X .
D X = C o n v 1 × 1 i n 2 C X
D X = [ A ; B ]
B = G T M B
C X = C o n v 1 × 1 2 C C [ A ; B ]
The Gated Transformer Module (GTM) consists of Grouped Attention (GA) and Gated Feed-Forward (GFF). The processing flow is outlined below. For the input feature B generates Q , K , V , where D W C o n v represents depthwise separated convolution and c h u n k divides the channels into five parts, each with the channel number of C .
Q K V r a w = D W C o n v 3 × 3 C o n v 1 × 1 C 5 C B B × 5 C × H × W
Q 1 , K 1 , Q 2 , K 2 , V = c h u n k Q K V r a w , 5 ; along   channel
Sorting is then performed by first flattening V to B × C × ( H × W ) , in ascending order. Then, Q 1 , K 1 , Q 2 , and K 2 are reordered by indexing operations to ensure that they are consistent with V .
V s o r t e d , i d x = s o r t V , dim = 2
The attention calculation is tuned by the temperature parameter, which controls the sensitivity of the similarity between the features, performs dynamic range adjustment, and calculates the result O by multiplying with V s o r t e d . The role of the temperature parameter is to control the degree of smoothing of the scores. A higher temperature results in a smoother attention distribution, while a lower temperature leads to a sharper attention distribution. d k indicates the size of the key vector.
A t t e n t i o n = Q × K T d K × t e m p e r a t u r e
O = A t t e n t i o n × V s o r t e d
Subsequently, a scatter operation is performed based on the previously obtained i d x to restore them to the original pixel positions, and finally, the reshape is returned to B × C × H × W .
O u t p u t = C o n v 1 × 1 C C Reshape O
In Gated Feed-Forward (GFF), for the input X B × C × H × W , the dimensionality convolution is performed first. Where the channel dimension expansion factor is r , the number of hidden channels C h = r × C .
U = C o n v 1 × 1 C 2 C h ( X ) B × 2 C h × H × W
Then, PixelShuffle upsampling is performed, a step that moves the spatial information in the channel dimensions onto H and W , with the spatial resolution scaled up by a factor of two and channel’s number scaled down by a factor of four. And the channel is split to form two branches, U 1 and U 2 . Large kernel convolution is performed for branch U 1 and dilation convolution is performed for branch U 2 .
U = P i x e d S h u f f l e ( U ) B × C h 2 × ( 2 H ) × ( 2 W )
U = [ U 1 ; U 2 ]
V 1 = D W C o n v 5 × 5 U 1 V 2 = D W C o n v 3 × 3 , d = 2 U 2
Then, a nonlinear activation of the second branch is conducted and then multiplied element by element with the first branch to complete the two-scale gated fusion, where m i s h is involved as a smooth nonmonotonic activation function.
m i s h V 2 = V 2 tanh ln 1 + e V 2 G = m i s h V 2 × V 1
By downsampling with PixelUnshuffle, the spatial information is moved back to the channel dimension, the spatial resolution is restored, and the channel’s number is scaled up by a factor of four. Finally, using convolution, we project the channel’s number back to the input size C .
G = P i x e d U n S h u f f l e G B × C h × H × W
O u t p u t = C o n v 1 × 1 C h C G B × C × H × W
The MFAM module enables the model to accurately identify lesion areas associated with pine wilt disease by modeling multi-level image features. Even if the differences in these areas are subtle, the model can efficiently capture this information through Grouped Attention and Gated Feed-Forward mechanisms. This method effectively solves the problem of early identification of subtle features of pine wilt disease, and provides strong characterization and learning capabilities for detecting subtle disease features in UAV remote sensing images.

2.3.3. Weighted and Linearly Scaled IoU Loss (WLIoU Loss)

The loss function is crucial in deep learning to measure the difference between actual labels and model predictions, and it directly affects the process of updating and optimizing network parameters [38]. In image detection tasks, selecting the appropriate loss function is critical because it will directly affect the model’s fitting effect on training data and play a key role in optimization goals [39]. Therefore, the loss function occupies a core position in neural network model training. For target detection tasks related to pine wilt disease, we chose the Intersection over Union (IoU) loss function because it can effectively handle difficult samples [40]. Building on this, we propose a mixed loss function, the Weighted and Linearly Scaled IoU Loss (WLIoU Loss), which combines IoU sample weight adjustment and linear stretch truncation. Compared to traditional loss functions, WLIoU significantly enhances model learning efficiency by modifying the weights between positive and negative samples. By optimizing the weighting parameters, this method effectively balances fitting ability and generalization, while mitigating the risk of overfitting. The specific processing flow of the WLIoU loss function is outlined below.
I o U calculates the overlap between the predicted bounding box and the ground truth box. In this case, inter represents the intersection area, while union refers to the union area.
I o U = inter union
The minimum outer box diagonal length calculation is performed first. C w and C h denote the size of the minimum outer rectangle that encloses both the real box and the predicted box. Where x p u l , y p u l and x p l r , y p l r are the upper-left and lower-right coordinates of the prediction frame, and x r u l , y r u l and x r l r , y r l r are the upper-left and lower-right coordinates of the real frame.
C w = max x p l r , x r l r min x p u l , x r u l C h = max y p l r , y r l r min y p u l , y r u l C 2 = C w 2 + C h 2
Next, the squared Euclidean distance is computed between the centers of the predicted frame and the real frame, where Δ x and Δ y are the differences between the coordinates of the two frame centers.
Δ x = x r u l + x r l r x p u l + x p l r 2 Δ y = y r u l + y r l r y p u l + y p l r 2 ρ 2 = Δ x 2 + Δ y 2
The Aspect Ratio Penalty Term is used to measure the difference in aspect ratios between two boxes. Here, the predicted box’s width and height is represented by w 1 , h 1 , while the ground truth box’s width and height is represented by w 2 , h 2 . The normalization factor 4 π 2 ensures that v 0 , 1 , and the coefficient α 1 dynamically adjusts the weight of the aspect ratio difference in the overall loss.
v = 4 π 2 arctan w 2 h 2 arctan w 1 h 1 2
α 1 = v 1 I o U + v
I o U is calculated by subtracting the center distance penalty and the aspect ratio penalty. Subsequently, the moderating factor κ is increased, decreasing its loss weight for easy-to-detect samples (high IoU) and increasing attention for difficult samples (low IoU).
I o U = I o U ρ 2 c 2 + v × α 1
w f = I o U κ
I o U = 1 I o U × w f
Finally, a linear stretching and truncation is performed to amplify the gradient signal in the high IoU interval and compress the very small value interval. Where d = 0.00 , u = 0.95 , and i w i denotes the normalization to the sum of all positive sample weights.
s u I o U = c l a m p I o U d u d
L W L I o U = 1 i w i i w i 1 s u I o U
WLIoU Loss takes into account overlap, position, and shape, as well as the focusing and weighting of difficult samples, which is more beneficial for addressing class imbalance and difficult samples, while simultaneously optimizing multiple objectives.

3. Results

3.1. Experimental Environment and Training Details

In this study, all experiments were conducted in the same hardware and software environments to avoid different experimental conditions affecting the results of SLMW-Net. AMD EPYC 9754 128-Core Processor CPU and NVIDIA GeForce RTX 4090 D GPU were the main hardware used in this experiment. Table 1 shows the experimental environment parameters. The experiment was set up with 300 training epochs, using SGD as the optimizer, an initial learning rate of 0.01, and a learning rate decay factor of 0.01. Other hyperparameters were set as follows: batch size 4, momentum value 0.937, and weight decay coefficient 0.0005. During the training phase, we employed WLIoU Loss as the loss function, updating parameters by calculating the model loss to accelerate convergence and enhance stability.

3.2. Evaluation Indicators

This study use indicators such as mAP@0.5, precision, recall, and mAP@0.5:0.95 to evaluate the effectiveness of the model. The relevant equations are given below.
As a comprehensive indicator, Mean Average Precision (mAP) is calculated by averaging individual Average Precision (AP) scores. The AP indicator measures the area under the precision-recall curve, which describes the balance between precision and recall at different thresholds, so mAP can provide a more comprehensive assessment of the performance of the detection model. Specifically, mAP@0.5 indicates the mean average precision with an IoU threshold of 0.5, whereas mAP@0.5:0.95 denotes the average mAP calculated in steps of 0.05 at IoU thresholds from 0.5 to 0.95.
m A P = N i = 1 A P i N
Precision reflects the likelihood of accurately identifying the target. TP refers to the correct detection of a diseased tree, while FP signifies the misdetection of a healthy pine tree as diseased.
P r e c i s i o n = T P T P + F P
Recall represents the likelihood of correctly identifying a diseased pine tree. FN refers to the incorrect prediction of a diseased tree as healthy.
R e c a l l = T P T P + F N

3.3. Model Performance Analysis

To comprehensively evaluate the spatiotemporal generalization capability of the proposed model in practical applications, we first conducted a 5-fold sample-based cross-validation. The entire dataset was uniformly divided into five equal subsets. In each iteration, one subset was used for validation while the remaining four were used for training. This process was repeated five times to ensure a balanced and robust assessment. For each fold, the model’s best performance was retained, and the final detection accuracy was reported as the average across all folds. As shown in Figure 3a, SLMW-Net demonstrates stable and consistent performance across different data partitions, indicating strong generalization under random sampling conditions. In addition, to further evaluate the model’s spatial transferability, we conducted an independent field-based cross-validation. Specifically, we trained the model using the ARen dataset and tested it on UAV imagery collected from Jiuyi Mountain in Yongzhou City, Hunan Province, which features distinct forest and terrain conditions. The results, presented in Figure 3b, show that SLMW-Net maintains satisfactory detection accuracy and robustness when applied to this geographically distinct region, further validating its potential for real-world deployment in diverse pine forest environments.
At the same time, to enhance the intuitiveness and interpretability of the model’s detection results, Figure 4 presents the original UAV remote sensing imagery, the normalized difference vegetation index (NDVI) image, and the target detection result image generated by the SLMW-Net model. These three types of images are precisely aligned spatially, clearly demonstrating the high spatial consistency between the suspected pine wilt disease infection areas identified by the model and the regions with low NDVI values. Specifically, the NDVI image reflects the physiological health status of vegetation within the study area, where regions with significantly low NDVI values typically indicate severe vegetation stress, degradation, or even non-vegetated areas. The bounding boxes predicted by SLMW-Net are highly overlapped with these low NDVI regions, indicating that the model exhibits strong biological relevance and spatial localization accuracy in identifying diseased areas.

3.4. Module Effectiveness Experiment

We propose three modules, namely SFEM, MFAM, and WLIoU Loss functions. To evaluate the performance of these modules, we first compare the SFEM module with other feature extraction modules. Subsequently, the MFAM module is evaluated and compared with various attention mechanisms. The WLIoU Loss function is compared with other commonly used loss functions. Under consistent experimental settings, the feasibility and performance of the proposed modules are verified by integrating each module into SLMW-Net for comparison.

3.4.1. Effectiveness of SFEM

In this research, we introduce the SFEM module designed to dynamically extract disease features from UAV remote sensing images of pine forests. By integrating learnable normalization, the SFEM module improves the representation of regions affected by nematodes, effectively overcoming the challenge of feature extraction in complex background environments. For comparison of the SFEM module’s performance, we conducted comparison experiments on Fourier Convolution (FFC) [41], Pyramid Convolution (PyConv) [42], and Multi-scale Convolution (MSCA) [43]. The comparison results are presented in Table 2. The research results show that both the SFEM and Fourier Convolution modules substantially improve detection performance for pine wilt disease. Although Fourier Convolution is known for its simplicity, flexibility, and computational efficiency, it is useful in distinguishing targets from complex backgrounds. Challenges exist that may reduce detection accuracy in practical applications. Consequently, the SFEM module is integrated into the backbone to improve both performance and efficiency in detecting pine wilt disease in visually complex scenarios.

3.4.2. Effectiveness of MFAM

In this research, we introduce a novel attentional mechanism, MFAM, aimed at improving the feature representation of pine wood nematode infestations captured in UAV images. By improving the ability to extract subtle infestation features, MFAM significantly boosts recognition accuracy. To validate its effectiveness, we compared MFAM with three existing attentional mechanisms: CBAM [44], Star Blocks [45], and SimAM [46], with the results presented in Table 3. Specifically, CBAM applies channel and spatial attention mechanisms separately to weight features. However, this approach may overlook the interdependencies between channel and spatial information, and the independent use of spatial and channel attention could result in incomplete feature representation. SimAM, while not increasing the number of network parameters, struggles to capture fine details and subtle features, particularly in the context of pine wilt disease detection. Star Blocks improves feature processing efficiency by mapping inputs to a high-dimensional nonlinear feature space via a star operation without expanding the network. Nevertheless, it does not perform as effectively as MFAM in extracting minor features of pine wilt disease. Our proposed MFAM module effectively captures fine details by integrating Grouped Attention with a Gated Feed-Forward network, thereby enhancing the ability to extract subtle features and improving its overall performance in detecting pine wood nematode infestation.

3.4.3. Effectiveness of WLIoU Loss

In this study, the proposed loss function, WLIoU Loss, integrates moderating factors with linear stretching and truncation mechanisms, enabling an equilibrium between detection performance and model complexity. The impact of various IoU loss functions on nematode disease detection is shown in Table 4. It shows that using WLIoU Loss leads to improved performance metrics for the model, emphasizing its effectiveness in localizing nematode lesions. Additionally, we assessed the performance of CIoU [47] and GIoU [48]. The results indicate that WLIoU Loss outperforms both GIoU and CIoU, particularly in handling challenging samples, as it better captures the distribution characteristics of pine wilt disease.

3.5. Ablation Experiment

In this research, we introduce SFEM, MFAM, and WLIoU Loss into the SLMW-Net model. SFEM enhances understanding of both the representation of pine trees and their spatial characteristics through convolution and learnable normalization layers. This improvement enables the model to concentrate on key features of infected pine trees in complex backgrounds, thereby increasing the accuracy of detecting diseased pine trees in diverse natural environments. MFAM emphasizes subtle features of pine trees initially infected with nematode disease, thereby enhancing the ability to perceive infected trees and extract fine details. WLIoU Loss facilitates the update of the model’s parameters through loss computation and adjusts the weights of positive and negative samples, thereby improving generalization ability and overall performance. Table 5 shows the outcomes of the ablation experiment. The data demonstrate that SLMW-Net achieves the highest performance across all key metrics, with an mAP@0.5 of 86.7%, Recall of 80.6%, Precision of 80.5%, and mAP@0.5:0.95 reaching 40.1%. These results validate the substantial advantages of SFEM, MFAM, and WLIoU Loss. Specifically, the combined application of these modules in SLMW-Net produces a cumulative effect, enabling the model to better differentiate the target from its surroundings, especially in complex image scenarios. In scenes with intricate or similar features, SFEM, MFAM, and WLIoU Loss effectively reduce misidentifications, efficiently enhance attention to subtle initial features, and ensure highly accurate target detection.

3.6. Comparison with State-of-the-Art Methods

3.6.1. SLMW-Net Performance on ARen Dataset

To thoroughly assess our model, we carry out comparison experiments with seven target detection approaches under identical testing conditions. These include Faster R-CNN [18], YOLOv3 [49], YOLOv5s [50], YOLOv7 [51], YOLOv8n [52], YOLOv9m [53], and YOLOv10n [54], allowing for a thorough analysis of the performance of SLMW-Net. Table 6 displays the results. In the two-stage detection model, candidate frames are initially generated using selective search. Features are then extracted from the input image using a pre-trained CNN, and target prediction is performed via a support vector machine. However, because each candidate frame requires separate feature extraction, this results in the generation of numerous redundant features, ultimately increasing detection time. Faster R-CNN addresses this issue by integrating candidate frame generation, classification, and regression into a single network, enabling joint training and significantly accelerating both the training and detection processes. Despite this, Faster R-CNN encounters difficulties in detecting small objects. Based on the results shown in Table 6, we observe that the one-stage detection models outperform the two-stage Faster R-CNN model in terms of overall performance. At the same time, Params and FLOPs serve as crucial metrics for a comprehensive evaluation of both the model size and the computational cost during inference. SLMW-Net shows advantages over other lightweight models in terms of the balance between computing overhead and performance. While lightweight models like YOLOv8n and YOLOv10n perform well in terms of computational efficiency, SLMW-Net makes a thorough trade-off between inspection accuracy and computing requirements. Overall, SLMW-Net optimizes both Params and FLOPs, effectively reducing computational complexity without sacrificing accuracy.
The visualization of the experimental results is shown in Figure 5. Figure 5a,b display pine trees initially infected with pine wilt disease. It can be seen that our proposed SLMW-Net achieves optimal performance, successfully attaining the highest detection accuracy without any misdetections. Figure 5c–e demonstrate the detection in a complex background. In Figure 5c, the other seven models experienced misdetections, incorrectly identifying houses as pine trees infected with pine wilt disease, while SLMW-Net did not experience misdetections and attained the highest accuracy among the correctly detected targets. In Figure 5d, all models except YOLOv9m experienced misdetections. Although YOLOv9m did not misdetect any targets, its detection accuracy was lower than that of our proposed SLMW-Net. YOLOv3, although having accuracy consistent with SLMW-Net for correctly detected targets, still had four misdetections. In Figure 5e, some features resembling pine wilt disease were mistakenly identified as infected pine trees by YOLOv3, YOLOv7, and YOLOv8n. Although Faster R-CNN, YOLOv5s, YOLOv9m, and YOLOv10 did not experience misdetections, their detection accuracies were lower than that of SLMW-Net. Figure 5f shows a more obvious pine wood nematode infestation, where the detection accuracy of the other models was not as high as that of our proposed SLMW-Net. In summary, SLMW-Net is able to more effectively extract the complex features of pine wilt disease, achieve higher precision detection, and meet the required accuracy for pine wilt disease detection.

3.6.2. SLMW-Net Performance on Roboflow Dataset

To further evaluate the performance of SLMW-Net, we obtain a public dataset for pine wilt disease detection from the Roboflow platform https://universe.roboflow.com/project-dkq3q/-9pmdt/dataset/8 (accessed on 16 June 2025). This dataset includes 4301 training images, 412 validation images, and 202 test images, each with a resolution of 640 × 640 pixels. The training images are adjusted for brightness in the range of −15% to +15%. We screen the images in this dataset and verify the performance with seven other models under the same experimental environment. Compared to the other models in Table 7, SLMW-Net performs the best in detecting pine wilt disease. We attribute the performance advantage of SLMW-Net to the following aspects: (a) SLMW-Net improves upon YOLOv11, incorporating several advanced technologies to enhance the model’s accuracy and speed while ensuring its robust detection performance; (b) SFEM, a key component of the feature extraction module, significantly enhances ability to extract complex features of pine wilt disease and improves its ability to distinguish features in non-forested areas; (c) MFAM, an attention mechanism introduced before the neck and the integration of GTM multiple times within the neck, successfully strikes a balance between global and local features, enabling the feature map to express its characteristics more clearly, especially the subtle changes in pine wilt disease; (d) WLIoU Loss optimizes the training performance by effectively adjusting the weights of different samples, balancing fitting ability and generalization capacity.

4. Discussion

The proposed SLMW-Net model in this study is primarily designed for pine wilt disease detection using UAV remote sensing imagery. To assess the model’s potential for satellite remote sensing applications, we carried out experiments utilizing Gaofen-2 satellite imagery, incorporating the blue, green, and red bands, to examine its generalization ability and real-world performance across various data sources. Specifically, we obtained 6424 images for training, 803 images for testing, and 803 images for validation, conducting a comparative analysis between the backbone and SLMW-Net models. Figure 6a shows that SLMW-Net retains a certain level of effectiveness in detecting regional PWD symptoms when applied to satellite imagery. Furthermore, Figure 6b indicates that, although the overall detection accuracy of SLMW-Net slightly declines on satellite images, primarily due to their relatively lower spatial resolution, it still demonstrates clear performance advantages over the backbone. The result not only validates the effectiveness of key architectural components within SLMW-Net on satellite remote sensing imagery but also further demonstrates the potential feasibility of this model in detecting regional pine wilt disease under satellite image conditions.

5. Conclusions

In order to explore the best method for detecting pine wilt disease, we propose an SLMW-Net accurate pine wilt disease detection method based on UAV images. First, we propose an SFEM module to enhance the feature extraction capability for detecting pine wilt disease in complex backgrounds. Then, the feature information is enhanced using MFAM to direct the model to focus on tiny features. Finally, WLIoU Loss was designed to calculate the loss of the model and guide the model parameter update. As a result, the method overcomes the problems of leakage and misdetection in pine wilt disease detection and improves the detection accuracy. In addition, SLMW-Net better matches the data features of UAV pine forest images. By enabling the timely and accurate detection of pine wilt disease, this study can assist producers in developing targeted control measures based on the identified nematode infection. In the experimental part, we divided 750 images collected in Anren, Hunan Province, into an 8:2 training set and validation set. As evidenced by the experimental results, SLMW-Net greatly improves the detection accuracy while maintaining the standard detection speed, and it outperforms other detection models. Meanwhile, we tested it on the public dataset of the Roboflow platform and found that our model still outperforms all other models, thus proving the effectiveness of our proposed model.
Despite promising results, several issues remain. Our primary dataset is geographically restricted to a single county, and while external validation with public datasets has been conducted, further validation is needed in regions with diverse climatic and ecological conditions. Additionally, early stage disease symptoms are often not visible, making it difficult to detect infections from images alone. To address these issues, future studies will concentrate on broadening the dataset to encompass multiple regions and incorporating IoT sensor technology with deep learning models to improve disease detection and prediction. The effectiveness of the model will also be assessed across different time periods and epidemic cycles to gain deeper insights into how temporal factors affect its accuracy and generalizability. Overcoming these limitations will contribute to the development of a more reliable early warning detection system, capable of effectively supporting forest management and ecological conservation in various environments.

Author Contributions

Conceptualization, G.Z.; investigation, Y.Y. and X.Y. (Xuewu Yan); methodology, X.Y. (Xiaoli Yuan) and X.Y. (Xuewu Yan); resources, Y.Y. and X.Y. (Xuewu Yan); supervision, G.Z.; validation, X.Y. (Xiaoli Yuan); writing—original draft, X.Y. (Xiaoli Yuan). All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the Forestry Department of Hunan Province Project (Grant No. XLSF2025007) and the Hunan Provincial Natural Science Foundation Project (Grant No. 2025JJ50385).

Data Availability Statement

The data related to the findings of this research are available upon request from the corresponding author.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Pan, S.; Yan, P.; Zhao, R.; Li, F.; Wang, L.; Wang, Y.; Lv, Z.; Ma, Y.; Yu, M.; Guo, X.; et al. Conduction of AgNPs with different surface charges in pine trees and their prevention and control of pine wood nematode disease. Chem. Eng. J. 2025, 520, 166146. [Google Scholar] [CrossRef]
  2. Luo, J.; Fan, J.; Huang, S.; Wu, S.; Zhang, F.; Li, X. Semi-supervised learning techniques for detection of dead pine trees with UAV imagery for pine wilt disease control. Int. J. Remote Sens. 2025, 46, 575–605. [Google Scholar] [CrossRef]
  3. Khuman, S.N.; Lee, H.Y.; Cho, I.-G.; Chung, D.; Lee, S.Y.; Lee, J.; Oh, J.-K.; Choi, S.-D. Monitoring of Organochlorine Pesticides Using Pine Needle, Pine Bark, and Soil Samples across South Korea: Source Apportionment and Implications for Atmospheric Transport. Chemosphere 2024, 370, 144043. [Google Scholar] [CrossRef] [PubMed]
  4. Lin, H.; Ning, L.; Dai, T.; Wei, W.; Ma, N.; Hong, M.; Wang, F.; You, C. Application of Bio-Based Nanopesticides for Pine Surface Retention and Penetration, Enabling Effective Control of Pine Wood Nematodes. Chem. Eng. J. 2024, 502, 158009. [Google Scholar] [CrossRef]
  5. Gao, R.; Liu, L.; Li, R.; Fan, S.; Dong, J.; Zhao, L. Predicting Potential Distributions of Monochamus Saltuarius, a Novel Insect Vector of Pine Wilt Disease in China. Front. For. Glob. Change 2023, 6, 1243996. [Google Scholar] [CrossRef]
  6. Hao, Z.; Huang, J.; Li, X.; Sun, H.; Fang, G. A Multi-Point Aggregation Trend of the Outbreak of Pine Wilt Disease in China over the Past 20 Years. For. Ecol. Manag. 2021, 505, 119890. [Google Scholar] [CrossRef]
  7. Back, M.A.; Bonifácio, L.; Inácio, M.L.; Mota, M.; Boa, E. Pine Wilt Disease: A Global Threat to Forestry. Plant Pathol. 2024, 73, 1026–1041. [Google Scholar] [CrossRef]
  8. Li, M.; Li, H.; Ding, X.; Wang, L.; Wang, X.; Chen, F. The Detection of Pine Wilt Disease: A Literature Review. Int. J. Mol. Sci. 2022, 23, 10797. [Google Scholar] [CrossRef]
  9. Wang, W.; Zhu, Q.; He, G.; Liu, X.; Peng, W.; Cai, Y. Impacts of Climate Change on Pine Wilt Disease Outbreaks and Associated Carbon Stock Losses. Agric. For. Meteorol. 2023, 334, 109426. [Google Scholar] [CrossRef]
  10. Sharma, A.; Cory, B.; McKeithen, J.; Frazier, J. Structural Diversity of the Longleaf Pine Ecosystem. For. Ecol. Manag. 2020, 462, 117987. [Google Scholar] [CrossRef]
  11. You, J.; Zhang, R.; Lee, J. A Deep Learning-Based Generalized System for Detecting Pine Wilt Disease Using RGB-Based UAV Images. Remote Sens. 2021, 14, 150. [Google Scholar] [CrossRef]
  12. Syifa, M.; Park, S.-J.; Lee, C.-W. Detection of the Pine Wilt Disease Tree Candidates for Drone Remote Sensing Using Artificial Intelligence Techniques. Engineering 2020, 6, 919–926. [Google Scholar] [CrossRef]
  13. Pan, J.; Lin, J.; Xie, T. Exploring the Potential of UAV-Based Hyperspectral Imagery on Pine Wilt Disease Detection: Influence of Spatio-Temporal Scales. Remote Sens. 2023, 15, 2281. [Google Scholar] [CrossRef]
  14. Iordache, M.-D.; Mantas, V.; Baltazar, E.; Pauly, K.; Lewyckyj, N. A Machine Learning Approach to Detecting Pine Wilt Disease Using Airborne Spectral Imagery. Remote Sens. 2020, 12, 2280. [Google Scholar] [CrossRef]
  15. Yu, R.; Luo, Y.; Zhou, Q.; Zhang, X.; Wu, D.; Ren, L. A Machine Learning Algorithm to Detect Pine Wilt Disease Using UAV-Based Hyperspectral Imagery and LiDAR Data at the Tree Level. Int. J. Appl. Earth Obs. Geoinf. 2021, 101, 102363. [Google Scholar] [CrossRef]
  16. Hu, G.; Wang, T.; Wan, M.; Bao, W.; Zeng, W. UAV Remote Sensing Monitoring of Pine Forest Diseases Based on Improved Mask R-CNN. Int. J. Remote Sens. 2022, 43, 1274–1305. [Google Scholar] [CrossRef]
  17. Li, F.; Liu, Z.; Shen, W.; Wang, Y.; Wang, Y.; Ge, C.; Sun, F.; Lan, P. A Remote Sensing and Airborne Edge-Computing Based Detection System for Pine Wilt Disease. IEEE Access 2021, 9, 66346–66360. [Google Scholar] [CrossRef]
  18. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 1137–1149. [Google Scholar] [CrossRef] [PubMed]
  19. Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.E.; Fu, C.-Y.; Berg, A.C. SSD: Single Shot MultiBox Detector. In Computer Vision—ECCV 2016; Springer: Cham, Switzerland, 2016; Volume 9905. [Google Scholar]
  20. Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You Only Look Once: Unified, Real-Time Object Detection. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 779–788. [Google Scholar]
  21. Redmon, J.; Farhadi, A. YOLO9000: Better, Faster, Stronger. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 6517–6525. [Google Scholar]
  22. Wu, K.; Zhang, J.; Yin, X.; Wen, S.; Lan, Y. An Improved YOLO Model for Detecting Trees Suffering from Pine Wilt Disease at Different Stages of Infection. Remote Sens. Lett. 2023, 14, 114–123. [Google Scholar] [CrossRef]
  23. Bochkovskiy, A.; Wang, C.-Y.; Liao, H.-Y.M. YOLOv4: Optimal speed and accuracy of object detection. arXiv 2020, arXiv:2004.10934. [Google Scholar]
  24. Sun, Z.; Ibrayim, M.; Hamdulla, A. Detection of Pine Wilt Nematode from Drone Images Using UAV. Sensors 2022, 22, 4704. [Google Scholar] [CrossRef]
  25. Wang, L.; Cai, J.; Wang, T.; Zhao, J.; Gadekallu, T.R.; Fang, K. Detection of Pine Wilt Disease Using AAV Remote Sensing with an Improved YOLO Model. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2024, 17, 19230–19242. [Google Scholar] [CrossRef]
  26. Ye, X.; Pan, J.; Shao, F.; Liu, G.; Lin, J.; Xu, D.; Liu, J. Exploring the Potential of Visual Tracking and Counting for Trees Infected with Pine Wilt Disease Based on Improved YOLOv5 and StrongSORT Algorithm. Comput. Electron. Agric. 2024, 218, 108671. [Google Scholar] [CrossRef]
  27. Zhu, X.; Wang, R.; Shi, W.; Liu, X.; Ren, Y.; Xu, S.; Wang, X. Detection of Pine-Wilt-Disease-Affected Trees Based on Improved YOLO V7. Forests 2024, 15, 691. [Google Scholar] [CrossRef]
  28. Huang, X.; Gang, W.; Li, J.; Wang, Z.; Wang, Q.; Liang, Y. Extraction of Pine Wilt Disease Based on a Two-Stage Unmanned Aerial Vehicle Deep Learning Method. J. Appl. Remote Sens. 2024, 18, 014503. [Google Scholar] [CrossRef]
  29. Yuan, J.; Wang, L.; Wang, T.; Bashir, A.K.; Al Dabel, M.M.; Wang, J.; Feng, H.; Fang, K.; Wang, W. YOLOv8-RD: High-Robust Pine Wilt Disease Detection Method Based on Residual Fuzzy YOLOv8. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2025, 18, 385–397. [Google Scholar] [CrossRef]
  30. Xu, S.; Huang, W.; Wang, D.; Zhang, B.; Sun, H.; Yan, J.; Ding, J.; Wang, J.; Yang, Q.; Huang, T.; et al. Automatic Pine Wilt Disease Detection Based on Improved YOLOv8 UAV Multispectral Imagery. Ecol. Inform. 2024, 84, 102846. [Google Scholar] [CrossRef]
  31. Yao, J.; Song, B.; Chen, X.; Zhang, M.; Dong, X.; Liu, H.; Liu, F.; Zhang, L.; Lu, Y.-B.; Xu, C.; et al. Pine-YOLO: A Method for Detecting Pine Wilt Disease in Unmanned Aerial Vehicle Remote Sensing Images. Forests 2024, 15, 737. [Google Scholar] [CrossRef]
  32. Khanam, R.; Hussain, M. YOLOv11: An Overview of the Key Architectural Enhancements. arXiv 2024, arXiv:2410.17725. [Google Scholar]
  33. Deng, Y.; Xi, H.; Zhou, G.; Chen, A.; Wang, Y.; Li, L.; Hu, Y. An Effective Image-Based Tomato Leaf Disease Segmentation Method Using MC-UNet. Plant Phenomics 2023, 5, 0049. [Google Scholar] [CrossRef] [PubMed]
  34. Wan, L.; Zhu, W.; Dai, Y.; Zhou, G.; Chen, G.; Jiang, Y.; Zhu, M.; He, M. Identification of Pepper Leaf Diseases Based on TPSAO-AMWNet. Plants 2024, 13, 1581. [Google Scholar] [CrossRef]
  35. Li, J.; Zhao, F.; Zhao, H.; Zhou, G.; Xu, J.; Gao, M.; Li, X.; Dai, W.; Zhou, H.; Hu, Y.; et al. A Multi-Modal Open Object Detection Model for Tomato Leaf Diseases with Strong Generalization Performance Using PDC-VLD. Plant Phenomics 2024, 6, 0220. [Google Scholar] [CrossRef]
  36. Hang, R.; Li, Z.; Liu, Q.; Ghamisi, P.; Bhattacharyya, S.S. Hyperspectral Image Classification with Attention-Aided CNNs. IEEE Trans. Geosci. Remote Sens. 2021, 59, 2281–2293. [Google Scholar] [CrossRef]
  37. Sadasivan, A.; Gananathan, K.; Dhanith, J.; Balasubramanian, S. A Systematic Survey of Graph Convolutional Networks for Artificial Intelligence Applications. Wiley Interdiscip. Rev. Data Min. Knowl. Discov. 2025, 15, 70012. [Google Scholar] [CrossRef]
  38. Hu, Y.; Zhang, Y.; Liu, S.; Zhou, G.; Li, M.; Hu, Y.; Li, J.; Sun, L. DMFGAN: A Multifeature Data Augmentation Method for Grape Leaf Disease Identification. Plant J. 2024, 120, 1278–1303. [Google Scholar] [CrossRef]
  39. Liu, Z.; Zhou, G.; Zhu, W.; Chai, Y.; Li, L.; Wang, Y.; Hu, Y.; Dai, W.; Liu, R.; Sun, L. Identification of Rice Disease under Complex Background Based on PSOC-DRCNet. Expert Syst. Appl. 2024, 249, 123643. [Google Scholar] [CrossRef]
  40. Putra, H.A.A.; Murni Arymurthy, A.; Chahyati, D. Enhancing Bounding Box Regression for Object Detection: Dimensional Angle Precision IoU-Loss. IEEE Access 2025, 13, 81029–81047. [Google Scholar] [CrossRef]
  41. Chu, T.; Chen, J.; Sun, J.; Lian, S.; Wang, Z.; Zuo, Z.; Zhao, L.; Xing, W.; Lu, D. Rethinking Fast Fourier Convolution in Image Inpainting. In Proceedings of the 2023 IEEE/CVF International Conference on Computer Vision (ICCV), Paris, France, 1–6 October 2023; pp. 23138–23148. [Google Scholar]
  42. Duta, I.C.; Liu, L.; Zhu, F.; Shao, L. Pyramidal Convolution: Rethinking Convolutional Neural Networks for Visual Recognition. arXiv 2020, arXiv:2006.11538. [Google Scholar]
  43. Guo, M.-H.; Lu, C.-Z.; Hou, Q.; Liu, Z.; Cheng, M.-M.; Hu, S.-M. SegNeXt: Rethinking Convolutional Attention Design for Semantic Segmentation. Adv. Neural Inf. Process. Syst. 2022, 35, 1140–1156. [Google Scholar]
  44. Woo, S.; Park, J.; Lee, J.-Y.; Kweon, I.S. CBAM: Convolutional Block Attention Module. In Computer Vision—ECCV 2018; Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y., Eds.; Springer International Publishing: Cham, Switzerland, 2018; pp. 3–19. [Google Scholar]
  45. Ma, X.; Dai, X.; Bai, Y.; Wang, Y.; Fu, Y. Rewrite the Stars. In Proceedings of the 2024 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 16–22 June 2024; pp. 5694–5703. [Google Scholar]
  46. Yang, L.; Zhang, R.-Y.; Li, L.; Xie, X. SimAM: A Simple, Parameter-Free Attention Module for Convolutional Neural Networks. In Proceedings of the 38th International Conference on Machine Learning, Virtual, 18–24 July 2021; PMLR 139. pp. 11863–11874. [Google Scholar]
  47. Zheng, Z.; Wang, P.; Ren, D.; Liu, W.; Ye, R.; Hu, Q.; Zuo, W. Enhancing Geometric Factors in Model Learning and Inference for Object Detection and Instance Segmentation. IEEE Trans. Cybern. 2022, 52, 8574–8586. [Google Scholar] [CrossRef]
  48. Rezatofighi, H.; Tsoi, N.; Gwak, J.; Sadeghian, A.; Reid, I.; Savarese, S. Generalized Intersection Over Union: A Metric and a Loss for Bounding Box Regression. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019. [Google Scholar]
  49. Redmon, J.; Farhadi, A. YOLOv3: An Incremental Improvement. arXiv 2018, arXiv:1804.02767. [Google Scholar]
  50. Khanam, R.; Hussain, M. What Is YOLOv5: A Deep Look into the Internal Features of the Popular Object Detector. arXiv 2024, arXiv:2407.20892. [Google Scholar]
  51. Wang, C.-Y.; Bochkovskiy, A.; Liao, H.-Y.M. YOLOv7: Trainable Bag-of-Freebies Sets New State-of-the-Art for Real-Time Object Detectors. In Proceedings of the 2023 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Vancouver, BC, Canada, 17–24 June 2023; pp. 7464–7475. [Google Scholar]
  52. Varghese, R.; M., S. YOLOv8: A Novel Object Detection Algorithm with Enhanced Performance and Robustness. In Proceedings of the 2024 International Conference on Advances in Data Engineering and Intelligent Computing Systems (ADICS), Chennai, India, 18–19 April 2024; pp. 1–6. [Google Scholar]
  53. Wang, C.-Y.; Yeh, I.-H.; Liao, H.-Y.M. YOLOv9: Learning What You Want to Learn Using Programmable Gradient Information. arXiv 2024, arXiv:2402.13616. [Google Scholar]
  54. Wang, A.; Chen, H.; Liu, L.; Chen, K.; Lin, Z.; Han, J.; Ding, G. YOLOv10: Real-Time End-to-End Object Detection. arXiv 2024, arXiv:2405.14458. [Google Scholar]
Figure 1. Data acquisition.
Figure 1. Data acquisition.
Plants 14 02490 g001
Figure 2. Structure of SLMW-Net. (a) The backbone, neck, and head of SLMW-Net; (b) shows the structure of SFEM; (c) shows the structure of MFAM.
Figure 2. Structure of SLMW-Net. (a) The backbone, neck, and head of SLMW-Net; (b) shows the structure of SFEM; (c) shows the structure of MFAM.
Plants 14 02490 g002
Figure 3. Cross-validation. (a) 5-fold cross-validation on ARen dataset; (b) field-based cross-validation on Jiuyi Mountain.
Figure 3. Cross-validation. (a) 5-fold cross-validation on ARen dataset; (b) field-based cross-validation on Jiuyi Mountain.
Plants 14 02490 g003
Figure 4. NDVI and detection result. (ad) Four pairs of experimental results.
Figure 4. NDVI and detection result. (ad) Four pairs of experimental results.
Plants 14 02490 g004
Figure 5. Comparison results of ARen dataset. (af) Six samples of ARen dataset.
Figure 5. Comparison results of ARen dataset. (af) Six samples of ARen dataset.
Plants 14 02490 g005
Figure 6. Performance of backbone and SLMW-Net on satellite images. (a) Visualization results; (b) Data statistics.
Figure 6. Performance of backbone and SLMW-Net on satellite images. (a) Visualization results; (b) Data statistics.
Plants 14 02490 g006
Table 1. Hardware and software environment settings.
Table 1. Hardware and software environment settings.
CategoryComponentSpecification
HardwareCPUAMD EPYC 9754 128-Core Processor
RAM1TB
GPUNVIDIA GeForce RTX 4090 D
SoftwareOSLinux
PythonPython 3.8.10
CUDA Toolkit11.8
CUDNNV8.7
PyTorch2.0.0
Table 2. The experimental findings regarding the effectiveness of SFEM.
Table 2. The experimental findings regarding the effectiveness of SFEM.
MethodsmAP@0.5 (%)Precision (%)Recall (%)mAP@0.5:0.95 (%)
FFC84.277.377.738.5
PyConv83.976.875.638.2
MSCA82.675.573.037.6
SFEM84.777.878.938.8
Table 3. The experimental findings regarding the effectiveness of MFAM.
Table 3. The experimental findings regarding the effectiveness of MFAM.
MethodsmAP@0.5 (%)Precision (%)Recall (%)mAP@0.5:0.95 (%)
CBAM83.175.974.837.2
SimAM83.677.876.237.8
Star Blocks84.276.475.338.3
MFAM84.978.976.338.9
Table 4. The experimental findings regarding the effectiveness of WLIoU Loss.
Table 4. The experimental findings regarding the effectiveness of WLIoU Loss.
MethodsmAP@0.5 (%)Precision (%)Recall (%)mAP@0.5:0.95 (%)
CIOU83.677.378.737.6
GIOU83.276.277.137.4
WLIoU Loss84.678.678.938.8
Table 5. Ablation experiment.
Table 5. Ablation experiment.
ABCmAP@0.5 (%)Precision (%)Recall (%)mAP@0.5:0.95 (%)
nonono83.976.976.338.4
yesnono84.777.878.938.8
noyesno84.978.976.338.9
nonoyes84.678.678.938.8
yesyesno85.278.778.239.6
yesnoyes85.479.279.439.5
noyesyes85.580.478.839.1
yesyesyes86.780.580.640.1
Table 6. Compare with other models on ARen dataset.
Table 6. Compare with other models on ARen dataset.
MethodsmAP@0.5
(%)
Precision
(%)
Recall
(%)
mAP@0.5:0.95 (%)Params (M)FLOPs (G)
Faster R-CNN59.558.561.323.228.3492.9
YOLOv382.275.775.739.361.5155.3
YOLOv5s83.476.477.737.37.015.9
YOLOv779.774.874.937.437.2105.1
YOLOv8n83.778.0 74.0 38.33.08.2
YOLOv9m83.479.574.639.920.277.5
YOLOv10n80.978.971.837.72.78.4
SLMW-Net86.780.580.640.13.98.8
Table 7. Compare with other models on Roboflow dataset.
Table 7. Compare with other models on Roboflow dataset.
MethodsmAP@0.5
(%)
Precision
(%)
Recall
(%)
mAP@0.5:0.95 (%)Params (M)FLOPs (G)
Faster R-CNN71.054.473.831.228.3492.9
YOLOv381.678.075.937.461.5155.3
YOLOv5s83.779.475.538.47.015.9
YOLOv784.678.979.838.037.2105.1
YOLOv8n85.180.678.239.33.08.2
YOLOv9m84.077.777.139.420.277.5
YOLOv10n82.377.475.238.22.78.4
SLMW-Net85.381.380.740.43.98.8
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Yuan, X.; Zhou, G.; Yan, Y.; Yan, X. Detection of Pine Wilt Disease in UAV Remote Sensing Images Based on SLMW-Net. Plants 2025, 14, 2490. https://doi.org/10.3390/plants14162490

AMA Style

Yuan X, Zhou G, Yan Y, Yan X. Detection of Pine Wilt Disease in UAV Remote Sensing Images Based on SLMW-Net. Plants. 2025; 14(16):2490. https://doi.org/10.3390/plants14162490

Chicago/Turabian Style

Yuan, Xiaoli, Guoxiong Zhou, Yongming Yan, and Xuewu Yan. 2025. "Detection of Pine Wilt Disease in UAV Remote Sensing Images Based on SLMW-Net" Plants 14, no. 16: 2490. https://doi.org/10.3390/plants14162490

APA Style

Yuan, X., Zhou, G., Yan, Y., & Yan, X. (2025). Detection of Pine Wilt Disease in UAV Remote Sensing Images Based on SLMW-Net. Plants, 14(16), 2490. https://doi.org/10.3390/plants14162490

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop