Next Article in Journal
Deep Learning-Based Assessment of Brainstem Volume Changes in Spinocerebellar Ataxia Type 2 (SCA2): A Study on Patients and Preclinical Subjects
Previous Article in Journal
Reconstruction Error Guided Instance Segmentation for Infrared Inspection of Power Distribution Equipment
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

MS-YOLOv11: A Wavelet-Enhanced Multi-Scale Network for Small Object Detection in Remote Sensing Images

Department of Astronautics, Space Engineering University, Beijing 101416, China
*
Author to whom correspondence should be addressed.
Sensors 2025, 25(19), 6008; https://doi.org/10.3390/s25196008
Submission received: 2 September 2025 / Revised: 21 September 2025 / Accepted: 25 September 2025 / Published: 29 September 2025
(This article belongs to the Section Remote Sensors)

Abstract

In remote sensing imagery, objects smaller than 32 × 32 pixels suffer from three persistent challenges that existing detectors inadequately resolve: (1) their weak signal is easily submerged in background clutter, causing high miss rates; (2) the scarcity of valid pixels yields few geometric or textural cues, hindering discriminative feature extraction; and (3) successive down-sampling irreversibly discards high-frequency details, while multi-scale pyramids still fail to compensate. To counteract these issues, we propose MS-YOLOv11, an enhanced YOLOv11 variant that integrates “frequency-domain detail preservation, lightweight receptive-field expansion, and adaptive cross-scale fusion.” Specifically, a 2D Haar wavelet first decomposes the image into multiple frequency sub-bands to explicitly isolate and retain high-frequency edges and textures while suppressing noise. Each sub-band is then processed independently by small-kernel depthwise convolutions that enlarge the receptive field without over-smoothing. Finally, the Mix Structure Block (MSB) employs the MSPLCK module to perform densely sampled multi-scale atrous convolutions for rich context of diminutive objects, followed by the EPA module that adaptively fuses and re-weights features via residual connections to suppress background interference. Extensive experiments on DOTA and DIOR demonstrate that MS-YOLOv11 surpasses the baseline in mAP@50, mAP@95, parameter efficiency, and inference speed, validating its targeted efficacy for small-object detection.

1. Introduction

With the rapid advancement of remote sensing technology, its applications in environmental monitoring [1], urban planning [2], and disaster assessment [3] have become increasingly widespread. In recent years, the application of deep learning techniques has significantly enhanced the feature extraction capabilities of convolutional neural networks [4]. Currently, mainstream object detection algorithms are divided into two-stage and single-stage detection methods [5]. Among them, the YOLO series [6] of single-stage detectors exhibits great potential in improving recognition accuracy and has demonstrated strong practicality and broad application potential across various computer vision tasks. Moreover, the real-time processing capability and computational efficiency of YOLO series algorithms provide significant advantages for deployment in aerospace systems [7]. With continuous iteration and optimization of the YOLO architecture, it can better adapt to general object detection tasks.
However, objects in remote sensing images are often extremely small, and their distinguishable features are limited due to high-altitude acquisition, which restricts effective feature extraction. Secondly, cluttered backgrounds combined with environmental factors such as weather variations and illumination differences exacerbate foreground–background confusion and reduce detection accuracy. Furthermore, the lack of texture and shape cues can lead to confusion among visually similar small objects (e.g., vehicles vs. oil tanks, ships vs. floating bridges). Meanwhile, weak target signals are easily overshadowed by background clutter, resulting in a significantly higher missed detection rate compared to medium and large objects [8]. Therefore, the accurate detection of small objects [9] in remote sensing images using YOLO series algorithms remains a major challenge. In summary, the core challenges of small object detection in such scenarios can be categorized into the following three aspects:
1. Weak target signals: Objects are easily obscured in wide fields of view and strong clutter, leading to a high missed detection rate.
2. Insufficient pixel information: Extreme scarcity of geometric contours and texture details makes it difficult for traditional convolution operations to capture discriminative representations.
3. Loss of high-frequency information: Continuous downsampling operations irreversibly eliminate edge and texture details, while existing pyramid structures still struggle to compensate for fine-grained structural information. To address these challenges, this paper proposes MS-YOLO (Multi-Scale YOLO), designed for remote sensing scenarios. Based on the YOLOv11 baseline, the algorithm introduces three innovative strategies: preservation of frequency domain detail, lightweight receptive field expansion, and adaptive cross-scale fusion. The specific contributions are as follows:
1.
Frequency-domain detail preservation: A 2D Haar wavelet transform is embedded at the input stage to explicitly decompose the image into a low-frequency approximation component and three high-frequency detail sub-bands. This process suppresses background noise while highlighting subtle edges and textures that are prone to being obscured.
2.
Lightweight receptive field expansion: Small-kernel depthwise separable convolutions are independently applied to each sub-band. This approach exponentially expands the receptive field without introducing redundant parameters, preserving global context while avoiding excessive smoothing of small objects.
3.
Adaptive cross-scale fusion: A Mix Structure Block (MSB) is designed, which includes the following:
  • An MSPLCK module that employs parallel dilated convolutions ( 7 × 7 13 × 13 19 × 19 ) for dense sampling, capturing multi-scale contextual information of extremely small targets;
  • An EPA module that utilizes dual channel-space attention mechanisms with residual connections to adaptively weight features, dynamically suppressing background interference and enhancing target features.

2. Related Work

2.1. Object Detection Methods

The research landscape of object detection has demonstrated a distinct evolutionary trend from traditional methods to deep learning techniques. Early conventional approaches primarily relied on combinations of handcrafted feature extractors and classifiers. Template matching methods such as the Viola–Jones detector (2001) [10] and HOG+SVM (2005) [11] achieved object localization through sliding window mechanisms combined with Haar features or histogram of oriented gradients, but suffered from high computational redundancy and poor real-time performance.
With the rise of deep learning, object detection technology has undergone fundamental transformation. The end-to-end learning paradigm based on convolutional neural networks (CNNs) has significantly improved detection accuracy and efficiency. Modern deep learning approaches primarily follow three technical pathways: two-stage detection [12], single-stage detection [13], and anchor-free detection. Two-stage methods first generate region proposals followed by classification and regression, with representative works including the R-CNN series (e.g., Fast R-CNN, Faster R-CNN) [14], and subsequent developments such as FPN (enhancing multi-scale feature fusion through feature pyramid networks) [15,16,17] and Mask R-CNN [18] (extended to instance segmentation tasks). These methods achieve high detection accuracy suitable for scenarios requiring strict precision, albeit with higher computational complexity. Single-stage methods like the YOLO series and SSD [19] directly predict object categories and locations on feature maps, trading some accuracy for speed advantages, making them more suitable for real-time applications. The YOLO series has undergone multiple iterations from YOLOv1 to the latest YOLOv12, incorporating technologies such as multi-scale prediction [20], CSPNet backbone networks [21], Transformer attention mechanisms [22], and neural architecture search (NAS) [23], continuously balancing precision and efficiency. Anchor-free methods [24] (e.g., CenterNet, CornerNet, and DETR) abandon preset anchor boxes, directly predicting key points or utilizing self-attention mechanisms for end-to-end detection, reducing hyperparameter dependency but facing challenges such as slow training convergence and high computational demands.
YOLOv11 demonstrates an optimal balance among accuracy, speed, and extensibility, making it particularly suitable for the challenging task of small object detection in remote sensing images. In contrast, although YOLOv10 [25] performs excellently in general object detection tasks—such as safety helmet monitoring on construction sites as referenced in related studies—these tasks typically involve larger objects and relatively simple backgrounds. Consequently, its architectural design and focus are not fully aligned with our requirements for multi-scale feature extraction, high-frequency detail preservation, and lightweight extensibility. Therefore, selecting YOLOv11 as the baseline and implementing targeted enhancements is both justified and effective.

2.2. Deep Learning Algorithms for Small Object Detection

The detection of small objects in remote sensing imagery represents a critical challenge in the field of computer vision. In recent years, researchers worldwide have conducted extensive studies across multiple directions, including feature enhancement and attention mechanisms, multi-scale feature fusion, model lightweighting and efficient detection, as well as Transformer-based novel approaches, significantly advancing the development of this field.
In terms of feature enhancement, researchers have improved the representational capacity for small objects by introducing attention mechanisms and feature reconstruction techniques. For instance, Huang et al., building upon Faster R-CNN, employed image upsampling and gradient ascent methods to reconstruct features while preserving detailed information. Wu et al. integrated spatial-to-depth transformations with attention mechanisms, enhancing both accuracy and the ability to distinguish densely distributed objects. However, such methods often incur substantial computational overhead, making it difficult to meet real-time detection requirements.
Multi-scale feature [26] fusion represents a crucial technical pathway for addressing scale variations. Shan et al. combined Region Proposal Networks (RPNs) with multi-angle RoI extraction [27], effectively leveraging multi-level features. Lv et al. proposed the RT-DETR model, which enhances multi-scale processing capabilities through cross-scale fusion and IoU-aware query selection mechanisms [28]. Liu et al. further designed SO-RTDETR [29], incorporating attention mechanisms to balance detection performance across objects of different scales. Although these methods achieve excellent accuracy, their complex structures and high training costs demand substantial computational resources.
Additionally, a series of improved algorithms based on the YOLO architecture have accelerated the advancement of small object detection. For example, LGFF-YOLO [30] enhances multi-scale feature representation through a Global Information Fusion Module (GIFM) and a Four-Leaf Clover Fusion Module (FLCM). FFCA-YOLO [31] introduces a Feature Enhancement Module (FEM) to expand the receptive field, while SF-YOLO [32] optimizes foreground–background discrimination via a Spatial Information Perception (SIP) module. Meanwhile, novel paradigms such as Mutual Learning (MA) [33] further improve detection performance through branch collaboration and feature alignment. Examples include MADet [34], which employs hybrid regression strategies to handle diverse objects, and DICN [35], which facilitates cross-domain knowledge transfer through network consistency [36].
In summary, CNNs—particularly YOLOv11—provide a strong baseline for small object detection in remote sensing, offering a balanced combination of speed, accuracy, and practical deployability. Their inductive bias toward local feature extraction aligns well with the core challenges of this task. While transformers [37] hold theoretical advantages in global modeling, their computational cost may be prohibitive for many real-time remote sensing applications. Therefore, MS-YOLOv11 adopts a hybrid approach: it enhances local detail extraction via wavelet transform and compensates for global context limitations through efficient attention and multi-scale mechanisms. This results in simultaneous improvements in both accuracy and speed, better meeting the practical demands of small object detection in remote sensing imagery.

3. Proposed Model

3.1. Overview of MS-YOLO Model

The original YOLOv11 architecture employs four C3K2 modules in its backbone network for feature extraction, which may induce feature degradation. To mitigate this limitation, we integrate a wavelet transform convolutional neural network (WTCNN) module into the backbone. This innovative module performs multi-scale target decomposition through wavelet transformation, enabling effective capture of target features across different scales. The proposed enhancement not only improves small object detection accuracy but also enhances multi-scale target detection performance in complex scenarios.
Within the neck network, the Mix Structure Block module plays a pivotal role in multi-scale feature fusion. By combining feature fusion techniques with residual connections, this module optimizes feature representation while capturing multi-scale characteristics. This design significantly improves feature robustness and discriminative power, enabling superior performance in detection tasks involving complex backgrounds and variable illumination conditions. The comprehensive network architecture is illustrated in Figure 1.

3.2. Algorithm Principle of Wavelet Transform

While the Fourier transform decomposes signals into global sinusoidal components, this approach presents limitations for analyzing image features that exhibit strong spatial locality, such as edges and texture patterns. Consider an image containing fine details in its upper-left quadrant versus smooth regions in the lower-right quadrant: these distinct areas demand independent spatial analysis, yet the Fourier transform inherently lacks spatial localization capabilities. This limitation becomes particularly evident when processing non-stationary image signals—for instance, abrupt frequency transitions at object boundaries—where the Fourier transform can only yield an aggregated frequency spectrum without indicating the spatial distribution of high-frequency edge components [38]. In remote sensing target detection applications, spatial localization proves especially critical. Essential image features like object boundaries manifest as localized high-frequency components. The wavelet transform addresses this fundamental limitation by directly encoding spatial position information (through image coordinates) in its high-frequency subbands (LH, HL, and HH). This stands in contrast to Fourier analysis, which merely indicates the global presence of high-frequency content without spatial reference. Moreover, wavelet decomposition naturally generates a pyramid-structured representation (exemplified by the Laplacian pyramid), facilitating hierarchical, coarse-to-fine feature extraction. This multi-resolution property makes wavelet analysis particularly suitable for computer vision tasks including target detection and image fusion. The mathematical framework of the two-dimensional Haar wavelet transform is formally presented in Algorithm 1.
Algorithm 1 2D Haar Wavelet Transform for Image Decomposition
Require: 
Input image A R C × H × W with C channels, H and W even
Ensure: 
Decomposed representation H HWT R 4 C × H / 2 × W / 2
1:
Initialize low-pass filter f L = 1 2 [ 1 , 1 ]
2:
Initialize high-pass filter f H = 1 2 [ 1 , 1 ]
3:
for  c = 1 to C do                   ▹ Process each channel
4:
    Initialize temporary tensors: s p e c L , s p e c H R H / 2 × W
5:
    for  i = 0 to H / 2 1  do
6:
        for  j = 0 to W 1  do
7:
            s p e c L [ i ] [ j ] f L [ 0 ] · A [ c ] [ 2 i ] [ j ] + f L [ 1 ] · A [ c ] [ 2 i + 1 ] [ j ]
8:
            s p e c H [ i ] [ j ] f H [ 0 ] · A [ c ] [ 2 i ] [ j ] + f H [ 1 ] · A [ c ] [ 2 i + 1 ] [ j ]
9:
        end for
10:
    end for
11:
    Initialize output subbands: L L , L H , H L , H H R H / 2 × W / 2
12:
    for  i = 0 to H / 2 1  do
13:
        for  j = 0 to W / 2 1  do
14:
            L L [ i ] [ j ] f L [ 0 ] · s p e c L [ i ] [ 2 j ] + f L [ 1 ] · s p e c L [ i ] [ 2 j + 1 ]
15:
            L H [ i ] [ j ] f H [ 0 ] · s p e c L [ i ] [ 2 j ] + f H [ 1 ] · s p e c L [ i ] [ 2 j + 1 ]
16:
            H L [ i ] [ j ] f L [ 0 ] · s p e c H [ i ] [ 2 j ] + f L [ 1 ] · s p e c H [ i ] [ 2 j + 1 ]
17:
            H H [ i ] [ j ] f H [ 0 ] · s p e c H [ i ] [ 2 j ] + f H [ 1 ] · s p e c H [ i ] [ 2 j + 1 ]
18:
        end for
19:
    end for
20:
    Stack subbands: H HWT [ 4 c : 4 c + 3 ] [ L L , L H , H L , H H ]
21:
end for
22:
return  H HWT
This work utilizes the Haar wavelet transform (WT) primarily for its computational efficiency and implementation advantages. We emphasize that our methodology is not fundamentally constrained to this specific wavelet basis, as other wavelet families could alternatively be implemented, though with potentially greater computational demands.
For a given image X , the single-level one-dimensional Haar WT (applied along either width or height dimensions) is implemented through: (1) depthwise convolution using the analysis kernels 1 2 [ 1 , 1 ] and 1 2 [ 1 , 1 ] , followed by (2) standard downsampling with stride 2. The two-dimensional extension combines these operations along both spatial dimensions, effectively realized as a depthwise convolution with stride 2 employing four distinct filter sets:
f L L = 1 2 1 1 1 1 , f L H = 1 2 1 1 1 1 , f H L = 1 2 1 1 1 1 , f H H = 1 2 1 1 1 1
Here, f L L operates as a low-pass filter, while f L H , f H L and f H H constitute complementary high-pass filters. The convolutional operation for each input channel yields the following:
[ X L L , X L H , X H L , X H H ] = Conv ( [ f L L , f L H , f H L , f H H ] , X ) ,
producing four output channels with each channel’s spatial resolution reduced by half relative to X . The components X L L , X L H , X H L , and X H H respectively represent the low-frequency approximation and horizontal, vertical, and diagonal high-frequency details. The orthogonal nature of the basis filters in Equation (1) permits exact reconstruction via transposed convolution:
X = ConvTranspose [ f L L , f L H , f H L , f H H ] , [ X L L , X L H , X H L , X H H ]
Cascaded wavelet decomposition is achieved through recursive application to the low-frequency components:
X L L ( i ) , X L H ( i ) , X H L ( i ) , X H H ( i ) = WT X L L ( i 1 )
where superscript ( i ) denotes decomposition level. This hierarchical process progressively enhances frequency resolution while reducing spatial resolution for low-frequency components.
Notably, convolutional operations in the wavelet domain enable expanded receptive fields. For instance, applying a 3 × 3 convolution to the second-level low-frequency component X L L ( 2 ) effectively establishes a nine-parameter kernel operating on a 12 × 12 receptive field in the original input X for low-frequency signal processing.

3.3. Wavelet Transform Convolutional Neural Network

The quadratic growth in parameters associated with increasing convolutional kernel sizes presents a significant challenge. Our solution involves a three-stage process: (1) wavelet-based filtering and downsampling of input low/high-frequency components, (2) small-kernel depthwise convolution on frequency-specific feature maps, and (3) output reconstruction via inverse wavelet transform. Mathematically, this operation is expressed as:
Y = IWT Conv W , WT ( X )
where X denotes the input tensor and W represents the weight tensor of the k × k depthwise convolution kernel, with input channels quadruple those of X. This approach not only isolates frequency-specific convolutions but also enables small kernels to operate on expanded input regions, effectively increasing their receptive fields. Figure 2 provides a schematic illustration.
We extend this single-level operation through cascaded decomposition, defined by the following recursive equations:
X L L ( i ) , X L H ( i ) , X H L ( i ) , X H H ( i ) = WT X L L ( i 1 )
Y L L ( i ) , Y L H ( i ) , Y H L ( i ) , Y H H ( i ) = Conv W ( i ) , X L L ( i ) , X L H ( i ) , X H L ( i ) , X H H ( i )
Here, X L L ( 0 ) serves as the input to the first decomposition layer. The output X H ( i ) at each level i comprises the three high-frequency sub-bands X L H ( i ) , X H L ( i ) , X H H ( i ) . Leveraging the linearity of the WT and IWT operations, the multi-level feature aggregation is performed recursively:
Z ( i ) = IWT Y L L ( i ) + Z ( i + 1 ) , Y L H ( i ) , Y H L ( i ) , Y H H ( i )
where Z ( i ) represents the aggregated output from level i, and the initial condition for the deepest level is set as Z ( + 1 ) = 0 . Unlike [39], which employs independent normalization for each frequency sub-band, we adopt learnable channel-wise scaling factors after the convolution operations, enabling the network to adaptively balance the contributions from different frequency components.
Figure 2 demonstrates WTCNN application to MobileNetV2’s third inverted residual block, showing two-level decomposition with 3 × 3 kernels.
The WTCNN integration offers two key advantages: (1) Exponential receptive field growth ( 2 · k ) with linear parameter increase ( · 4 · c · k 2 ) through -level decomposition, and (2) enhanced low-frequency response via recursive WT decomposition of input low frequencies, complementing standard convolution’s high-frequency bias.
These technical merits yield practical benefits including improved scalability versus large-kernel methods, enhanced corruption robustness, and stronger shape (versus texture) sensitivity.
We further visualize multi-scale feature extraction through wavelet-transformed image features (Figure 3). The decomposition clearly separates image content into distinct frequency bands, producing four characteristic feature maps.
The figure showcases avian image decomposition via 2D discrete wavelet transform. The original image (“WT”) reveals basic silhouette, while subbands highlight: (LL) morphological structures, (LH) vertical feather textures, (HL) horizontal beak/claw edges, and (HH) diagonal microstructures—effectively demonstrating hierarchical feature extraction from macro-shapes to micro-patterns.

3.4. Mix Structure Block

The mixed structure block integrates two key components: a multi-scale parallel large convolutional kernel module and an enhanced parallel attention module [40], as shown in Figure 4.

3.4.1. Multi-Scale Parallel Large Convolutional Kernel Module (MSPLCK)

The MSPLCK module synergistically combines large receptive fields with multi-scale capabilities. Large-kernel convolutional networks not only provide expanded effective receptive fields but also exhibit significantly stronger shape bias over texture bias. For small object detection tasks, extensive receptive fields are critical—enabling the network to capture the holistic structure of tiny targets through global contextual information, while precisely reconstructing object contours via shape bias.
The processing begins with the original feature map x, which undergoes batch normalization:
x ^ = BatchNorm ( x )
The module’s data flow is mathematically expressed as:
x 1 = PWConv ( x ^ ) , x 2 = Conv ( x 1 ) , x 3 = Concat ( DWDConv 19 ( x 2 ) , DWDConv 13 ( x 2 ) , DWDConv 7 ( x 2 ) )
where PWConv denotes pointwise convolution, Conv represents standard 5 × 5 convolution, and DWDConvk indicates dilated depthwise separable convolution with effective kernel size k (7 × 7 kernel with dilation rate 3 for DWDConv19, 5 × 5 kernel with dilation rate 3 for DWDConv13, and 3 × 3 kernel with dilation rate 3 for DWDConv7). The Concat operation merges features along the channel dimension.
This parallel architecture, employing three distinct kernel sizes, facilitates comprehensive multi-scale feature extraction. While larger dilated convolutions provide extensive receptive fields and pronounced shape bias—aiding in locating small targets with long-range dependencies—the smaller variants focus on fine local details and texture information, enhancing discriminative capability for small targets. The concatenation operation triples the feature dimension of x 3 relative to x.
y = x + PWConv ( GELU ( PWConv ( x ) ) )
Subsequently, x 3 is processed through a multilayer perceptron (MLP) consisting of two pointwise convolutional layers with GELU activation, transforming its feature dimension to match x. The MLP output combines with the original input via residual connection, effectively integrating diverse feature types while accommodating the specific representation requirements of small targets.

3.4.2. Enhanced Parallel Attention Module

The Enhanced Parallel Attention (EPA) module represents an innovative architecture that strategically combines complementary attention mechanisms through parallel processing. Our analysis of small target characteristics demonstrates that channel attention mechanisms excel at encoding global semantic information, while pixel attention operations show superior capability in modeling spatially fine-grained features. This architectural design enables simultaneous extraction of both location-specific and globally shared feature representations, making it particularly effective for small target detection tasks.
The EPA framework incorporates three parallel attention pathways operating on batch-normalized feature maps:
x ^ = BatchNorm ( x )
The pixel attention branch specializes in capturing spatially fine-grained features, which are critical for precise localization and classification of small targets. It is shown in Figure 5.
This component comprises two principal submodules: a feature transformation branch ( P F s ) and a spatial attention gate ( P A s ), which operate as follows:
P F s = Conv 3 × 3 ( PWConv ( x ^ ) )
P A s = σ ( PWConv ( x ^ ) )
F s = P F s P A s
where PWConv denotes point-wise convolution operations and σ represents the sigmoid activation function.The extended pixel attention mechanism incorporates additional nonlinear transformations:
P A p = σ ( PWConv ( GELU ( PWConv ( x ^ ) ) )
F p = x ^ P A p
The channel attention branch provides global contextual modeling through channel-wise feature recalibration:
C A c = σ ( PWConv ( GELU ( PWConv ( GAP ( x ^ ) ) ) )
F c = x ^ C A c
where GAP denotes global average pooling operations for spatial information aggregation.
Feature integration is achieved through concatenation and nonlinear transformation:
F = Concat ( F s , F c , F p )
y = x + PWConv ( GELU ( PWConv ( F ) ) )
The EPA module demonstrates exceptional performance for small object detection due to its dual-capacity architecture. For global semantic information, the channel attention branch generates channel-wise weights to enhance discriminative features of targets, while for spatial details, the pixel attention produces spatial masks to focus on local structures of small objects. This parallel processing of different feature hierarchies through specialized attention mechanisms enables comprehensive modeling of small object characteristics while maintaining computational efficiency.

4. Experiment

4.1. Datasets

In this paper, we utilize two open source experimental datasets: the DIOR-R dataset [41] and the DOTA dataset [42]. We used the DIOR-R dataset to assess the model’s performance. It comprises 23,190 remote sensing images with 192,512 annotated target instances, covering 20 commonly encountered object categories. All object instances are labeled with an oriented bounding box that indicates their position and orientation in the image. The distribution is 60% for training, 20% for validation, and 20% for testing. The images are resized to 640 × 640 pixels to serve as the model’s input. The DOTA (Dataset for Object Detection in Aerial Images) is a notable large-scale dataset for oriented bounding box (OBB) detection in remote sensing. It encompasses 2806 images, varying in size from 800 × 800 to 4000 × 4000 pixels, with 188,282 instances across 15 categories. This dataset facilitates a comprehensive evaluation of object detection algorithms on diverse aerial images and is employed to verify the framework’s accuracy in oriented object detection tasks.The oriented bounding box (OBB) head predicts the rotation angle θ for each object, parameterized within the range [ 0 ,   90 ° ) and represented directly by its value for regression.

4.2. Evaluation Metrics

We evaluate model performance using five metrics: Precision (P), Recall (R), mean Average Precision (mAP), F1-score, detection speed (FPS), and mAP for small objects (denoted as mAP(small)). Based on false positives (FP), true positives (TP), false negatives (FN), and true negatives (TN), P and R are defined as follows:
R = T P T P + F N
P = T P T P + F P
IoU threshold is set to 0.5. Subsequently, using the previously obtained recall and precision rates, we plot the Precision–Recall (P-R) curve and compute the Average Precision (AP) as the area under this curve. We evaluate the model using several standard mean Average Precision (mAP) metrics following the COCO evaluation protocol:
  • mAP@0.50: mAP at a single IoU threshold of 0.50;
  • mAP@0.75: mAP at a stricter IoU threshold of 0.75;
  • mAP@0.50:0.95: Average mAP computed over multiple IoU thresholds from 0.50 to 0.95 in steps of 0.05;
  • mAP(small): Specifically measures mAP@0.50 for small objects, defined as those with area less than 32 × 32 pixels according to the COCO standard.
The final mAP value used for overall model evaluation is defined as the mean value of Average Precision (AP) across all target categories:
A P = 0 1 P ( R ) d R , m AP = i = 1 N cls A P N cls
where N cls denotes the number of categories. Furthermore, a higher F1-score indicates stronger robustness of the detection model, defined as follows:
F 1 = 2 × P × R P + R
Additionally, we report inference speed in FPS (Frames Per Second). All metrics are computed on the test set unless otherwise specified.

4.3. Experiment Setup

As presented in Table 1, the hardware environment consists of an NVIDIA A100—SXM4—80 GB GPU (NVIDIA, Santa Clara, CA, USA). The software environment runs on Ultralytics 8.3.54 Python—3.8.20 torch—1.9.0 + cu111. We employ the SGD algorithm for end-to-end network optimization, with hyperparameters configured as follows: initial learning rate ( l r 0 ) of 0.01, final learning rate ( l r f ) of 0.01, batch size of 16, image size ( i m g s z ) of 640, 300 training epochs, 8 workers.

4.4. Experiment Results

Table 2, Table 3 and Table 4 present the performance metrics of the YOLOv11-OBB and MS-YOLOv11-OBB models on the DIOR-R and DOTA datasets, respectively. By analyzing the models’ performance across different categories in the DIOR-R dataset, we observe that the improved model achieves significant results in detecting small targets. Specifically, for small target categories such as Storage Tank, and Vehicle, the model’s mAP50 values reach 85.87%, and 81.13%, respectively. These results demonstrate that the model maintains high detection accuracy even when dealing with small and challenging targets. Furthermore, the average mAP50 across all categories reaches 0.8833, further validating the superior overall performance of the model. Additionally, the mAP50 for all categories exceeds 69.61%.
To further compare with the original YOLOv11 and verify the robustness of MS-YOLOv11, we conducted comparative experiments on the DOTA and DIOR-R datasets. The results show that our method achieves better performance, with the improved MS-YOLOv11-OBB model showing overall improvements in Recall, F1-Score, and mAP on the DIOR-R dataset. Compared to the original YOLOv11, MS-YOLOv11 improves the average P, R, F1, and mAP50 by 0.22%, 3.57%, 2.32%, and 2.58%, respectively. Notably, for the Dam category, the original YOLOv11 achieves an mAP of only 56.61%, while MS-YOLOv11 reaches 72.85%, marking a significant improvement of 16.24%. For small target categories such as Storage Tank, Ship, and Vehicle, the model also achieves high precision. Finally, on the DOTA dataset, the proposed MS-YOLOv11 improves P, R, F1, and mAP by 2.75%, 1.98%, 2.19%, and 3.13%, respectively, compared to the original YOLOv11. From these comparisons, we can preliminarily conclude that MS-YOLOv11 performs better in detecting various remote sensing targets and outperforms the original YOLOv11 in terms of precision and accuracy.
Figure 6 presents the Precision–Recall (P-R) curves for two models, visually reflecting the trade-off between precision and recall at different confidence thresholds. The curve trends indicate that the P-R curve of MS-YOLO V11obb is smoother, closer to the top-right corner of the graph, and the precision decreases more slowly as recall increases. This suggests that MS-YOLO V11bb maintains better performance across different confidence thresholds, demonstrating higher stability and accuracy.
Lastly, to visually demonstrate the detection performance of MS-YOLOv11, we present some detection results of MS-YOLOv11 and YOLOv11 on the DIOR-R dataset, as shown in Figure 7. YOLOv11 exhibits varying degrees of missed and false detections, while MS-YOLOv11 addresses these issues. Particularly when significant overlap exists between multiple target objects, MS-YOLOv11’s bounding box predictions are more accurate. This further indicates that our improvements enhance the original YOLO performance, contributing to higher accuracy in remote sensing target extraction.

4.5. Ablation Experiments

In this work, the Haar wavelet transform is employed for its efficiency and simplicity, although our approach is agnostic to the choice of basis, with others available at the cost of increased computational demand. To further validate the choice of the Haar wavelet and explore the impact of different frequency-domain transformations, we conducted comparative experiments with various wavelet bases under identical experimental settings. As shown in Table 5, while more complex wavelets like Sym8 and Bior2.4 achieved competitive accuracy, possibly due to their higher regularity and better texture representation, the difference in mAP was marginal (<0.5%). Crucially, the Haar wavelet demonstrated a significant advantage in terms of inference speed (FPS), outperforming other bases by a considerable margin. This aligns with our design principle of achieving a optimal balance between accuracy and efficiency for practical applications. Therefore, we conclude that the Haar wavelet is the most suitable choice for the WTConv module in MS-YOLOv11.
According to the ablation study results shown in Table 6, we analyzed the contributions of different components in the MS-YOLOv11 model. By locally enhancing the Backbone with the WTConv module, we improved the mAP from 85.75% to 86.96%, while reducing the number of parameters by approximately 61,591, as the WTConv module replaced the repeatedly used C3K2 modules, achieving a 1.21% increase in accuracy. When independently replacing the neck network with the proposed Mix Structure Block, the mAP increased from 85.75% to 87.81% (an improvement of 2.06%). Overall, while keeping the number of parameters essentially unchanged, MS-YOLOv11 increased the FPS from 302.17 to 447.81, achieving a 2.58% higher mAP compared to the original YOLOv11, and also improved the mAP@75 and mAP@50-95 by 4.33% and 4.24%, respectively. These results demonstrate the effective enhancement of MS-YOLOv11 over the original YOLOv11 performance.

4.6. Comparison with Other Detection Models

To validate the superiority of the proposed MS-YOLOv11, we conducted comprehensive comparisons with several representative detection methods on the DIOR-R dataset, including the baseline YOLOv11, YOLOv8, classical oriented object detectors (Rotated Faster R-CNN, Rotated RetinaNet, RoI Transformer, Gliding Vertex, AOPG, GGHL, Oriented RepPoints), and state-of-the-art approaches (LSK-S*, DCFL). As summarized in Table 7, classical methods generally underperformed in remote sensing scenarios, with mAP values ranging from 62.0% to 66.3%. Notably, MS-YOLOv11 achieved the highest mAP of 88.33%, outperforming YOLOv11 (85.75%) and YOLOv8 (84.96%), and significantly surpassing all other methods. For small object categories—Ship (SH), Storage Tank (STO), and Vehicle (VE)—MS-YOLOv11 attained detection accuracies of 95.03%, 85.87%, and 81.13%, respectively, demonstrating consistent advantages over competing approaches. These results confirm that MS-YOLOv11 not only enhances overall detection performance but also exhibits exceptional capability in small object detection, validating the effectiveness of our architectural improvements.
Benefiting from the WTConv module utilization of 2D Haar Wavelet Transform for multi-level decomposition of input images, it expands the receptive field while extracting both global and detailed features. Secondly, the network further introduces the Mix Structure Block module to enhance small target and multi-scale feature extraction capabilities, acquiring multi-scale features that are ultimately optimized through feature fusion and residual connections. MS-YOLOv11 demonstrates higher adaptability and superior detection performance in remote sensing target detection, effectively addressing issues of missed and false detections. Meanwhile, in terms of speed, MS-YOLOv11 achieves an 145.64 FPS improvement compared to YOLOv11. Furthermore, future research could integrate multi-source remote sensing data and employ techniques like domain adaptive learning to further enhance the accuracy of algorithm and generalization capabilities.

5. Conclusions

The proposed MS-YOLOv11 algorithm significantly improves small target detection in remote sensing images by integrating the WTConv module with 2D Haar Wavelet Transform and the Mix Structure Block module. The WTConv module expands the receptive field through multi-level decomposition while jointly extracting global and local features, whereas the Mix Structure Block enhances multi-scale feature fusion. Experimental results demonstrate that MS-YOLOv11 achieved higher detection accuracy on both the DIOR-R and DOTA datasets, with mAP improvements of 2.58% and 3.13%, respectively. The detection performance demonstrates significant improvement in Storage Tank small target category, increasing from 78.50 % to 81.13 % , a rise of 2.63 % . Compared to classical algorithms (e.g., Rotated Faster R-CNN) and the original YOLOv11, the model markedly reduces missed and false detections while delivering more precise bounding box predictions. Future work will focus on incorporating multi-source remote sensing data and domain adaptation learning to further enhance the algorithm’s generalization in complex scenarios.

Author Contributions

Conceptualization, X.L. and H.L.; Methodology, H.L.; Software, H.L.; Validation, H.L.; Formal Analysis, Y.Z.; Investigation, H.L.; Resources, X.L.; Data Curation, Z.W. and Q.L.; Writing—Original Draft Preparation, H.L.; Writing—Review & Editing, L.W.; Visualization, H.L.; Supervision, L.W.; Project Administration, L.W.; Funding Acquisition, X.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data is contained within the article.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Li, J.; Pei, Y.; Zhao, S.; Xiao, R.; Sang, X.; Zhang, C. A review of remote sensing for environmental monitoring in China. Remote Sens. 2020, 12, 1130. [Google Scholar] [CrossRef]
  2. Wellmann, T.; Lausch, A.; Andersson, E.; Knapp, S.; Cortinovis, C.; Jache, J.; Scheuer, S.; Kremer, P.; Mascarenhas, A.; Kraemer, R.; et al. Remote sensing in urban planning: Contributions towards ecologically sound policies? Landsc. Urban Plan. 2020, 204, 103921. [Google Scholar] [CrossRef]
  3. Yamazaki, F.; Matsuoka, M. Remote sensing technologies in post-disaster damage assessment. J. Earthq. Tsunami 2007, 1, 193–210. [Google Scholar] [CrossRef]
  4. Li, Z.; Liu, F.; Yang, W.; Peng, S.; Zhou, J. A survey of convolutional neural networks: Analysis, applications, and prospects. IEEE Trans. Neural Netw. Learn. Syst. 2021, 33, 6999–7019. [Google Scholar] [CrossRef] [PubMed]
  5. Padilla, R.; Netto, S.L.; Da Silva, E.A.B. A survey on performance metrics for object-detection algorithms. In Proceedings of the 2020 International Conference on Systems, Signals and Image Processing (IWSSIP), Niteroi, Brazil, 1–3 July 2020; pp. 237–242. [Google Scholar]
  6. Wang, L.; Bai, J.; Li, W.; Jiang, J. Research Progress of YOLO Series Target Detection Algorithms. J. Comput. Eng. Appl. 2023, 59, 15. [Google Scholar]
  7. Liu, Z.; Gao, Y.; Du, Q. Yolo-class: Detection and classification of aircraft targets in satellite remote sensing images based on yolo-extract. IEEE Access 2023, 11, 109179–109188. [Google Scholar] [CrossRef]
  8. Betti, A.; Tucci, M. YOLO-S: A lightweight and accurate YOLO-like network for small target detection in aerial imagery. Sensors 2023, 23, 1865. [Google Scholar] [CrossRef]
  9. Qu, J.; Tang, Z.; Zhang, L.; Zhang, Y.; Zhang, Z. Remote sensing small object detection network based on attention mechanism and multi-scale feature fusion. Remote Sens. 2023, 15, 2728. [Google Scholar]
  10. Wang, Y.-Q. An analysis of the Viola-Jones face detection algorithm. Image Process. Line 2014, 4, 128–148. [Google Scholar] [CrossRef]
  11. Dadi, H.S.; Pillutla, G.K.M. Improved face recognition rate using HOG features and SVM classifier. IOSR J. Electron. Commun. Eng. 2016, 11, 34–44. [Google Scholar] [CrossRef]
  12. Zhou, X.; Koltun, V.; Krähenbühl, P. Probabilistic two-stage detection. arXiv 2021, arXiv:2103.07461. [Google Scholar] [CrossRef]
  13. Li, B.; Liu, Y.; Wang, X. Gradient harmonized single-stage detector. In Proceedings of the AAAI Conference on Artificial Intelligence, Honolulu, HI, USA, 27 January–1 February 2019; Volume 33, pp. 8577–8584. [Google Scholar]
  14. Liu, B.; Zhao, W.; Sun, Q. Study of object detection based on Faster R-CNN. In Proceedings of the 2017 Chinese Automation Congress (CAC), Jinan, China, 20–22 October 2017; pp. 6233–6236. [Google Scholar]
  15. Zhang, G.; Yu, W.; Hou, R. Mfil-fcos: A multi-scale fusion and interactive learning method for 2d object detection and remote sensing image detection. Remote Sens. 2024, 16, 936. [Google Scholar] [CrossRef]
  16. Yuan, Z.; Gong, J.; Guo, B.; Wang, C.; Liao, N.; Song, J.; Wu, Q. Small object detection in uav remote sensing images based on intra-group multi-scale fusion attention and adaptive weighted feature fusion mechanism. Remote Sens. 2024, 16, 4265. [Google Scholar]
  17. Li, J.; Huang, K. Lightweight multi-scale feature fusion network for salient object detection in optical remote sensing images. Electronics 2024, 14, 8. [Google Scholar] [CrossRef]
  18. He, K.; Gkioxari, G.; Doll’ar, P.; Girshick, R. Mask r-cnn. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2961–2969. [Google Scholar]
  19. Yilmaz, E.N.; Navruz, T.S. Real-Time Object Detection: A Comparative Analysis of YOLO, SSD, and EfficientDet Algorithms. In Proceedings of the 2025 7th International Congress on Human-Computer Interaction, Optimization and Robotic Applications (ICHORA), Ankara, Turkiye, 23–24 May 2025; pp. 1–9. [Google Scholar]
  20. Sun, G.; Zhang, Y.; Sun, W.; Liu, Z.; Wang, C. Multi-scale prediction of the effective chloride diffusion coefficient of concrete. Constr. Build. Mater. 2011, 25, 3820–3831. [Google Scholar] [CrossRef]
  21. Wang, C.-Y.; Liao, H.-Y.M.; Wu, Y.-H.; Chen, P.-Y.; Hsieh, J.-W.; Yeh, I.-H. CSPNet: A new backbone that can enhance learning capability of CNN. In Proceedings of the Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, Seattle, WA, USA, 14–19 June 2020; pp. 390–391. [Google Scholar]
  22. Choi, S.R.; Lee, M. Transformer architecture and attention mechanisms in genome data analysis: A comprehensive review. Biology 2023, 12, 1033. [Google Scholar] [CrossRef]
  23. Elsken, T.; Metzen, J.H.; Hutter, F. Neural architecture search: A survey. J. Mach. Learn. Res. 2019, 20, 1–21. [Google Scholar]
  24. Zhang, S.; Chi, C.; Yao, Y.; Lei, Z.; Li, S.Z. Bridging the gap between anchor-based and anchor-free detection via adaptive training sample selection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 9759–9768. [Google Scholar]
  25. Wang, S.; Park, S.; Kim, J.; Kim, J. Safety Helmet Monitoring on Construction Sites Using YOLOv10 and Advanced Transformer Architectures with Surveillance and Body-Worn Cameras. J. Constr. Eng. Manag. 2025, 151, 04025186. [Google Scholar] [CrossRef]
  26. Cheng, Y.; Wang, W.; Zhang, W.; Yang, L.; Wang, J.; Ni, H.; Guan, T.; He, J.; Gu, Y.; Tran, N.N. A multi-feature fusion and attention network for multi-scale object detection in remote sensing images. Remote Sens. 2023, 15, 2096. [Google Scholar]
  27. Chen, J.; Kao, S.-h.; He, H.; Zhuo, W.; Wen, S.; Lee, C.-H.; Chan, S.-H.G. Run, don’t walk: Chasing higher FLOPS for faster neural networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023; pp. 12021–12031. [Google Scholar]
  28. Zhao, Y.; Lv, W.; Xu, S.; Wei, J.; Wang, G.; Dang, Q.; Liu, Y.; Chen, J. Detrs beat yolos on real-time object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 16–22 June 2024; pp. 16965–16974. [Google Scholar]
  29. Liu, J.; Cao, Y.; Wang, Y.; Guo, C.; Zhang, H.; Dong, C. SO-RTDETR for Small Object Detection in Aerial Images. Preprints 2024, 2024101980. [Google Scholar] [CrossRef]
  30. Peng, H.; Xie, H.; Liu, H.; Guan, X. LGFF-YOLO: Small object detection method of UAV images based on efficient local–global feature fusion. J. Real-Time Image Process. 2024, 21, 167. [Google Scholar] [CrossRef]
  31. Wei, J.; Liang, J.; Song, J.; Zhou, P. YOLO-PBESW: A Lightweight Deep Learning Model for the Efficient Identification of Indomethacin Crystal Morphologies in Microfluidic Droplets. Micromachines 2024, 15, 1136. [Google Scholar] [CrossRef] [PubMed]
  32. Cao, X.; Su, Y.; Geng, X.; Wang, Y. YOLO-SF: YOLO for fire segmentation detection. IEEE Access 2023, 11, 111079–111092. [Google Scholar] [CrossRef]
  33. Zhang, Y.; Xiang, T.; Hospedales, T.M.; Lu, H. Deep mutual learning. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 4320–4328. [Google Scholar]
  34. Xie, X.; Lang, C.; Miao, S.; Cheng, G.; Li, K.; Han, J. Mutual-assistance learning for object detection. IEEE Trans. Pattern Anal. Mach. Intell. 2023, 45, 15171–15184. [Google Scholar] [CrossRef]
  35. Jiao, Y.; Yao, H.; Xu, C. Dual instance-consistent network for cross-domain object detection. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 45, 7338–7352. [Google Scholar] [CrossRef] [PubMed]
  36. Yao, B.; Zhang, C.; Meng, Q.; Sun, X.; Hu, X.; Wang, L.; Li, X. SRM-YOLO for Small Object Detection in Remote Sensing Images. Remote Sens. 2025, 17, 2099. [Google Scholar]
  37. Park, S.; Kim, J.; Kim, J.; Wang, S. Fault diagnosis of air handling units in an auditorium using real operational labeled data across different operation modes. J. Comput. Civ. Eng. 2025, 39, 04025065. [Google Scholar] [CrossRef]
  38. Finder, S.E.; Amoyal, R.; Treister, E.; Freifeld, O. Wavelet convolutions for large receptive fields. In Proceedings of the European Conference on Computer Vision, Milan, Italy, 29 September 2024; pp. 363–380. [Google Scholar]
  39. Ding, X.; Zhang, X.; Han, J.; Ding, G. Scaling up your kernels to 31 ✕ 31: Revisiting large kernel design in cnns. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 11963–11975. [Google Scholar]
  40. Lu, L.; Xiong, Q.; Xu, B.; Chu, D. Mixdehazenet: Mix structure block for image dehazing network. In Proceedings of the 2024 International Joint Conference on Neural Networks (IJCNN), Yokohama, Japan, 30 June–5 July 2024; pp. 1–10. [Google Scholar]
  41. Li, K.; Wan, G.; Cheng, G.; Meng, L.; Han, J. Object detection in optical remote sensing images: A survey and a new benchmark. ISPRS J. Photogramm. Remote Sens. 2020, 159, 296–307. [Google Scholar]
  42. Xia, G.-S.; Bai, X.; Ding, J.; Zhu, Z.; Belongie, S.; Luo, J.; Datcu, M.; Pelillo, M.; Zhang, L. DOTA: A large-scale dataset for object detection in aerial images. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 3974–3983. [Google Scholar]
  43. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards real-time object detection with region proposal networks. IEEE Trans. Pattern Anal. Mach. Intell. 2016, 39, 1137–1149. [Google Scholar]
  44. Hou, L.; Lu, K.; Xue, J.; Li, Y. Shape-adaptive selection and measurement for oriented object detection. In Proceedings of the AAAI Conference on Artificial Intelligence, Oxford, UK, 19–21 May 2021; Volume 36, pp. 923–932. [Google Scholar]
  45. Ding, J.; Xue, N.; Long, Y.; Xia, G.-S.; Lu, Q. Learning RoI transformer for oriented object detection in aerial images. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 2849–2858. [Google Scholar]
  46. Xu, Y.; Fu, M.; Wang, Q.; Wang, Y.; Chen, K.; Xia, G.-S.; Bai, X. Gliding vertex on the horizontal bounding box for multi-oriented object detection. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 43, 1452–1459. [Google Scholar] [CrossRef]
  47. Yang, X.; Yang, X.; Yang, J.; Ming, Q.; Wang, W.; Tian, Q.; Yan, J. Learning high-precision bounding box for rotated object detection via kullback-leibler divergence. In Proceedings of the Advances in Neural Information Processing Systems, Virtual, 6–14 December 2021; Volume 34, pp. 18381–18394. [Google Scholar]
  48. Li, Z.; Hou, B.; Wu, Z.; Ren, B.; Yang, C. FCOSR: A simple anchor-free rotated detector for aerial object detection. Remote Sens. 2023, 15, 5499. [Google Scholar] [CrossRef]
  49. Li, W.; Chen, Y.; Hu, K.; Zhu, J. Oriented reppoints for aerial object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 1829–1838. [Google Scholar]
  50. Yang, W.; Li, Q.; Li, Q.; Huang, H. Remote Sensing Image Object Detection Method with Multi-Scale Feature Enhancement Module. In Proceedings of the 2025 5th International Conference on Neural Networks, Information and Communication Engineering (NNICE), Guangzhou, China, 10–12 January 2025; pp. 457–460. [Google Scholar]
  51. Xu, C.; Ding, J.; Wang, J.; Yang, W.; Yu, H.; Yu, L.; Xia, G.-S. Dynamic coarse-to-fine learning for oriented tiny object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023; pp. 7318–7328. [Google Scholar]
  52. Yu, J.; Huang, W.; Shen, W. An Improved YOLOv8 Algorithm for Oriented Object Detection in Remote Sensing Images. In Proceedings of the 2024 6th International Conference on Robotics and Computer Vision (ICRCV), Wuxi, China, 20–22 September 2024; pp. 88–92. [Google Scholar]
Figure 1. Architectural overview of MS-YOLOv11 network.
Figure 1. Architectural overview of MS-YOLOv11 network.
Sensors 25 06008 g001
Figure 2. Operational principle of wavelet transform convolutional neural network.
Figure 2. Operational principle of wavelet transform convolutional neural network.
Sensors 25 06008 g002
Figure 3. Wavelet transform-based multi-scale feature extraction demonstration.
Figure 3. Wavelet transform-based multi-scale feature extraction demonstration.
Sensors 25 06008 g003
Figure 4. Architecture of the Mix Structure Block, comprising both the multi-scale parallel large convolution kernel module and enhanced parallel attention module.
Figure 4. Architecture of the Mix Structure Block, comprising both the multi-scale parallel large convolution kernel module and enhanced parallel attention module.
Sensors 25 06008 g004
Figure 5. Architectural diagram of the EPA module.
Figure 5. Architectural diagram of the EPA module.
Sensors 25 06008 g005
Figure 6. Precision–Recall curve of YOLOV11obb and MS-YOLOV11obb on DIOR-R.
Figure 6. Precision–Recall curve of YOLOV11obb and MS-YOLOV11obb on DIOR-R.
Sensors 25 06008 g006
Figure 7. Comparison example of YOLOv11obb and MS-YOLOv11obb test results on DIOR-R.
Figure 7. Comparison example of YOLOv11obb and MS-YOLOv11obb test results on DIOR-R.
Sensors 25 06008 g007
Table 1. Experimental hardware and software environment configuration.
Table 1. Experimental hardware and software environment configuration.
CategoryComponentSpecification/Version
Hardware EnvironmentCPUIntel Xeon Platinum 8369B @ 2.90 GHz (Intel, Santa Clara, CA, USA)
RAM512 GB DDR4
GPUNVIDIA A100-SXM4-80 GB (x1)
GPU
Memory
80 GB
Software EnvironmentOSUbuntu 20.04.6 LTS
CUDA11.1
cuDNN8.0.5
Python3.8.20
PyTorch1.9.0+cu111
FrameworkUltralytics YOLOv11 (v8.3.54)
Table 2. The DOTA dataset results of MS-YOLOV11 OBB(D01: plane, D02: ship, D03: storage tank, D04: baseball diamond, D05: tennis court, D06: basketball court, D07: ground track field, D08: harbor, D09: bridge, D10: large vehicle, D11: small vehicle, D12: helicopter, D13: roundabout, D14: soccer ball field, D15: swimming pool).
Table 2. The DOTA dataset results of MS-YOLOV11 OBB(D01: plane, D02: ship, D03: storage tank, D04: baseball diamond, D05: tennis court, D06: basketball court, D07: ground track field, D08: harbor, D09: bridge, D10: large vehicle, D11: small vehicle, D12: helicopter, D13: roundabout, D14: soccer ball field, D15: swimming pool).
ModelClassDetection Categories
D01D02D03D04D05D06D07D08D09D10D11D12D13D14D15AVG
YOLOv11obbP0.93470.90310.93370.81540.97410.66990.74650.84490.66050.86440.66020.50240.75610.68800.71120.7738
R0.88710.85720.59700.71220.90010.53290.56650.83590.44260.83250.71710.56720.50500.49290.71170.6772
F10.91030.87950.72830.76030.92310.59360.64410.84030.52680.84810.68200.53280.61770.55610.71140.7170
AP0.93080.90160.76740.77640.94620.58710.63140.53330.48000.88240.72210.57490.59130.50210.68150.7219
MS-YOLOv11obbP0.94010.77200.94210.83790.99800.66100.77440.87190.68750.90000.74170.52180.84630.69090.89330.8013
R0.88630.75310.87590.74020.86510.47230.57360.81070.43110.84560.77490.60090.88750.54320.89170.6970
F10.91240.76240.72400.78600.90010.55090.65900.84020.52390.87190.75800.59450.69350.60820.89250.7389
AP0.93390.78850.76920.80650.92540.55890.66350.86860.48560.90090.80190.64240.67180.55180.94100.7532
Note: P = Precision, R = Recall, AP = Average Precision per class.
Table 3. The Dior-R dataset results of MS-YOLOV11 OBB(D01: Airplane, D02: Airport, D03: Baseball Diamond, D04: Basketball Court, D05: Bridge, D06: Chimney, D07: Dam, D08: Expressway Service Area, D09: Expressway Toll Station, D10: Harbor).
Table 3. The Dior-R dataset results of MS-YOLOV11 OBB(D01: Airplane, D02: Airport, D03: Baseball Diamond, D04: Basketball Court, D05: Bridge, D06: Chimney, D07: Dam, D08: Expressway Service Area, D09: Expressway Toll Station, D10: Harbor).
ModelClassDetection Categories (D01–D10)
D01D02D03D04D05D06D07D08D09D10
YOLOv11obbP0.98620.80460.97640.93940.84890.97670.63260.93260.97080.9041
R0.91620.77860.95330.79020.55520.94730.55910.91670.80310.8431
F10.94990.79140.96470.85840.67130.96180.59360.92460.87910.8726
AP0.93550.84290.98290.85670.66570.96820.56610.94490.87110.9103
MS-YOLOv11obbP0.98740.84520.98350.93170.84580.97290.72810.94990.89000.8588
R0.92950.83580.93830.83330.64340.96490.68420.93560.80360.8387
F10.95750.84050.96040.87980.73090.96890.70550.94270.84460.8487
AP0.96000.89930.98330.86370.71130.97670.72850.98030.88140.9008
Table 4. The Dior-R dataset results of MS-YOLOV11 OBB(D11: Golf Field, D12: Ground Track Field, D13: Overpass, D14: Ship, D15: Stadium, D16: Storage Tank, D17: Tennis Court, D18: Train Station, D19: Vehicle, D20: Windmill).
Table 4. The Dior-R dataset results of MS-YOLOV11 OBB(D11: Golf Field, D12: Ground Track Field, D13: Overpass, D14: Ship, D15: Stadium, D16: Storage Tank, D17: Tennis Court, D18: Train Station, D19: Vehicle, D20: Windmill).
ModelClassDetection Categories (D11–D20)
D11D12D13D14D15D16D17D18D19D20AVG
YOLOv11obbP0.93160.76970.91760.93910.96360.94970.98280.73520.95350.93150.9023
R0.88190.55790.65230.90620.94740.71260.96270.79210.57860.85630.7955
F10.90600.64690.76260.92240.95540.81420.97270.76260.72020.89230.8411
AP0.95070.65070.77500.96290.97400.85270.98750.77570.78500.89100.8575
MS-YOLOv11obbP0.92930.80270.85140.94090.99110.93390.97560.83920.91730.91580.9045
R0.91130.60650.69100.90980.95240.77660.96340.85340.66760.88390.8312
F10.92020.69100.76290.92510.97130.84800.96950.84630.77280.89960.8643
AP0.95280.69610.78260.95030.98330.85870.99040.87020.81130.88580.8833
Note: P = Precision, R = Recall, F1 = F1-Score, AP = Average Precision per class, AVG = Average.
Table 5. Performance comparison of MS-YOLOv11 with different wavelet bases on the DIOR-R dataset.
Table 5. Performance comparison of MS-YOLOv11 with different wavelet bases on the DIOR-R dataset.
Wavelet BasemAP@50 (%)mAP@50:95 (%)mAP(small) (%)FPSParams (M)
Haar (Ours)88.3370.9084.21447.82.87
Db488.1570.7583.95421.52.87
Db888.4170.8884.10415.22.87
Sym488.2870.8083.87423.12.87
Sym888.5071.0284.35412.72.87
Bior1.387.9570.4583.70440.12.87
Bior2.488.4570.9584.28418.32.87
Table 6. Ablation study results.
Table 6. Ablation study results.
Model VariantParametersmAP@50 (%)mAP@75 (%)mAP@50-95 (%)FPS
YOLOv11-OBB2,802,64785.7571.8666.66302.17
YOLOv11-OBB + WTConv2,741,59186.9673.6268.88420.35
YOLOv11-OBB + Mix3,254,02587.8175.7970.50376.43
MS-YOLOv112,865,38388.3376.1970.90447.81
Note: mAP@50 = mean Average Precision at IoU 50%, FPS = Frames Per Second.
Table 7. Comparison with other detection models.
Table 7. Comparison with other detection models.
MethodBackboneSH | STO | VEmAP
Faster R-CNN-O [43]R-5079.4 | 67.6 | 46.262.0
SASM [44]R-5083.6 | 63.9 | 43.662.2
RoI Transformer [45]R-5081.2 | 62.5 | 43.864.8
Gliding Vertex [46]R-5081.0 | 62.5 | 43.262.9
KLD [47]R-5080.9 | 68.0 | 47.864.6
FCOSR [48]R-5081.0 | 62.5 | 43.163.4
Oriented RepPoints [49]R-5085.1| 65.3 | 48.066.3
LSKNet-S * [50]LSK-S81.2 | 70.9 | 49.871.6
† DCFL [51]R-50- | - | 50.971.0
YOLO V8obb * [52]CSP- | - | -86.3
YOLO V8obbCSP94.91 | 81.95 | 79.7884.96
YOLO V11obbCSP96.29 | 85.27 | 78.5085.75
MS-YOLOv11obbCSP95.03 | 85.87 | 81.1388.33
Note: SH = Ship, STO = Storage Tank, VE = Vehicle, mAP = mean Average Precision. * denotes improved versions of the baseline models. † indicates the model presented in the cited reference.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Liu, H.; Li, X.; Wang, L.; Zhang, Y.; Wang, Z.; Lu, Q. MS-YOLOv11: A Wavelet-Enhanced Multi-Scale Network for Small Object Detection in Remote Sensing Images. Sensors 2025, 25, 6008. https://doi.org/10.3390/s25196008

AMA Style

Liu H, Li X, Wang L, Zhang Y, Wang Z, Lu Q. MS-YOLOv11: A Wavelet-Enhanced Multi-Scale Network for Small Object Detection in Remote Sensing Images. Sensors. 2025; 25(19):6008. https://doi.org/10.3390/s25196008

Chicago/Turabian Style

Liu, Haitao, Xiuqian Li, Lifen Wang, Yunxiang Zhang, Zitao Wang, and Qiuyi Lu. 2025. "MS-YOLOv11: A Wavelet-Enhanced Multi-Scale Network for Small Object Detection in Remote Sensing Images" Sensors 25, no. 19: 6008. https://doi.org/10.3390/s25196008

APA Style

Liu, H., Li, X., Wang, L., Zhang, Y., Wang, Z., & Lu, Q. (2025). MS-YOLOv11: A Wavelet-Enhanced Multi-Scale Network for Small Object Detection in Remote Sensing Images. Sensors, 25(19), 6008. https://doi.org/10.3390/s25196008

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop