Next Article in Journal
A Semi-Automated, Hybrid GIS-AI Approach to Seabed Boulder Detection Using High Resolution Multibeam Echosounder
Previous Article in Journal
Spatio-Temporal Characteristics of the Morphological Development of Gully Erosion on the Chinese Loess Plateau
Previous Article in Special Issue
Remote Sensing Image Compression via Wavelet-Guided Local Structure Decoupling and Channel–Spatial State Modeling
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

HAF-YOLO: Dynamic Feature Aggregation Network for Object Detection in Remote-Sensing Images

College of Weaponry Engineering, Naval University of Engineering, Wuhan 430030, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2025, 17(15), 2708; https://doi.org/10.3390/rs17152708
Submission received: 23 June 2025 / Revised: 27 July 2025 / Accepted: 29 July 2025 / Published: 5 August 2025

Abstract

The growing use of remote-sensing technologies has placed greater demands on object-detection algorithms, which still face challenges. This study proposes a hierarchical adaptive feature aggregation network (HAF-YOLO) to improve detection precision in remote-sensing images. It addresses issues such as small object size, complex backgrounds, scale variation, and dense object distributions by incorporating three core modules: dynamic-cooperative multimodal fusion architecture (DyCoMF-Arch), multiscale wavelet-enhanced aggregation network (MWA-Net), and spatial-deformable dynamic enhancement module (SDDE-Module). DyCoMF-Arch builds a hierarchical feature pyramid using multistage spatial compression and expansion, with dynamic weight allocation to extract salient features. MWA-Net applies wavelet-transform-based convolution to decompose features, preserving high-frequency detail and enhancing representation of small-scale objects. SDDE-Module integrates spatial coordinate encoding and multidirectional convolution to reduce localization interference and overcome fixed sampling limitations for geometric deformations. Experiments on the NWPU VHR-10 and DIOR datasets show that HAF-YOLO achieved mAP50 scores of 85.0% and 78.1%, improving on YOLOv8 by 4.8% and 3.1%, respectively. HAF-YOLO also maintained a low computational cost of 11.8 GFLOPs, outperforming other YOLO models. Ablation studies validated the effectiveness of each module and their combined optimization. This study presents a novel approach for remote-sensing object detection, with theoretical and practical value.

1. Introduction

In recent years, the rapid development of remote-sensing technologies, coupled with a significant reduction in manufacturing costs, has driven their widespread adoption in various fields, such as agricultural monitoring [1], geological exploration [2], infrastructure inspection [3], and smart-city management [4]. Remote-sensing satellites provide wide-area geographical information from a high-altitude perspective, offering efficient data support for real-time monitoring [5], object tracking [6], and data analysis. However, images captured by unmanned aerial vehicles (UAVs) are often highly complex. This complexity arises from several factors, including intricate background interference, severe illumination variations, extremely small size of some objects, and wide variability in object scales. These factors make small-object detection one of the biggest challenges in the field of remote sensing [7]. Specifically, small objects in remote-sensing images, such as vehicles, small buildings, or vegetation, typically occupy only a tiny fraction of the image (e.g., fewer than 32 × 32 pixels). Their blurred feature representations and susceptibility to background interference further exacerbate the difficulty of accurate detection [8].
Traditional object-detection methods rely on handcrafted feature extraction algorithms, such as histogram of oriented gradients (HOG) [9] and scale-invariant feature transform (SIFT) [10]. Despite moderate success in earlier stages, these methods adapt poorly to complex scenes, particularly when addressing issues such as low object resolution and scale variation [11]. With breakthroughs in deep learning, convolutional neural networks (CNNs) [12] have significantly improved the robustness of object detection by automatically learning hierarchical feature representations. CNNs extract semantic information through convolution and down-sampling operations. However, multiple down-sampling stages often cause a loss of spatial details, thereby limiting detection precision [13].
To solve this issue, researchers have proposed multiscale feature fusion approaches, such as feature pyramid networks (FPNs) [13] and bidirectional feature pyramid networks (Bi-FPNs) [14], which enhance feature representation for small objects by integrating high-resolution shallow features with deep semantic information. Among single-stage detection algorithms, the You Only Look Once (YOLO) series has stood out because of its efficient end-to-end framework, which achieves a balance between accuracy and speed by integrating deeper backbone networks with attention mechanisms. However, object detection in remote-sensing images continues to pose unique challenges: Certain objects in satellite images are extremely small and densely distributed, leading to high false-negative rates; at the same time, the wide range of object scales imposes stringent requirements on the network’s multiscale detection capabilities.
To solve these issues, we propose a hierarchical adaptive fusion network (HAF-YOLO). The network consists of three main modules: the dynamic-cooperative multimodal fusion architecture (DyCoMF-Arch), multiscale wavelet-enhanced aggregation network (MWA-Net), and spatial-deformable dynamic enhancement module (SDDE-Module). The DyCoMF-Arch module constructs a hierarchical feature pyramid through multistage spatial compression and expansion operations and uses a dynamic weight allocation mechanism to accurately extract salient features from objects at multiple scales. The MWA-Net employs wavelet-transform-based convolution to decompose features and establishes both a high-frequency detail preservation mechanism and a dynamic weighted fusion strategy. The SDDE-Module integrates spatial coordinate encoding with multidirectional convolutional feature extraction to improve the model’s adaptability to the spatial distribution of small objects.
These innovations effectively address challenges related to scale variation, complex backgrounds, and the difficulty of small-object detection in remote-sensing data, while enhancing the network’s multiscale feature representation and small-object localization accuracy. Compared with existing methods, including traditional multiscale fusion networks such as FPN and Bi-FPN, our approach introduces more dynamic and adaptive mechanisms tailored for remote sensing complexities. FPN and Bi-FPN architectures are effective in fusing features across different scales, but they often adopt fixed fusion strategies, which may not fully adapt to the spatial and contextual variations of small objects in high-resolution remote sensing imagery. Similarly, YOLOv8 improves detection performance through architectural refinements and decoupled heads, yet it still suffers from limited adaptability in handling dense and geometrically variant small objects due to its fixed convolutional design and standard upsampling modules. In contrast, our proposed HAF-YOLO architecture introduces adaptive fusion strategies, including dynamic weight allocation, wavelet-enhanced multiscale decomposition, and spatial-deformable attention mechanisms. These enable our model to dynamically recalibrate feature importance across layers and regions, improving robustness against background clutter and geometric deformations.
The contributions of this study are summarized as follows:
  • We propose DyCoMF-Arch, which constructs a multigranularity feature space through multilevel progressive sampling operations. By integrating BiFPN-Concat for cross-level feature interaction, the architecture enhances multiscale object-detection capabilities under complex remote-sensing scenes.
  • We design MWA-Net as a replacement for the original C2f module. This network constructs multipath branches to fully extract image features and employs a dynamic fusion branch based on the BiFPN structure to achieve dynamic weighted feature fusion. This approach enhances the representation of small objects in complex backgrounds and solves the issue of feature detail loss encountered by the C2f module in remote-sensing scenes.
  • To address the limitations of the YOLOv8 detection head in detecting small objects in remote-sensing images, we propose SDDE-Module. This module introduces a coordinate enhancement layer to embed absolute coordinate information and adds a multibranch convolutional architecture comprising horizontal, vertical, and deformable convolution branches. This design improves the localization of dense small objects under interference conditions and overcomes the limitations of fixed sampling patterns in adapting to geometric deformations.

2. Related Work

2.1. Traditional Object Detection Methods

The development of traditional object-detection methods dates to the early 21st century. The early approaches mainly relied on manual processing combined with classifiers. The Viola–Jones framework [15], which utilized Haar features and an AdaBoost classifier for face detection, was limited to specific object types and struggled in complex scenes. Subsequently, methods based on sliding windows and feature extraction became mainstream. The HOG proposed by Dalal et al. [9], when combined with a support vector machine (SVM) classifier, enhanced robustness in pedestrian detection. However, its limited feature representation capabilities hindered its ability to handle multiscale objects and complex backgrounds.
In 2014, Girshick et al. introduced region-based convolutional neural networks (R-CNN) [16], marking the beginning of the deep learning era in object detection. Its core pipeline included selective search for generating region proposals, independent CNN-based feature extraction, and classification with an SVM coupled with bounding box regression. Although it significantly improved accuracy on the Pascal VOC dataset, it suffered from computational redundancy and complex training. To address these issues, Girshick proposed Fast R-CNN [17], which introduced the region of interest (ROI) pooling layer. With Fast R-CNN, the entire image could be input into a CNN to generate shared feature maps and extract features of region proposals. In addition, the multitask loss function was employed to further improve training efficiency. In 2015, Ren et al. advanced the design with Faster R-CNN [18], replacing selective search with a region proposal network (RPN). Utilizing an anchor-box mechanism to directly generate region proposals, it enabled joint optimization of proposal generation and object detection. This end-to-end architecture set the standard for two-stage detection algorithms, although its computational complexity still limited real-time application. To improve detection precision and address sample imbalance, Cai et al. proposed Cascade R-CNN [19] in 2018. This method employed a multistage cascade of detectors to progressively increase the intersection-over-union (IoU) thresholds for candidate boxes and ground-truth objects. The iterative optimization reduced false-positive rates and achieved outstanding performance in complex backgrounds and densely packed small-object scenes. It proved particularly effective in detecting small objects such as vehicles and ships in remote-sensing images [8]. However, the inherent dependency on region proposal generation in two-stage methods limited their real-time applicability, prompting researchers to shift toward single-stage detection algorithms.
In 2016, Liu et al. proposed the single-shot multibox detector (SSD) [20], a representative single-stage detection algorithm. By predefining multiscale anchor boxes on feature maps at various levels, it directly predicted object classes and bounding box offsets. Using VGG16 as the backbone network and incorporating a multiscale feature fusion strategy, SSD achieved a balance between detection speed and recall rate for small objects. Nevertheless, its performance in detecting objects at extreme scales was constrained by the depth of the feature pyramid. To improve feature transmission and gradient flow, Huang et al. introduced DenseNet [21], where each layer was directly connected to all subsequent layers. This dense connectivity enhanced feature reuse and alleviated the vanishing gradients in deep networks, providing detailed features for small-object detection in high-resolution remote-sensing images. However, its high memory consumption limited its application in resource-constrained environments.

2.2. YOLO Series Object Detection Methods

The YOLO series, as a canonical representative of single-stage object-detection algorithms, has consistently maintained a technological edge in real-time detection through its efficient end-to-end architecture. In 2016, Redmon et al. introduced YOLOv1 [22], which reformulated object detection as a regression task. By dividing the input image into an S × S grid and directly predicting bounding box coordinates, confidence scores, and class probabilities, it achieved a breakthrough in detection speed. However, this version exhibited poor performance in detecting dense small objects and suffered from limited localization precision due to coarse grid partitioning. To overcome these limitations, YOLOv2 (YOLO9000) [23], released in 2017, introduced the anchor box mechanism. It utilized predefined multiscale priors to enhance multiscale object detection, adopted batch normalization and a multiscale training strategy, and achieved a favorable balance between accuracy and speed on the Pascal VOC and COCO datasets. In 2018, YOLOv3 [24] improved upon its predecessors by employing the enhanced Darknet-53 backbone with residual connections to mitigate vanishing gradients. It also integrated an FPN to facilitate cross-level feature fusion, increasing the average precision (AP) for small-object detection by approximately 15%.
Released in 2020, YOLOv4 [25] adopted the CSPDarknet53 backbone to reduce computational redundancy, introduced the Mish activation function to enhance nonlinear expressiveness, and utilized the CIoU loss function to refine bounding box regression accuracy. Around the same time, YOLOv5 [26] emerged with adaptive anchor-box computation and hybrid data augmentation strategies and demonstrated greater deployment flexibility in industrial applications. Its modular design supported multiple model sizes suitable for various hardware configurations. In 2021, YOLOX innovatively removed the anchor-box design and employed a decoupled detection head along with the SimOTA dynamic label assignment strategy, improving the detection precision of small objects on the COCO dataset. In 2022, YOLOv6 [27] implemented a RepVGG-style re-parameterization design and channel distillation technique, which reduced computational costs by 30% while maintaining accuracy. YOLOv7 [28] integrated dynamic label assignment with a cascaded feature scaling strategy. Its extended version, YOLOv7-X, achieved 46.8% mAP on the VisDrone2019 remote-sensing dataset, representing a 7.3% improvement over YOLOv5. In 2023, YOLOv8 [29] introduced the task-aligned assigner (TAA) and dynamic non-maximum suppression (NMS), excelling in dense object scenes. YOLOv9 [30] improved feature propagation paths using programmable gradient information (PGI) and enhanced multiscale feature interaction with the general efficient layer aggregation network (GELAN), effectively mitigating semantic degradation in deep networks. YOLOv10 [31], released in 2024, adopted a training paradigm without NMS. By applying a consistency loss function to suppress redundant prediction boxes directly, it significantly reduced the false-positive rate in remote sensing tasks.
In this study, YOLOv8 was selected as the baseline detection model for three primary reasons. First, it achieves an optimal balance between accuracy and efficiency on remote-sensing datasets. As shown in our baseline experiments (Section 4.4.1), YOLOv8 achieved a strong mAP50 performance of 75.0% on the DIOR dataset while maintaining a relatively low computational cost of 8.2 GFLOPs. Second, YOLOv8 provides a modular and extensible architecture that facilitates seamless integration of the three proposed modules: DyCoMF-Arch, MWA-Net, and SDDE-Module. This structural flexibility enables targeted enhancements in feature fusion, scale sensitivity, and spatial adaptability. Given these advantages, YOLOv8 serves as a practical and technically sound foundation for further enhancement in the remote-sensing object-detection domain.

2.3. Application of YOLO Series Algorithms in Remote-Sensing Image Detection

In remote-sensing image object detection, detecting small and multiscale objects presents multiple technical challenges, which mainly stem from limitations associated with object properties, imaging conditions, and algorithmic design. Small objects typically occupy only a minuscule proportion of image pixels. Their sparse feature information is easily overwhelmed by complex background noise [32]. In remote-sensing images, vehicles or small buildings often appear as blurry patches due to resolution constraints, lacking prominent texture or shape features [33]. Traditional CNNs extract semantic information through layer-by-layer down-sampling, but this process leads to the loss of small-object details. This issue is particularly prominent in deep networks, where low-level features of objects are gradually diluted, making it difficult for detectors to distinguish objects from the background [34].
Complex background interference is another key challenge for small-object detection. Natural scene elements, such as clouds, vegetation, and shadows, share similar color and texture characteristics with small objects, leading to rising false-positive rates [35]. Koyun et al. [36] demonstrated that in UAV images over farmland or forested areas, local textures of crops or trees may be misidentified as small man-made objects. Wu et al. [37] proposed the AAPW-YOLO network, which enhances the YOLOv8 backbone network by replacing standard convolution with AKConv to improve small-object feature extraction. In addition, they employed the Wise-IoU gradient enhancement strategy to balance the gradient contributions of anchor boxes of varying quality, thereby improving regression precision and reducing false-positive and false-negative rates caused by small object size. Ji et al. [38] proposed the YOLO-TLA network, which integrates a small-object detection layer into the neck network of YOLOv5 to improve small-object detection precision. Yin et al. [39] redefined the channel partitioning strategy in CSPNet and integrated it into the neck network of YOLOv4. By combining a bidirectional multiscale feature weighting mechanism, they significantly improved small-object detection performance. However, the method underperforms in rotated object detection.
The considerable scale variation among multiscale objects further complicates detection. Remote-sensing images cover extensive areas. The same scene may simultaneously contain buildings spanning several meters and sub-meter vehicles. As a result, object sizes may cover multiple orders of magnitude. Although FPNs enhance scale adaptability by fusing features across layers, their top-down feature propagation mechanism may cause the loss of small-object information from shallow features during fusion. Liu et al. [40] determined that when fusing high-resolution shallow features, FPNs are susceptible to strong responses from large objects, which weakens the feature representation of small objects. Furthermore, the dense distribution of objects in remote-sensing images exacerbates the complexity of multiscale detection. Occlusion and overlap among adjacent objects interfere with the detector’s localization and classification accuracy. To resolve the issue, Zhao et al. [41] proposed MS-YOLOv7, which introduces a fourth detection head based on YOLOv7 to extract features at various scales and enhance multiscale object-detection capability. However, this approach significantly increases the number of model parameters. Li et al. [42] proposed YOLO-DRS, which integrates a lightweight multiscale module (LEC) to fuse multiscale feature information and improve the model’s ability to extract and recognize multiscale objects.
These studies have contributed various improvements to YOLO-based object detection schemes. However, challenges remain in the algorithm improvement process, including insufficient preservation and utilization of spatial information, rigid feature fusion strategies, limited feature detection capabilities, and increased model complexity. To overcome these limitations, we propose HAF-YOLO in this study.

3. Methodology

To address the challenges identified in remote-sensing satellite image detection, we propose HAF-YOLO, based on the YOLOv8 architecture. This framework establishes a collaborative mechanism for multiscale feature capture, cross-level information fusion, and adaptive feature enhancement to improve multiscale feature representation and small-object detection capabilities. The overall architecture of the proposed method is illustrated in Figure 1. First, we designed DyCoMF-Arch to enhance responses to salient objects. This module performs multilevel spatial compression to capture cross-scale remote-sensing features and integrates shallow textures and deep semantics using BiFPN-Concat with weighted fusion. Then, addressing the difficulty of small-object detection in remote-sensing scenes using the C2f module in YOLOv8, we replaced the C2f module with the MWA-Net, which constructs multiple feature extraction paths and dynamically fuses them to strengthen the representation of small objects in complex backgrounds. Finally, at the detection head, we introduced the SDDE-Module, which is positioned before the detection head. This module integrates multiple perception mechanisms to enhance the detector’s ability to identify objects in remote-sensing images.

3.1. DyCoMF-Arch

To address the challenges of significant object scale variation and high background complexity in remote-sensing data, we designed the DyCoMF-Arch. This architecture establishes a synergistic mechanism for multiscale feature capture, cross-level information fusion, and adaptive feature enhancement, thereby improving the network’s multiscale feature representation capabilities. The core design philosophy behind this architecture is to construct a hierarchical feature pyramid using multistage spatial compression and expansion operations, combined with a dynamic weight allocation mechanism for the precise extraction of salient features in remote-sensing images.
As illustrated in Figure 1, during the feature extraction phase, the network first constructs a multigranularity feature space through three stages of progressive sampling. Average pooling operations with strides of 8, 4, and 2 are applied to down-sample image layers of varying resolutions from the backbone network. Large-stride pooling captures broad contextual information, medium-stride pooling extracts regional structural features, and small-stride pooling preserves fine-grained local details. Concurrently, features from the spatial pyramid pooling fast (SPPF) layer are upsampled by a factor of two to align high-level semantic information with spatial details, forming a multiresolution feature map covering multiple scales. Cross-level feature interaction is achieved using BiFPN-Concat, which performs weighted concatenation of five heterogeneous feature pathways and dynamically adjusts the contribution weights of each scale. Compared to conventional FPNs that propagate information unidirectionally, this design enables shallow spatial details to retroactively refine deep semantic features while simultaneously guiding high-level semantics to enhance low-level representations. After feature interaction, a 1 × 1 convolution is applied to compress and rearrange information along the channel dimension, eliminating redundancies introduced by multiscale concatenation. Subsequently, we introduced the HiGA-Module of our method to further enhance the discriminative power of features. Figure 2 illustrates the network architecture of the HiGA-Module.
The core innovation of the HiGA-Module lies in establishing a synergistic optimization mechanism across both the channel and spatial dimensions, enabling balanced extraction of deep semantic and fine-grained spatial features in remote-sensing images. The module adopts a dual-path parallel architecture. In the main path, a channel compression module reduces dimensionality. This operation can be mathematically defined as
C = F c o m p r e s s ( C 1 ) = σ ( C o n v 1 × 1 C 1 ) ,
where C 1 H × W × c 1 is the input feature map; σ is the Hard–Swish activation function. This design compresses the channel dimension to one-quarter of the original input, reducing the computational complexity of subsequent operations. In contrast to traditional squeeze-and-excitation (SE) modules that reweight channels, this method utilizes linear dimensionality reduction in combination with a non-saturating activation function, thus preserving key feature information while mitigating gradient vanishing issues. It is suitable for handling high-dimensional feature spaces in high-resolution remote-sensing images. For spatial feature extraction, the module employs a decoupled depth-wise separable convolution architecture as follows:
S = Conv 1 × 1 ( DWConv 3 × 3 C ) .
This module first uses grouped depth-wise convolution to extract spatial features, followed by 1 × 1 convolution to expand channels. The design enables localized modeling of spatial correlations by depth-wise convolution and global integration of channel information by pointwise convolution. It maintains a large receptive field while enhancing the network’s ability to represent geometric objects in remote-sensing images, such as building edges and road networks that have strong directional characteristics.
To address the prevalent issues of significant intra-class variance and strong background interference in remote-sensing object detection, we further propose chunked gated adaptive attention (CGAA) based on the multihead self-attention (MHSA) framework [43]. This module employs a chunk-wise serial processing strategy to decompose the input feature x B × N × C ( N = H × W ) into m sub-blocks for parallel computation. Each sub-block corresponds to localized-global fusion across different spatial scales. The attention weight matrix is computed as
Attention ( Q , K , V ) = Softmax Q K T d k V ,
where Q , K , V m × n × d ( n = N / m ) are the query, key, and value matrices after chunking, respectively. This chunk-wise computation establishes a hierarchical attention field across scales, allowing small-scale objects to obtain stronger feature activation within localized chunks, whereas large-scale objects capture long-range dependencies through inter-chunk interactions. The final output is combined with the input by residual connections, ensuring training stability. Compared with existing feature fusion structures in the YOLO series, such as the PANet used in YOLOv7, DyCoMF-Arch introduces a more novel approach by constructing a multilevel spatial compression and expansion mechanism to dynamically establish a multiscale contextual representation. Furthermore, it innovatively incorporates BiFPN-Concat into remote-sensing object-detection tasks to achieve bidirectional feature stream interaction. This approach addresses the limitation of traditional single-channel or single-dimensional spatial attention mechanisms, allowing the network to retain critical edge information in densely distributed object regions while effectively suppressing background interference.

3.2. MWA-Net

The conventional C2f module employs a serially stacked bottleneck architecture for feature transformation. However, its single-path design tends to cause the progressive attenuation of shallow detail features during the propagation toward deeper layers. In remote-sensing images, small objects occupy a limited number of pixels and feature high texture complexity. As a result, conventional convolution with down-sampling operations exacerbates the loss of high-frequency information, which is manifested as blurred object boundaries and localization inaccuracies. We propose the MWA-Net to address these limitations of the C2f module in YOLOv8, namely, the loss of fine-grained details, inadequate multiscale modeling, and weakened semantic information for small objects. MWA-Net was specifically designed to enhance the representation of small objects under complex backgrounds by introducing mechanisms for high-frequency detail preservation, multigranularity feature interaction paths, and dynamic weighted fusion strategies. The overall architecture is illustrated in Figure 3.
The frequency-domain decomposition branch replaces conventional convolutional layers with wavelet-transform-based convolution (WTConv2d) [44], decomposing the input feature maps into low-frequency approximation and high-frequency detail components. This process can be formulated as a convolutional implementation of the two-dimensional discrete wavelet transform:
W T ( X ) L L , W T ( X ) L H , W T ( X ) H L , W T ( X ) H H = W ( X ψ d b 1 ) ,
where ψ d b 1 is the Daubechies wavelet basis function. The LL component retains the macroscopic structural features of the object, and the LH, HL, and HH components capture vertical, horizontal, and diagonal high-frequency details, respectively. This explicit frequency-domain separation strategy allows the network to preserve edge oscillation characteristics of small objects at shallow layers, effectively mitigating the high-frequency information loss inherent to traditional convolution operations.
The feature refinement path employs a “compression–expansion” architecture in conjunction with depth-wise separable convolutions to strike a balance between computational efficiency and representational capacity. First, a 1 × 1 convolution expands the channel dimension to p × 2 c , forming a high-dimensional feature space to enhance nonlinear representational power. Next, a depth-wise separable convolution (DWConv) is applied for spatial feature extraction, which reduces computational complexity O ( d k 2 C i n H W + C i n C o u t H W ) by a factor of approximately 1 / C o u t compared to standard convolution, significantly reducing the processing overhead for high-resolution remote-sensing images. Finally, another 1 × 1 convolution compresses the channels back to the original dimensionality, completing the feature distillation process. This “expansion–compression” mechanism strengthens local feature interaction through a temporary high-dimensional mapping space, while DWConv retains a 3 × 3 receptive field and captures edge oscillation patterns of small objects using per-channel convolution. The process is formalized as
F r e f i n e ( X ) = C o n v 1 × 1 ( D W C o n v ( C o n v 1 × 1 ( X ) ) ) .
The nonlinear expansion path consists of multiple bottleneck modules, each constructed with dual 3 × 3 convolutional residual structures. By cascading multiple bottleneck blocks, the network constructs a deep nonlinear mapping space, thereby enhancing its discriminative ability in scenes with both complex backgrounds and co-existing small objects. The operation is defined by
F b o t t l e n e c k ( X ) = X + C o n v 3 × 3 ( C o n v 3 × 3 ( X ) ) .
Features from all three paths are fused using BiFPN-Concat. During feature concatenation, the outputs from the WTConv decomposition (y01, y02), the feature refinement path (y1), the original feature partitions (y2, y3), and the n feature tensors from the bottleneck expansion are combined into a total of 5 + n feature tensors. Unlike equal-weighted concatenation (Concat), this mechanism utilizes gradient competition during backpropagation to autonomously strengthen the fusion weights of high-frequency detail features while suppressing interference from redundant background features. In the frequency-domain decomposition path, the high-frequency components generated by the wavelet transform carry edge oscillation information of small objects. Meanwhile, the original feature partition path retains more spatial contextual information. Using a cross-domain correlation model, the dynamic weighting strategy adaptively adjusts the contribution ratios of features across different frequency bands. This design overcomes the limitations of fixed-weight fusion where high-frequency features are prone to smoothing effects in deep networks, thereby providing more stable object-detection performance under complex backgrounds.
In contrast to methods such as AAPW-YOLO, which replaces standard convolution with AKConv to improve edge extraction, MWA-Net utilizes WTConv2d to perform explicit decomposition of feature maps into high- and low-frequency components, thus preserving the texture boundaries of small objects. Meanwhile, its dynamic weight allocation strategy mitigates the over-processing of high-frequency information typically encountered in Bi-FPN-based multifrequency fusion. The traditional C2f module suffers from limited capacity in multipath modeling and fine-grained semantic retention; MWA-Net addresses this issue with a three-path structure that facilitates comprehensive feature extraction across different representation levels.

3.3. SDDE-Module

We also propose the SDDE-Module to address the limitations of the YOLOv8 detection head in small-object detection within remote-sensing images, such as its insufficient spatial sensitivity and limited adaptability to deformation features. This module enhances the model’s adaptability to the spatial distribution of small objects by integrating spatial coordinate encoding, multidirectional convolutional feature extraction, and an adaptive deformation-aware mechanism. The architecture is illustrated in Figure 4. The detection head in YOLOv8 lacks explicit encoding of absolute coordinate information in remote sensing scenes, which makes the localization of dense small objects susceptible to interference. Furthermore, its fixed sampling patterns are inadequate for adapting to the geometric deformations inherent in object imaging. This module systematically addresses these limitations through a hierarchical feature enhancement strategy.
In remote-sensing images, the spatial distribution characteristics of objects are critical for detection tasks. Traditional convolution operations lack explicit positional encoding. As a result, the spatial information of small objects is diluted in deep networks. To address this issue, the SDDE-Module introduces a coordinate enhancement layer. The structure of this layer is shown in Figure 5. This layer embeds absolute coordinate information into the feature map using a normalized grid generation algorithm (Figure 5). Specifically, given an input feature map X B × C × H × W , a normalized coordinate grid is first generated as follows:
x norm = 2 x W 1 1 , x [ 0 , W 1 ] y norm = 2 y H 1 1 , y [ 0 , H 1 ] .
Then, the polar coordinate r = x norm 2 + y norm 2 and azimuthal angle θ = arctan ( y norm / x norm ) are further expanded and concatenated with the original input to form the enhanced feature X enhanced = Concat ( X , x norm , y norm , r , θ ) . This design is well-suited for detecting densely arranged objects in remote-sensing images, such as windmills and ships. Given the highly regular spatial distribution of these objects, coordinate priors can be utilized to enhance localization accuracy. Therefore, the design is particularly effective for dense small objects in remote-sensing images.
To address the diversity of object shapes and interference from complex backgrounds, the SDDE-Module adopts a multibranch convolutional architecture. This targeted design enables the extraction of feature information from various directions and dimensions, thereby improving the model’s object recognition capabilities. The two branches in the architecture use (1 × 3) and (3 × 1) convolution kernels to extract directional features along the horizontal and vertical axes, enhancing the discriminability of spatial features. Their mathematical formulas are given as follows:
F horizontal = Conv 1 × 3 ( X enhanced ) ,
F vertical = Conv 3 × 1 ( X enhanced ) ,
where Fhorizontal and Fvertical are the features extracted in the horizontal and vertical directions, respectively; Conv1×3 and Conv3×1 are the convolution operations with (1 × 3) and (3 × 1) kernels, respectively; and Xenhanced is the input feature map.
Another deformable convolution branch uses a two-layer convolutional network to effectively predict the sampling offset Δ p 2 k 2 and modulation factor m [ 0 , 1 ] k 2 . This mechanism dynamically adjusts the receptive field according to the object’s shape and location, allowing for more accurate feature capture. Specifically, the branch first samples features at initial locations using a conventional kernel and then predicts the offset through the two-layer convolutional network. Afterwards, the sampling points are moved to better locations to dynamically adjust the receptive field:
F deform = k m k X ( p 0 + p k + Δ p k ) ,
where p k is the original sampling position of the standard convolution kernel; Δ p k is the predicted offset. By dynamically adjusting the receptive field, this structure effectively captures irregular object contours, enhancing the model’s capability to represent deformed objects. Finally, the outputs from the multiple branches are fused through element-wise summation:
F out = SiLU ( BN ( F horizontal + F vertical + F deform ) ) .
Unlike YOLO-TLA, which enhances small-object detection by simply adding a shallow detection head, the SDDE-Module introduces a more sophisticated design by jointly optimizing coordinate encoding and deformable convolutions. This is further augmented by a multibranch directional convolution mechanism, resulting in a unified structure that integrates spatial awareness and morphological flexibility. Unlike YOLO-DRS, which relies on lightweight attention mechanisms, the SDDE-Module embeds polar coordinate priors through the coordinate enhancement layer to impose explicit spatial constraints. This approach significantly improves localization precision in scenes with densely aligned objects. Additionally, the three-branch convolutional design—comprising horizontal, vertical, and deformable kernels—enables precise modeling of nonlinear deformations caused by viewpoint and angle variations, capturing detailed object structures more effectively.

4. Experiments and Results

4.1. Experimental Environment Configuration

We utilized PyTorch as the deep learning framework and adopted YOLOv8n as the baseline YOLOv8 model for training. All experiments were conducted on a 64-bit Windows 10 system powered by an Intel Core i9 CPU and an NVIDIA GeForce RTX 3090 GPU. Python 3.8 was chosen as the programming language, and PyTorch 2.0.0 and CUDA 11.6 were employed as the deep-learning frameworks.
In this study, we adopted the YOLOv8 architecture and implemented it using the Ultralytics framework. During model training, the stochastic gradient descent (SGD) optimizer was utilized with an initial learning rate of 0.01, a cosine annealing learning rate schedule, and a weight decay of 0.0005. To improve training stability, a three-epoch warm-up phase was employed, during which the momentum was set to 0.8, and the initial bias learning rate was set to 0.1.
For the loss function, the default configuration was used, which includes the bounding box regression loss (box loss), classification loss (cls loss), and distribution focal loss (DFL). In addition, to enhance the diversity of the training data, various data augmentation techniques were applied, including hue variation, saturation adjustment, brightness variation, random translation, scaling, horizontal flipping, mosaic augmentation, and random erasing, to improve the robustness of the model. Mosaic augmentation was disabled after the first 10 epochs to achieve more stable convergence outcomes.
Table 1 summarizes the detailed training environment, and Table 2 presents the training parameters.

4.2. Experimental Datasets

To comprehensively evaluate the recognition capability of the proposed network across datasets of different scales, we selected the NWPU VHR-10 [45] dataset and the DIOR [46] dataset as our experimental benchmarks. The class labels for both datasets are illustrated in Figure 6.
The NWPU VHR-10 dataset, constructed by Northwestern Polytechnical University, is a medium-scale, high-resolution remote-sensing object-detection dataset comprising 800 optical remote-sensing images. Among them, 650 images form the “positive set” containing objects, while the remaining 150 images constitute the “negative set” with no objects. The dataset covers 10 classes of geospatial objects, such as airplanes, ships, storage tanks, stadiums, and transportation infrastructure, totaling 3651 instances. All objects were manually annotated using horizontal bounding boxes. The annotations were formatted as the coordinates of the top-left and bottom-right corners along with the corresponding class label. This dataset is particularly suitable for evaluating small-object detection performance and algorithm robustness.
The DIOR dataset is a large-scale optical remote-sensing object-detection dataset containing 23,463 images of 800 × 800 pixels and 192,472 instances. It covers 20 object classes, including airplanes, airports, bridges, and windmills, encompassing both transportation infrastructure and complex scenes. Data were collected from over 80 regions worldwide, capturing a wide variety of seasons, weather conditions, and imaging settings. The spatial resolution ranges from 0.5 m to 30 m, offering pronounced intra-class variance and inter-class similarity. Its large-scale diversity makes it a key benchmark for assessing deep-learning model performance in complex remote-sensing scenes.

4.3. Evaluation Metrics

We employed precision, recall, average precision (AP), and mean average precision (mAP) as the metrics to evaluate model performance. Model scale was quantified using the number of parameters (Para) and computational cost measured in GFLOPs. The mAP metric was further divided into mAP50(%) and mAP50–95(%): mAP50 was computed at an IoU threshold of 0.5 to reflect general detection capability, and mAP50–95 averaged AP values over multiple IoU thresholds from 0.5 to 0.95, offering a more rigorous assessment of model performance. The following metrics were used: True positive (TP) refers to correctly predicted positive samples; false positive (FP) stands for negative samples incorrectly classified as positive; true negative (TN) represents correctly predicted negative samples; false negative (FN) denotes positive samples incorrectly classified as negative. The formulas for precision and recall are
Precision = TP TP + FP ,
Recall = TP TP + FN .
Average precision (AP) is calculated as the area under the precision–recall curve (AUC) for a single class. Mean average precision (mAP) is the average of AP values across all classes, computed as
AP = 0 1 P ( R ) d R ,
mAP = 1 N i = 1 N AP i

4.4. Experimental Results and Analysis

4.4.1. Ablation Experiments

To verify the effectiveness of the three proposed modules (DyCoMF-Arch, MWA-Net, and SDDE-Module), ablation experiments were conducted using both the DIOR and NWPU VHR-10 datasets. The fundamental YOLOv8 model was taken as the reference framework, and different combinations of the proposed modules were integrated into the network to investigate their mutual effects. The results of the ablation experiments on the DIOR and NWPU VHR-10 datasets are presented in Table 3 and Table 4, respectively. A check mark (√) indicates that the corresponding module was used, whereas a cross (×) denotes that it was not used.
According to the ablation results presented in Table 3 and Table 4, the proposed DyCoMF-Arch, MWA-Net, and SDDE-Module had significant performance gains in remote-sensing object-detection tasks and displayed synergistic optimization effects when combined. On the DIOR dataset, the benchmarks of YOLOv8 model achieved an mAP50 of 75.0%. After the introduction of the DyCoMF-Arch module, the mAP50 increased to 76.2% (+1.2%) with only a 0.3 M parameter increment (from 3.0 M to 3.3 M), demonstrating high parameter efficiency. The MWA-Net module, utilizing wavelet-transform-based convolution to preserve high-frequency detail features, achieved an mAP50 of 76.1% (+1.1%) with a 0.7 M parameter increment (3.0 M to 3.7 M). The SDDE-Module enhanced spatial adaptability through a multibranch deformable convolution structure and reached 76.3% mAP50 (+1.3%) with just a 0.2 M increase in parameters (3.0 M to 3.2 M), indicating strong adaptability to geometric deformations. When two modules were combined, the joint integration of DyCoMF-Arch and SDDE-Module achieved an mAP50 of 77.5% (+2.5%) on the DIOR dataset, surpassing the linear summation of their individual contributions. This result suggests complementarity between multiscale feature fusion and deformation-aware mechanisms. When all three modules were jointly employed, the model attained mAP50 values of 78.1% and 85.0% on the DIOR and NWPU datasets, respectively—improvements of 3.1% and 4.8% over the baseline. The associated computational cost increased from 8.2 GFLOPs to 11.8 GFLOPs (+44%). These results reflect a good balance between performance gains and computational cost.
Notably, different module combinations had dataset-specific behaviors. On the DIOR dataset, the SDDE-Module alone yielded the highest individual performance boost (+1.3% mAP50). In contrast, on the NWPU dataset, the combined DyCoMF-Arch and SDDE-Module configuration achieved a 3.6% improvement in mAP50 (from 80.2% to 83.8%). Thus, a greater demand for multiscale feature interaction was observed under complex background conditions. Furthermore, the combination of MWA-Net and SDDE-Module achieved 77.2% mAP50 on DIOR (+2.2%), while requiring fewer additional parameters than the sum of each module individually, suggesting the presence of parameter sharing and optimization between modules. Overall, the architectural design of the proposed network delivered significant improvements in remote-sensing object-detection performance at an acceptable computational cost.

4.4.2. Comparative Analysis Between HAF-YOLO and YOLOv8

The confusion matrices before and after the improvements are shown in Figure 7, where (a) and (c) represent the confusion matrices of the original YOLOv8 model on the NWPU VHR-10 and DIOR datasets, respectively, and (b) and (d) represent the confusion matrices of the improved HAF-YOLO model on the NWPU VHR-10 and DIOR datasets, respectively. In each confusion matrix, the horizontal axis denotes the ground-truth class labels, and the vertical axis represents the predicted class labels. The values within the cells indicate the proportion of predicted labels corresponding to each ground-truth class. A deeper color indicates a higher proportion, and blank cells signify a value of zero.
By comparing the confusion matrices across the two datasets, it can be observed that matrices (b) and (d), representing the improved model, had a higher concentration of correct predictions along the diagonal. The detection precision for all object classes was enhanced. Moreover, the values in the bottom row of the improved model’s confusion matrices were markedly reduced, indicating a lower false-negative rate across all objects. This comparative analysis of confusion matrices confirms that the proposed model achieved higher accuracy in detecting diverse object classes while reducing both false-negative and false-positive rates.
Figure 8 and Figure 9 illustrate the training curves for precision, recall, mAP50, and mAP50–95 of both YOLOv8 and HAF-YOLO across the NWPU VHR-10 and DIOR datasets. From the analysis of these figures, it can be observed that, because of the smaller scale of the NWPU VHR-10 dataset, both YOLOv8 and HAF-YOLO had relatively large training fluctuations. Nonetheless, HAF-YOLO consistently maintained a slightly higher overall performance trend compared to the original YOLO model. On the larger DIOR dataset, training curves were more stable, and HAF-YOLO consistently demonstrated a much better trend in all metrics compared to YOLOv8. The training curves on both datasets proved the effectiveness of HAF-YOLO.
To comprehensively evaluate the performance of the proposed HAF-YOLO architecture in remote-sensing object-detection tasks, we selected a variety of test samples from two representative remote sensing image datasets: NWPU VHR-10 and DIOR. These samples cover scenes involving small objects, multiscale objects, and densely arranged objects, thereby enabling a thorough assessment of the model’s capabilities in multiple dimensions. Figure 10, Figure 11 and Figure 12 present visual comparisons of detection results between YOLOv8 and HAF-YOLO across these datasets. In the original images, red bounding boxes indicate the ground-truth objects with precise position and class information. In the post-detection results from both networks, circular markers highlight missed or falsely detected instances, offering an intuitive and detailed basis for analyzing model detection performance.
Figure 10 demonstrates the enhanced detection performance of the improved network in scenes with densely arranged objects. Two classes of typical, densely distributed objects were selected: (a) a high-density storage tank array characterized by regular grid-like geometric patterns, and (b) a cluster of airplanes with significant intra-class variance and tight spacing between instances. Quantitative analyses suggest that both the baseline and the improved models were capable of initially identifying dense objects. However, the improved model showed a marked advantage in terms of detection confidence. This performance gain was attributed to the proposed feature separation mechanism, which introduces a spatial attention module to enhance the model’s ability to resolve the edge features of overlapping objects.
Figure 11 validates the effectiveness of HAF-YOLO across three representative small-object detection scenes: (a) low-resolution vehicle detection, where the average object size in pixels is small; (b) windmill detection, characterized by strong directional features and background texture similarity; and (c) small-ship detection in navigational channels, where the object color closely resembles the background. The baseline YOLOv8 network consistently had missed detections in all three scenes, primarily due to feature attenuation effects from conventional convolution operations. MWA-Net introduced a wavelet-domain decomposition mechanism that explicitly separated input features into four sub-bands: LL, LH, HL, and HH. Among them, high-frequency components effectively preserved edge oscillation features of small objects. Within the feature refinement path, the use of a depth-wise separable convolution architecture maintained a 3 × 3 receptive field while enabling efficient extraction of high-frequency detail features. Experimental results suggest that HAF-YOLO achieved full detection of all samples in the three small-object scenes. Its detection confidence significantly outperformed the baseline. It was demonstrated that HAF-YOLO enhanced small-object representation capabilities.
Figure 12 shows a comparison of the performance of HAF-YOLO and YOLOv8 in multiscale object-detection scenes. The test images include object groups with significant scale differences. The baseline model experienced notable performance degradation. The main reason is that the unidirectional information flow in traditional feature pyramids limits effective cross-scale feature interaction. The proposed DyCoMF-Arch constructs a hierarchical feature space through three-stage progressive sampling and leverages BiFPN-Concat for cross-level feature weighted fusion, establishing a bidirectional optimization mechanism between deep semantic features and shallow texture details. The results show that HAF-YOLO successfully detected all scale levels of objects and surpassed YOLOv8 in detection precision.
To provide a more intuitive demonstration of the proposed modules’ effectiveness, we compared activation heatmaps of the baseline and improved networks for detecting small objects, densely packed objects, and multiscale objects. As shown in Figure 13, in representative remote sensing scenes (a) and (b), involving small objects such as ships and windmills, the baseline model had high-frequency information loss at early convolutional stages, leading to weak activation of object regions and susceptibility to false activations from background textures such as road patterns. In contrast, HAF-YOLO, using the wavelet decomposition mechanism in MWA-Net, achieved frequency-space decoupling and enhancement of low-frequency contour information and high-frequency edge oscillation features, boosting average activation strength in object regions. In (c), for the densely arranged storage tank detection, the SDDE-Module’s spatial coordinate encoding effectively established priors on object spatial distributions and reinforced positional consistency through polar coordinate information. In (d), within a mixed multiscale object scene, the HiGA-Module’s chunked gated attention mechanism demonstrated its hierarchical feature selection capabilities: for large-scale building clusters, global attention fields were used to capture long-range dependencies, whereas for small-scale vehicle objects, local attention focused on key texture regions, reducing interference between features of different scales. Notably, the improved network excelled in background suppression. It substantially reduced false activations in interference-prone areas such as cloud shadows and road networks. Therefore, the effectiveness of the dynamic weight fusion strategy in eliminating feature redundancy was fully verified.
To better assess HAF-YOLO’s detection improvements across categories, Table 5 and Table 6 detail Precision, Recall, AP@50, and AP@50–95 for YOLOv8 and HAF-YOLO on DIOR and NWPU VHR-10 datasets. Table 5 shows HAF-YOLO consistently outperforms YOLOv8, especially for small or dense targets like vehicle, windmill, airplane, airport, and storage tank, with AP@50 gains of 2.9%, 2.1%, 3.9%, 4.1%, and 1.8%, and AP@50–95 gains of 3.2%, 3.3%, 4.3%, 7.3%, and 1.5%. Complex-structure categories (e.g., bridge, overpass, dam) also see marked improvements, indicating stronger feature representation and localization. Table 6 further confirms HAF-YOLO’s superiority on NWPU VHR-10, with notable AP@50–95 boosts of 4.0%, 6.2%, 2.7%, and 4.1% in storage tank, basketball court, bridge, and vehicle. Combining these quantitative results with the visual analysis in Figure 10, Figure 11, Figure 12 and Figure 13, we can systematically conclude that HAF-YOLO demonstrates stronger detection capabilities in scenarios involving small object sizes, high object densities, and large-scale variations, further proving its practicality in the field of remote-sensing object detection.

4.4.3. Comparative Experiments with Other Models

We compared the proposed HAF-YOLO model with several classical models in the field of object detection. Specifically, we selected representative multistage detectors, including Fast R-CNN and Faster R-CNN, along with classical single-stage detectors such as SSD, and several versions of the YOLO series: YOLOv3, YOLOv5, YOLOv7, YOLOv9, YOLOv10, YOLOv11, and YOLOv12. Specifically, state-of-the-art YOLO variants optimized for remote-sensing image object detection were selected, as discussed in Section 2.3. These include AAPW-YOLO [37] (enhancing YOLOv8), YOLO-TLA [38] (based on YOLOv5), MS-YOLOv7 [41], and YOLO-DRS [42]. All experiments were conducted under the same experimental settings. Table 7 and Table 8 present the comparative results across models on the NWPU VHR-10 and DIOR datasets, respectively, in terms of parameters, floating-point operations (FLOPs), precision, recall, and mAP metrics.
According to the comparative results listed in Table 7 and Table 8, the proposed HAF-YOLO demonstrated a significant advantage in overall performance for object detection in remote-sensing images. On the NWPU VHR-10 dataset, HAF-YOLO achieved 87.6% precision, 79.8% recall, 85.0% mAP@50, and 50.2% mAP@50:95 with only 4.3 M parameters and a computational cost of 11.8 GFLOPs, outperforming all contrastive models across all four key metrics. Compared to the most recent version in the YOLO series, YOLOv11 (9.4 M parameters, 21.6 GFLOPs), HAF-YOLO not only achieved a +0.8% improvement in mAP@50:95 but also compressed the model size by 54.3% and reduced computation by 45.8%. These results verify that the DyCoMF-Arch effectively integrates cross-scale features while reducing computational redundancy by using multistage spatial compression and a BiFPN-based concatenated weighted fusion strategy.
On the large-scale DIOR remote-sensing dataset, HAF-YOLO attained a mAP@50:95 of 55.4%, representing a 1.3-point improvement over YOLOv11, while also boosting precision by 2.7% to 87.1%. Despite the presence of more complex background interference and a higher number of small-object instances in this dataset, HAF-YOLO maintained a recall of 71.9%, surpassing YOLOv11 by 3%. The SDDE-Module, through its integration of spatial coordinate encoding and deformable convolution, significantly enhanced the spatial localization capability for dense small objects. Notably, the traditional two-stage detector Faster R-CNN achieved a mAP@50:95 of only 45.1% on DIOR, lagging HAF-YOLO by 10.3%. This result verifies the applicability of single-stage detection frameworks combined with hierarchical adaptive fusion strategies in remote-sensing detection tasks.
From a model efficiency perspective, HAF-YOLO achieved a superior balance between parameter count and computational complexity. Compared with YOLOv5 (2.5 M parameters, 7.1 GFLOPs), HAF-YOLO achieved a 3.1% improvement in mAP@50:95 on the NWPU VHR-10 dataset at the cost of only a 1.7-fold increase in parameters. Therefore, substantial performance gain was brought by the multigranularity feature capture mechanism of the DyCoMF-Arch. Compared to the lightweight design of YOLOv9 (1.9 M parameters, 7.6 GFLOPs), HAF-YOLO improved the mAP@50:95 by 2.1% on the DIOR dataset. In addition, the computational cost of HAF-YOLO was constrained to 11.8 GFLOPs, which was significantly lower than that of comparable models such as YOLOv10 (21.4 GFLOPs) and YOLOv12 (21.2 GFLOPs).
Compared to the state-of-the-art YOLOv11, HAF-YOLO achieved a 1.3% increase in mAP@50:95 on the DIOR dataset while utilizing only 54.7% of the parameters. This result validates the effectiveness of the proposed synergistic optimization mechanism in balancing model efficiency and detection accuracy. Moreover, HAF-YOLO exhibited stronger generalization capability in cross-dataset evaluations: The degradation rate of mAP@50:95 from NWPU VHR-10 to DIOR was only 10.2%, which was significantly lower than that of YOLOv8 (14.7%) and YOLOv7 (22.8%). This result demonstrates that the multistage spatial compression and adaptive feature enhancement strategies substantially improve model robustness to scene variations in remote sensing.
To further evaluate the applicability of the proposed model in real-world environments, we analyzed the frames per second (FPS) performance of HAF-YOLO in comparison to other models. The relevant data is shown in Table 7 and Table 8. On the NWPU VHR-10 dataset, HAF-YOLO achieved 161 FPS, outperforming high-precision models such as YOLOv11 (128 FPS) and YOLOv12 (90 FPS), and approaching YOLO-DRS (169 FPS). On the DIOR dataset, HAF-YOLO reached 164 FPS, surpassing the inference speed of most similar models while maintaining superior detection accuracy (mAP50:95 of 55.4%). Compared to YOLOv11, which has a slightly higher mAP50 on DIOR (76.8%), its speed is only 132 FPS, whereas our model improved by 32 FPS, increasing mAP50:95 by 1.3%, while reducing the number of parameters by 54.7%. Similarly, compared to AAPW-YOLO (158 FPS) and YOLO-TLA (195 FPS), our method provided significantly better detection accuracy on NWPU VHR-10 with only a slight reduction in speed.
In addition to the standard YOLO series detectors, comparisons with specialized remote-sensing-optimized models provide deeper insights. On the NWPU VHR-10 dataset, AAPW-YOLO achieved an mAP@50 of 83.6% because of its adaptive convolution targeting small objects; however, HAF-YOLO surpassed it with a 1.4% margin. Similarly, YOLO-TLA, which incorporates an additional detection layer tailored for small-object detection based on YOLOv5, reached an mAP@50 of 82.1%, but still lagged behind HAF-YOLO by 2.9%. MS-YOLOv7, by incorporating an additional detection head to handle multiscale objects, achieved an mAP@50 of 81.2%, but at the cost of higher computational overhead (13.3 GFLOPs) compared to HAF-YOLO’s 11.8 GFLOPs. YOLO-DRS, focusing on lightweight multiscale fusion, achieved an mAP@50 of 83.6% with fewer parameters and GFLOPs, but its mAP@50 was still 1.4% lower than HAF-YOLO’s, and its mAP@50:95 was 0.6% lower. On the DIOR dataset, HAF-YOLO demonstrated even more significant advantages over specialized models. AAPW-YOLO achieved an mAP@50 of 76.8%, whereas HAF-YOLO outperformed it by 1.3%. MS-YOLOv7 achieved an mAP@50 of 75.5%, lower than HAF-YOLO’s 78.1%. YOLO-DRS achieved an mAP@50 of 76.8%, also 1.3% lower than HAF-YOLO.
The comparative experimental results on the two datasets validate the effectiveness of HAF-YOLO. The visual comparison in Figure 14 further corroborates the superiority of HAF-YOLO. The test images shown in Figure 14 include a large-scale target of a bridge and four small-scale targets of vehicles. For the detection of small vehicle targets, the HAF-YOLO network successfully detected all the small vehicle targets. Although some of the comparative networks achieved higher detection precision on certain targets than HAF-YOLO, they failed to detect all the small vehicle targets. The comparative experimental results highlight the advantages of the HAF-YOLO network in detecting small-scale targets and multiscale, cross-scale objects.
Figure 15 presents radar plots comparing the AP@50 values for different YOLO series models and HAF-YOLO across object categories. Figure 15a shows results for 20 object classes on the DIOR dataset, and Figure 15b shows results for 10 categories on the NWPU VHR-10 dataset. Different colors represent different models, as indicated in the legend at the top of each plot, enabling a clear visual comparison of model performance across object types. The radar plots and the model-wise per-class performance show that the detection performance varied with object classes. For large-scale objects (e.g., bridges, dams, and airports), HAF-YOLO showed obvious improvements in detection accuracy. Similarly, for small objects such as vehicles and ships, it achieved the highest mAP values, validating its enhanced capability in small-object detection. Overall, HAF-YOLO consistently outperformed other models across all object classes and achieved broader coverage, as shown by the radar plots for both datasets.

5. Conclusions

In this study, we proposed the HAF-YOLO model to address challenges in object detection for remote-sensing images, such as the vulnerability of small-object features to loss, complex background interference, diverse object scales, and densely distributed objects. Three core modules were proposed: DyCoMF-Arch, MWA-Net, and SDDE-Module, which respectively refine the conventional YOLOv8 framework from the perspectives of feature fusion, scale enhancement, and spatial adaptability.
DyCoMF-Arch constructs a hierarchical feature pyramid by multistage spatial compression and expansion operations and combines a dynamic weight allocation mechanism to achieve precise cross-scale feature fusion. MWA-Net employs wavelet-transform-based convolution to decompose features, explicitly separating low-frequency contours and high-frequency details. This approach, combined with a dynamic weighted fusion strategy, enhances the feature representation for small objects and effectively mitigates the high-frequency information loss common in traditional convolution operations. The SDDE-Module integrates absolute coordinate information through a coordinate enhancement layer and couples it with multidirectional convolutions and a deformable sampling mechanism to enhance the model’s adaptability to the spatial distribution of small objects. Experiments on the NWPU VHR-10 and DIOR datasets show that HAF-YOLO achieved mAP50 scores of 85.0% and 78.1%, respectively, surpassing the baseline YOLOv8 by 4.8% and 3.1%, while maintaining a computational cost of only 11.8 GFLOPs, which was lower than that of other models in the YOLO series. Comparative experiments show that HAF-YOLO achieved a more optimal balance between parameter efficiency and detection accuracy. Visualization results further confirm that HAF-YOLO significantly improved detection confidence in scenes with dense objects, small objects, and multiscale scenes while markedly reducing both the false-negative rate and the false-positive rate. This model provides efficient and reliable technical support for remote-sensing image analysis in domains such as agricultural monitoring and urban planning.
In this investigation, although the HAF-YOLO network demonstrated superior detection performance, it remained constrained by a relatively high number of computational parameters. In future, we will explore the integration of synthetic aperture radar (SAR) image datasets to improve model robustness under complex conditions through dual-modality data, as well as investigate more lightweight enhancements to better balance accuracy and detection speed, thereby maintaining superior performance in future aerial image-detection tasks.

Author Contributions

Conceptualization, J.L. and P.Z.; methodology, P.Z., J.L. and J.Z.; software, P.Z., Y.L. and J.S.; validation, P.Z., J.Z., Y.L. and J.S.; formal analysis, P.Z., Y.L. and J.L.; investigation, P.Z., J.Z., Y.L. and J.S.; resources, P.Z., J.L. and J.Z.; data curation, Y.L., P.Z. and J.S.; writing—original draft preparation, P.Z. and J.Z.; writing—review and editing, P.Z., J.L. and Y.L.; visualization, Y.L., P.Z. and J.S.; supervision, J.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

The original contributions presented in this study are included in the article. Further inquiries can be directed to the corresponding author.

Acknowledgments

I would like to express my heartfelt gratitude to all the authors whose works have contributed to this paper. Their invaluable insights, rigorous research, and scholarly discussions have greatly enriched my understanding and helped shape the foundation of this study. Without their efforts and dedication to advancing knowledge, this paper would not have been possible.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

APAverage precision
CGAAChunked gated adaptive attention
CNNConvolutional neural network
FLOPFloating-point operation
FNFalse negative
FPFalse positive
FPNFeature pyramid network
GELANGeneral efficient layer aggregation network
HOGHistogram of oriented gradients
IoUIntersection over union
MHSAMultihead self-attention
NMSNon-maximum suppression
PGIProgrammable gradient information
ROIRegion of interest
RPNRegion proposal network
SARSynthetic aperture radar
SDDESpatial-deformable dynamic enhancement
SESqueeze-and-excitation
SPPFSpatial pyramid pooling fast
SSDSingle-shot multibox detector
SVMSupport vector machine
TAATask-aligned assigner
TNTrue negative
TPTrue positive
UAVUnmanned aerial vehicle
YOLOYou Only Look Once

References

  1. Zheng, Z.; Yuan, J.; Yao, W.; Kwan, P.; Yao, H.; Liu, Q.; Guo, L. Fusion of UAV-acquired visible images and multispectral data by applying machine-learning methods in crop classification. Agronomy 2024, 14, 2670. [Google Scholar] [CrossRef]
  2. Zhang, Z.; Yao, F.; Li, J. Dynamic Penetration Test Based on YOLOv5. In Proceedings of the 2022 3rd International Conference on Geology, Mapping and Remote Sensing (ICGMRS), Zhoushan, China, 22–24 April 2022. [Google Scholar]
  3. Morita, M.; Kinjo, H.; Sato, S.; Tansuriyavong; Sulyyon; Anezaki, T. Autonomous Flight Drone for Infrastructure (Transmission Line) Inspection. In Proceedings of the 2017 International Conference on Intelligent Informatics and Biomedical Sciences (ICIIBMS), Okinawa, Japan, 24–26 November 2017. [Google Scholar]
  4. Menkhoff, T.; Tan, E.K.B.; Ning, K.S.; Hup, T.G.; Pan, G. Tapping Drone Technology to Acquire 21st Century Skills: A Smart City Approach. In Proceedings of the 2017 IEEE SmartWorld, San Francisco, CA, USA, 4–8 August 2017. [Google Scholar]
  5. Ryoo, D.-W.; Lee, M.-S.; Lim, C.-D. Design of a Drone-Based Real-Time Service System for Facility Inspection. In Proceedings of the 2023 14th International Conference on Information and Communication Technology Convergence (ICTC), Jeju Island, Republic of Korea, 11–13 October 2023. [Google Scholar]
  6. Sun, X.; Zhang, W. Implementation of Target Tracking System Based on Small Drone. In Proceedings of the 2019 IEEE 4th Advanced Information Technology, Electronic and Automation Control Conference (IAEAC), Chengdu, China, 20–22 December 2019. [Google Scholar]
  7. Wang, Z.; Xia, F.; Zhang, C. FD_YOLOX: An Improved YOLOX Object Detection Algorithm Based on Dilated Convolution. In Proceedings of the 2023 IEEE 18th Conference on Industrial Electronics and Applications (ICIEA), Ningbo, China, 18–22 August 2023. [Google Scholar]
  8. Zong, H.; Pu, H.; Zhang, H.; Wang, X.; Zhong, Z.; Jiao, Z. Small Object Detection in UAV Image Based on Slicing Aided Module. In Proceedings of the 2022 IEEE 4th International Conference on Power, Intelligent Computing and Systems (ICPICS), Shenyang, China, 29–31 July 2022. [Google Scholar]
  9. Dalal, N.; Triggs, B. Histograms of Oriented Gradients for Human Detection. In Proceedings of the 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR’05), San Diego, CA, USA, 20–25 June 2005. [Google Scholar]
  10. Lowe, D.G. Object Recognition from Local Scale-Invariant Features. In Proceedings of the Seventh IEEE International Conference on Computer Vision, Corfu, Greece, 20–27 September 1999. [Google Scholar]
  11. Pan, W.; Huan, W.; Xu, L. Improving High-Voltage Line Obstacle Detection with Multi-Scale Feature Fusion in YOLO Algorithm. In Proceedings of the 2024 6th International Conference on Electronics and Communication, Network and Computer Technology (ECNCT), Guangzhou, China, 19–21 July 2024. [Google Scholar]
  12. LeCun, Y.; Boser, B.; Denker, J.S.; Henderson, D.; Howard, R.E.; Hubbard, W.; Jackel, L.D. Backpropagation applied to handwritten zip code recognition. Neural Comput. 1989, 1, 541–551. [Google Scholar] [CrossRef]
  13. Lin, T.-Y.; Goyal, P.; Girshick, R.; He, K.; Dollar, P. Feature Pyramid Networks for Object Detection. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017. [Google Scholar]
  14. Tan, M.; Pang, R.; Le, Q.V. EfficientDet: Scalable and Efficient Object Detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 14–19 June 2020. [Google Scholar]
  15. Viola, P.; Jones, M. Rapid Object Detection Using a Boosted Cascade of Simple Features. In Proceedings of the 2001 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR), Kauai, HI, USA, 8–14 December 2001. [Google Scholar]
  16. Girshick, R.; Donahue, J.; Darrell, T.; Malik, J. Rich Feature Hierarchies for Accurate Object Detection and Semantic Segmentation. In Proceedings of the 2014 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Columbus, OH, USA, 23–28 June 2014. [Google Scholar]
  17. Girshick, R. Fast R-CNN. In Proceedings of the 2015 IEEE International Conference on Computer Vision (ICCV), Santiago, Chile, 7–13 December 2015. [Google Scholar]
  18. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards real-time object detection with region proposal networks. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 1137–1149. [Google Scholar] [CrossRef]
  19. Cai, Z.; Vasconcelos, N. Cascade R-CNN: Delving into high quality object detection. arXiv 2017, arXiv:1712.00726. [Google Scholar]
  20. Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.-Y.; Berg, A.C. SSD: Single Shot MultiBox Detector. In Proceedings of the European Conference on Computer Vision (ECCV), Amsterdam, The Netherlands, 8–16 October 2016. [Google Scholar]
  21. Huang, G.; Liu, Z.; Weinberger, K.Q. Densely connected convolutional networks. arXiv 2016, arXiv:1608.06993. [Google Scholar]
  22. Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You Only Look Once: Unified, Real-Time Object Detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016. [Google Scholar]
  23. Redmon, J.; Farhadi, A. YOLO9000: Better, Faster, Stronger. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017. [Google Scholar]
  24. Farhadi, A.; Redmon, J. YOLOv3: An Incremental Improvement. In Proceedings of the Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018. [Google Scholar]
  25. Bochkovskiy, A.; Wang, C.-Y.; Liao, H.-Y.M. YOLOv4: Optimal speed and accuracy of object detection. arXiv 2020, arXiv:2004.10934. [Google Scholar]
  26. Jocher, G.; Stoken, A.; Chaurasia, A.; Borovec, J.; Kwon, Y.; Michael, K.; Changyu, L.; Fang, J.; Skalski, P.; Hogan, A.; et al. YOLOv5. Available online: https://github.com/ultralytics/yolov5 (accessed on 27 August 2024).
  27. Li, C.; Li, L.; Jiang, H.; Weng, K.; Geng, Y.; Li, L.; Ke, Z.; Li, Q.; Cheng, M.; Nie, W.; et al. YOLOv6: A single-stage object detection framework for industrial applications. arXiv 2022, arXiv:2209.02976. [Google Scholar]
  28. Wang, C.-Y.; Bochkovskiy, A.; Liao, H.-Y.M. YOLOv7: Trainable Bag-of-Freebies Sets New State-of-the-Art for Real-Time Object Detectors. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023. [Google Scholar]
  29. Joher, G.; Chaurasia, A.; Qiu, J. YOLOv8. Available online: https://github.com/ultralytics/ultralytics (accessed on 27 August 2024).
  30. Wang, C.-Y.; Yeh, I.-H.; Liao, H.-Y.M. YOLOv9: Learning what you want to learn using programmable gradient information. arXiv 2024, arXiv:2402.13616. [Google Scholar]
  31. Wang, A.; Chen, H.; Liu, L.; Chen, K.; Lin, Z.; Han, J.; Ding, G. YOLOv10: Real-time end-to-end object detection. arXiv 2024, arXiv:2405.14458. [Google Scholar]
  32. Hu, J.; Pang, T.; Peng, B.; Shi, Y.; Li, T. A small object detection model for drone images based on multi-attention fusion network. Image Vis. Comput. 2025, 155, 105436. [Google Scholar] [CrossRef]
  33. Liu, M.; Wang, X.; Zhou, A.; Fu, X.; Ma, Y.; Piao, C. UAV-YOLO: Small object detection on unmanned aerial vehicle perspective. Sensors 2020, 20, 2238. [Google Scholar] [CrossRef]
  34. Wang, X.; He, N.; Hong, C.; Wang, Q.; Chen, M. Improved YOLOX-X based UAV aerial photography object detection algorithm. Image Vis. Comput. 2023, 135, 104697. [Google Scholar] [CrossRef]
  35. Koyun, O.C.; Keser, R.K.; Akkaya, I.B.; Töreyin, B.U. Focus-and-detect: A small object detection framework for aerial images. Signal Process. Image Commun. 2022, 104, 116675. [Google Scholar] [CrossRef]
  36. Qu, J.; Tang, Z.; Zhang, L.; Zhang, Y.; Zhang, Z. Remote sensing small object detection network based on attention mechanism and multi-scale feature fusion. Remote Sens. 2023, 15, 2728. [Google Scholar] [CrossRef]
  37. Wu, Y.; Mu, X.; Shi, H.; Hou, M. An object detection model AAPW-YOLO for UAV remote sensing images based on adaptive convolution and reconstructed feature fusion. Sci. Rep. 2025, 15, 16214. [Google Scholar] [CrossRef]
  38. Ji, C.L.; Yu, T.; Gao, P.; Wang, F.; Yuan, R.Y. YOLO-TLA: An efficient and lightweight small object detection model based on YOLOv5. J. Real-Time Image Process. 2024, 21, 141. [Google Scholar] [CrossRef]
  39. Hu, W.; Jiang, X.; Tian, J.; Ye, S.; Liu, S. Land target detection algorithm in remote sensing images based on deep learning. Land 2025, 14, 1047. [Google Scholar] [CrossRef]
  40. Liu, S.; Qi, L.; Qin, H.; Shi, J.; Jia, J. Path Aggregation Network for Instance Segmentation. In Proceedings of the 2018 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–23 June 2018. [Google Scholar]
  41. Zhao, L.L.; Zhu, M.L. MS-YOLOv7: YOLOv7 based on multi-scale for object detection on UAV aerial photography. Drones 2023, 7, 188. [Google Scholar] [CrossRef]
  42. Liao, H.; Zhu, W. YOLO-DRS: A bioinspired object detection algorithm for remote sensing images incorporating a multi-scale efficient lightweight attention mechanism. Biomimetics 2023, 8, 458. [Google Scholar] [CrossRef] [PubMed]
  43. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, L.; Polosukhin, I. Attention is all you need. Adv. Neural Inf. Process. Syst. 2017, 30, 5998–6008. [Google Scholar]
  44. Finder, S.E.; Amoyal, R.; Treister, E.; Freifeld, O. Wavelet Convolutions for Large Receptive Fields. In Proceedings of the European Conference on Computer Vision, Milan, Italy, 29 September–4 October 2024. [Google Scholar]
  45. Cheng, G.; Han, J.; Zhou, P.; Guo, L. Multi-class geospatial object detection and geographic image classification based on collection of part detectors. ISPRS J. Photogramm. Remote Sens. 2014, 98, 119–132. [Google Scholar] [CrossRef]
  46. Li, K.; Wan, G.; Cheng, G.; Meng, L.; Han, J. Object Detection in Optical Remote Sensing Images: A Survey and A New Benchmark. arXiv 2019, arXiv:1909.00133. [Google Scholar] [CrossRef]
Figure 1. Structure of HAF-YOLO.
Figure 1. Structure of HAF-YOLO.
Remotesensing 17 02708 g001
Figure 2. Network structure of the HiGA-Module.
Figure 2. Network structure of the HiGA-Module.
Remotesensing 17 02708 g002
Figure 3. Architecture of the MWA-Net.
Figure 3. Architecture of the MWA-Net.
Remotesensing 17 02708 g003
Figure 4. Architecture of the SDDE-Module.
Figure 4. Architecture of the SDDE-Module.
Remotesensing 17 02708 g004
Figure 5. Principle of the coordinate enhancement layer.
Figure 5. Principle of the coordinate enhancement layer.
Remotesensing 17 02708 g005
Figure 6. Data distribution: (a) DIOR dataset; (b) NWPU VHR-10 dataset.
Figure 6. Data distribution: (a) DIOR dataset; (b) NWPU VHR-10 dataset.
Remotesensing 17 02708 g006
Figure 7. Confusion matrix comparison: (a) Confusion matrix of YOLOv8 on the NWPU VHR-10 dataset, (b) confusion matrix of HAF-YOLO on the NWPU VHR-10 dataset, (c) confusion matrix of YOLOv8 on the DIOR dataset, and (d) confusion matrix of HAF-YOLO on the DIOR dataset.
Figure 7. Confusion matrix comparison: (a) Confusion matrix of YOLOv8 on the NWPU VHR-10 dataset, (b) confusion matrix of HAF-YOLO on the NWPU VHR-10 dataset, (c) confusion matrix of YOLOv8 on the DIOR dataset, and (d) confusion matrix of HAF-YOLO on the DIOR dataset.
Remotesensing 17 02708 g007
Figure 8. Training curves on the NWPU VHR-10 dataset: (a) Precision, (b) recall, (c) mAP50, (d) mAP50–95.
Figure 8. Training curves on the NWPU VHR-10 dataset: (a) Precision, (b) recall, (c) mAP50, (d) mAP50–95.
Remotesensing 17 02708 g008
Figure 9. Training curves on DIOR dataset: (a) Precision, (b) recall, (c) mAP50, (d) mAP50–95.
Figure 9. Training curves on DIOR dataset: (a) Precision, (b) recall, (c) mAP50, (d) mAP50–95.
Remotesensing 17 02708 g009
Figure 10. Detection of scenes involving densely arranged objects: (a) Storage tanks, (b) airplanes.
Figure 10. Detection of scenes involving densely arranged objects: (a) Storage tanks, (b) airplanes.
Remotesensing 17 02708 g010
Figure 11. Small-object detection scenes: (a) Vehicles, (b) windmills, and (c) ships.
Figure 11. Small-object detection scenes: (a) Vehicles, (b) windmills, and (c) ships.
Remotesensing 17 02708 g011
Figure 12. (ac) Multiscale object-detection scenes.
Figure 12. (ac) Multiscale object-detection scenes.
Remotesensing 17 02708 g012
Figure 13. Detection heatmaps: (a) Small ship objects, (b) small windmill objects, (c) densely arranged objects, (d) multiscale objects.
Figure 13. Detection heatmaps: (a) Small ship objects, (b) small windmill objects, (c) densely arranged objects, (d) multiscale objects.
Remotesensing 17 02708 g013
Figure 14. Visualization of detection experiments using different networks: (a) Original images, (b) YOLOv3, (c) YOLOv5, (d) YOLOv6, (e) YOLOv7, (f) YOLOv8, (g) YOLOv9, (h) YOLOv10, (i) YOLOv11, (j) YOLOv12, (k) AAPW-YOLO, (l) YOLO-TLA, (m) MS-YOLOv7, (n) YOLO-DRS, and (o) HAF-YOLO.
Figure 14. Visualization of detection experiments using different networks: (a) Original images, (b) YOLOv3, (c) YOLOv5, (d) YOLOv6, (e) YOLOv7, (f) YOLOv8, (g) YOLOv9, (h) YOLOv10, (i) YOLOv11, (j) YOLOv12, (k) AAPW-YOLO, (l) YOLO-TLA, (m) MS-YOLOv7, (n) YOLO-DRS, and (o) HAF-YOLO.
Remotesensing 17 02708 g014aRemotesensing 17 02708 g014b
Figure 15. Comparative analysis of mAP@50 (%) results across different models from experiments with the (a) DIOR dataset and (b) NWPU VHR-10 dataset.
Figure 15. Comparative analysis of mAP@50 (%) results across different models from experiments with the (a) DIOR dataset and (b) NWPU VHR-10 dataset.
Remotesensing 17 02708 g015
Table 1. Experimental training environment.
Table 1. Experimental training environment.
ParameterConfiguration
CPUIntel(R) Core(TM) i9-10940X
GPUNVIDIA GeForce RTX 3090
SystemWindows 10
Deep learning frameworkPyTorch 2.0.0
GPU acceleratorCUDA 11.6
Integrated development environmentPyCharm
Scripting languagePython 3.8
Table 2. Training parameters.
Table 2. Training parameters.
ParameterConfiguration
Epochs150
Workers10
Batch10
lr00.01
Momentum0.937
weight_decay0.0005
Network optimizerSGD
box loss weight7.5
cls loss weight0.5
dfl loss weight1.5
mosaic1.0
IOUC-IOU
translate0.1
scale0.5
fliplr0.5
Table 3. Ablation experiments using the DIOR dataset.
Table 3. Ablation experiments using the DIOR dataset.
YOLOv8DyCoMF-ArchMWA-NetSDDE-ModuleP
(%)
R
(%)
mAP50
(%)
mAP50:95
(%)
Param
(M)
GFLOPs
×××84.666.975.052.13.08.2
××85.468.876.253.33.39.0
××85.568.076.153.23.710.4
××85.267.876.353.73.28.7
×85.969.476.853.94.011.2
×86.269.877.554.23.69.6
×86.470.977.254.93.910.9
87.171.978.155.44.311.8
Table 4. Ablation experiments using the NWPU VHR-10 dataset.
Table 4. Ablation experiments using the NWPU VHR-10 dataset.
YOLOv8DyCoMF-ArchMWA-NetSDDE-ModuleP
(%)
R
(%)
mAP50
(%)
mAP50:95
(%)
Param
(M)
GFLOPs
×××84.975.180.247.43.08.2
××85.677.181.447.93.39.0
××85.776.881.547.53.710.4
××85.476.681.748.13.28.7
×86.177.882.448.24.011.2
×86.278.183.848.73.69.6
×86.278.583.548.83.910.9
87.679.885.050.24.311.8
Table 5. Comparison of recognition performance for each category on the DIOR dataset.
Table 5. Comparison of recognition performance for each category on the DIOR dataset.
P (%)R (%)AP50 (%)AP50:95 (%)
YOLOv8HAF-YOLOYOLOv8HAF-YOLOYOLOv8HAF-YOLOYOLOv8HAF-YOLO
All0.8470.8710.6690.7190.750.7810.5210.554
Airplane0.9440.9620.6350.7080.7880.8270.5320.575
Airport0.8050.8320.8030.8760.8570.8980.5850.658
Baseball field0.9540.9700.6980.7320.8470.8580.6880.712
Basketball court0.9240.9450.8450.8730.8890.9020.7570.794
Bridge0.7390.7680.3490.4050.4320.4870.2390.290
Chimney0.9700.9680.7250.7450.7650.7920.6450.682
Dam0.6980.7050.6570.7120.6980.7240.3910.435
Expressway service area0.8360.8520.8110.8560.8630.8880.6060.660
Expressway toll station0.9080.9120.5610.6250.6550.6960.5000.545
Golf course0.7730.8250.7860.8480.8210.8690.5870.685
Ground track field0.7200.7350.7720.8250.7920.8170.5980.640
Harbor0.7650.7850.6110.6420.6680.6840.4740.510
Overpass0.8440.8600.5230.5620.6170.6420.4130.445
Ship0.9290.9420.8320.8650.9010.9140.5490.575
Stadium0.8430.8850.5950.6520.7830.8020.6040.625
Storage tank0.9610.9680.5530.5820.7330.7510.4570.472
Tennis court0.9500.9580.8600.8750.9110.9230.7640.788
Train station0.6330.6350.6500.7050.6530.670.3390.395
Vehicle0.8890.9050.3340.3680.4840.5130.2680.300
Windmill0.8660.9020.7810.8280.8440.8650.4250.458
Table 6. Comparison of recognition performance for each category on the NWPU VHR-10 dataset.
Table 6. Comparison of recognition performance for each category on the NWPU VHR-10 dataset.
P (%)R (%)AP50 (%)AP50:95 (%)
YOLOv8HAF-YOLOYOLOv8HAF-YOLOYOLOv8HAF-YOLOYOLOv8HAF-YOLO
All0.8460.8760.7510.7980.8040.8500.4750.502
Airplane0.9360.9520.9360.9650.9770.9880.5810.591
Ship0.8550.8570.8550.6860.7140.7680.4390.439
Storage tank0.8250.9010.8250.8740.7600.8710.3840.424
Baseball diamond0.9270.9410.9270.9720.9810.9710.7140.716
Tennis court0.8880.8630.8880.9040.8770.9220.5120.545
Basketball court0.6230.7360.6230.5220.5250.6060.2860.348
Ground track field0.9540.9480.9540.9430.9720.9710.7320.736
Harbor0.8600.8540.8600.9030.9230.9250.4680.510
Bridge0.7150.8170.7150.4770.5460.6330.2090.236
Vehicle0.8740.8860.8740.7380.7630.8440.4300.471
Table 7. Comparative results on the NWPU VHR-10 dataset.
Table 7. Comparative results on the NWPU VHR-10 dataset.
ModelP
(%)
R
(%)
mAP50
(%)
mAP50:95
(%)
Param
(M)
GFLOPsFPS
Traditional Methods
Faster R-CNN85.276.882.346.826.545.641
SSD83.574.180.544.224.132.763
YOLO Series
YOLOv387.179.984.548.912.118.9178
YOLOv582.776.381.547.12.57.1198
YOLOv684.47882.148.64.211.8196
YOLOv781.178.981.242.96.013.3175
YOLOv884.975.180.247.43.08.2188
YOLOv983.872.678.746.31.97.6168
YOLOv1072.671.57545.17.221.4142
YOLOv1187.678.283.649.49.421.6128
YOLOv1283.774.179.645.89.221.290
RS-Specific SOTA
AAPW-YOLO86.578.083.649.03.510.5158
YOLO-TLA85.077.582.148.03.29.0195
MS-YOLOv781.579.081.247.06.514.5176
YOLO-DRS86.078.583.649.63.910.9169
Ours
HAF-YOLO87.679.885.050.24.311.8161
Table 8. Comparative results on the DIOR dataset.
Table 8. Comparative results on the DIOR dataset.
ModelP
(%)
R
(%)
mAP50
(%)
mAP50:95
(%)
Param
(M)
GFLOPsFPS
Traditional Methods
Faster R-CNN84.765.970.845.126.545.643
SSD82.363.568.242.724.132.764
YOLO Series
YOLOv38362.868.547.012.118.9185
YOLOv583.465.372.749.22.57.1194
YOLOv682.462.770.278.54.211.8203
YOLOv785.167.275.552.66.013.3182
YOLOv884.666.975.052.13.08.2190
YOLOv983.866.974.953.31.97.6163
YOLOv1083.163.271.548.27.221.4145
YOLOv1184.468.976.854.19.421.6132
YOLOv1284.667.675.953.99.221.288
RS-Specific SOTA
AAPW-YOLO85.068.576.854.03.510.5154
YOLO-TLA83.867.074.552.03.29.0192
MS-YOLOv785.567.575.552.86.514.5182
YOLO-DRS84.869.076.854.03.910.9173
Ours
HAF-YOLO87.171.978.155.44.311.8164
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhang, P.; Liu, J.; Zhang, J.; Liu, Y.; Shi, J. HAF-YOLO: Dynamic Feature Aggregation Network for Object Detection in Remote-Sensing Images. Remote Sens. 2025, 17, 2708. https://doi.org/10.3390/rs17152708

AMA Style

Zhang P, Liu J, Zhang J, Liu Y, Shi J. HAF-YOLO: Dynamic Feature Aggregation Network for Object Detection in Remote-Sensing Images. Remote Sensing. 2025; 17(15):2708. https://doi.org/10.3390/rs17152708

Chicago/Turabian Style

Zhang, Pengfei, Jian Liu, Jianqiang Zhang, Yiping Liu, and Jiahao Shi. 2025. "HAF-YOLO: Dynamic Feature Aggregation Network for Object Detection in Remote-Sensing Images" Remote Sensing 17, no. 15: 2708. https://doi.org/10.3390/rs17152708

APA Style

Zhang, P., Liu, J., Zhang, J., Liu, Y., & Shi, J. (2025). HAF-YOLO: Dynamic Feature Aggregation Network for Object Detection in Remote-Sensing Images. Remote Sensing, 17(15), 2708. https://doi.org/10.3390/rs17152708

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Article metric data becomes available approximately 24 hours after publication online.
Back to TopTop