Next Article in Journal
Coupling Machine Learning and Physically Based Hydrological Models for Reservoir-Based Streamflow Forecasting
Previous Article in Journal
YOLO-SRMX: A Lightweight Model for Real-Time Object Detection on Unmanned Aerial Vehicles
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

SDRFPT-Net: A Spectral Dual-Stream Recursive Fusion Network for Multispectral Object Detection

College of Intelligence Science and Technology, National University of Defense Technology, Changsha 410073, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2025, 17(13), 2312; https://doi.org/10.3390/rs17132312
Submission received: 13 May 2025 / Revised: 27 June 2025 / Accepted: 4 July 2025 / Published: 5 July 2025

Abstract

Multispectral object detection faces challenges in effectively integrating complementary information from different modalities in complex environmental conditions. This paper proposes SDRFPT-Net (Spectral Dual-stream Recursive Fusion Perception Target Network), a novel architecture that integrates three key innovative modules: (1) the Spectral Hierarchical Perception Architecture (SHPA), which adopts a dual-stream separated structure with independently parameterized feature extraction paths for visible and infrared modalities; (2) the Spectral Recursive Fusion Module (SRFM), which combines hybrid attention mechanisms with recursive progressive fusion strategies to achieve deep feature interaction through parameter-sharing multi-round recursive processing; and (3) the Spectral Target Perception Enhancement Module (STPEM), which adaptively enhances target region representation and suppresses background interference. Extensive experiments on the VEDAI, FLIR-aligned, and LLVIP datasets demonstrate that SDRFPT-Net significantly outperforms state-of-the-art methods, with improvements of 2.5% in mAP50 and 5.4% in mAP50:95 on VEDAI, 11.5% in mAP50 on FLIR-aligned, and 9.5% in mAP50:95 on LLVIP. Ablation studies further validate the effectiveness of each proposed module. The proposed method provides an efficient and robust solution for multispectral object detection in remote sensing image interpretation, making it particularly suitable for all-weather monitoring applications from aerial and satellite platforms, as well as in intelligent surveillance and autonomous driving domains.

1. Introduction

As one of the core tasks in computer vision, object detection plays a crucial role in remote sensing image interpretation, intelligent surveillance, autonomous driving, and urban planning [1,2,3]. Remote sensing image interpretation specifically requires robust detection algorithms to identify and locate various objects on the Earth’s surface from data acquired by different platforms including drones, aircraft, and satellites. With the rapid development of remote sensing technology, the capability to acquire high-resolution remote sensing images has significantly improved, providing rich data support for the identification and localization of various targets on the Earth’s surface [4]. However, due to the unique characteristics of remote sensing platforms, such as varying acquisition angles, diverse imaging conditions, and complex ground scenes, traditional single-modality object detection methods often demonstrate limited performance under complex environmental conditions, especially when targets are in low-light conditions, adverse weather, or cluttered backgrounds [5,6].
In practical applications, visible light sensors can capture rich color, texture, and shape information, but they are susceptible to the “same object but different spectrum” phenomenon, exhibiting unstable performance, especially under varying lighting conditions [7]. This limitation is particularly evident in remote sensing imagery where atmospheric conditions and diurnal changes can significantly impact image quality. In contrast, infrared sensors are more sensitive to temperature and radiation, performing well in low-light environments, but their low resolution and indistinct edge features make fine-grained target representation difficult [8]. Despite significant advancements in deep convolutional neural networks (CNNs), detection technologies utilizing only a single data source still face enormous challenges in increasingly complex environments [9].
Multispectral fusion object detection provides an effective solution for all-weather, all-time target detection by integrating complementary information from different sensors. As illustrated in Figure 1, the complementary nature of visible and infrared modalities is particularly evident across different lighting conditions—visible images excel in daylight scenarios with rich texture and color information, while infrared images provide superior target detection capabilities in low-light and nighttime conditions through thermal signatures. However, multispectral object detection faces several critical challenges that limit its practical deployment:
First, the heterogeneous nature of visible and infrared modalities poses significant difficulties in feature alignment and fusion. Visible images provide rich texture and color information but are severely affected by illumination variations, while infrared images capture thermal radiation patterns that remain stable across lighting conditions but lack fine-grained details. This fundamental difference in imaging principles leads to substantial semantic gaps between modalities, making direct fusion problematic.
Second, existing CNN-based fusion methods are primarily limited to simple element-wise addition, multiplication, and feature concatenation operations [10,11,12,13]. These naïve fusion strategies fail to model the complex complementary relationships between modalities, resulting in suboptimal feature representations. While these strategies improve single-modality detection performance to some extent, they fail to adequately consider deep interactions and correlations between modalities. The lack of explicit cross-modal interaction mechanisms prevents the network from fully exploiting the synergistic potential of multi-modal data, resulting in poor adaptability [14].
Third, the computational burden of processing dual-stream features poses challenges for real-time applications, particularly in remote sensing scenarios where large-scale imagery must be processed efficiently. Most existing methods simply double the network parameters by duplicating the backbone for each modality, leading to significant computational overhead without proportional performance gains. This inefficiency becomes critical when deploying models on resource-constrained platforms such as drones or satellites.
Fourth, target perception in multispectral images is complicated by varying object scales, occlusions, and cluttered backgrounds common in remote sensing scenarios. Small objects that occupy only a fraction of the image pixels are particularly challenging as their features can be easily overwhelmed by background noise during the fusion process. The problem is exacerbated in remote sensing applications where objects exhibit extreme scale variations and complex terrain textures interfere with detection.
To address these fundamental challenges, this paper proposes the Spectral Dual-stream Recursive Fusion Perception Target Network (SDRFPT-Net), a novel multispectral object detection architecture. Our key insight is that effective multispectral fusion requires: (1) modality-specific feature extraction that preserves unique spectral characteristics, (2) recursive cross-modal interaction that progressively refines feature representations without excessive computational cost, and (3) explicit target enhancement mechanisms that suppress background interference while highlighting object regions. Unlike existing methods, SDRFPT-Net innovatively proposes a Spectral Hierarchical Perception Architecture (SHPA) based on YOLOv10, providing a solid foundation for multi-modal feature extraction, and achieves deep feature interaction and efficient fusion through the Spectral Recursive Fusion Module (SRFM), finally using the Spectral Target Perception Enhancement Module (STPEM) to enhance target region representation and suppress background interference.
Compared to existing research, the main contributions of this paper are as follows:
(1)
We propose a spectral dual-stream separated architecture (SHPA) developed based on YOLOv10, with independently parameterized feature extraction paths for visible and infrared modalities, effectively preserving modality-specific information while adapting to the unique characteristics of each spectral domain;
(2)
We develop a novel spectral recursive fusion module (SRFM) that combines hybrid attention mechanisms with parameter-sharing recursive processing, achieving deep feature interaction while maintaining computational efficiency through cyclic weight reuse;
(3)
We design a spectral target perception enhancement module (STPEM) that adaptively enhances target region representation and suppresses background interference through lightweight mask prediction and similarity-based feature weighting;
(4)
We conduct extensive experiments on three benchmark datasets (VEDAI, FLIR-aligned, and LLVIP), demonstrating that our SDRFPT-Net significantly outperforms state-of-the-art methods in multispectral object detection across various environmental conditions and application scenarios.
The remainder of this paper is organized as follows: Section 2 reviews related work; Section 3 introduces the architecture and key modules of SDRFPT-Net in detail; Section 4 presents the experimental results and analysis; Section 5 concludes the paper and indicates future research directions.

2. Materials and Methods

This section provides a comprehensive review of multispectral object detection, feature fusion strategies, and applications of YOLO series algorithms in multispectral object detection, with a particular focus on remote sensing implementations.

2.1. Multispectral Object Detection

Multispectral object detection technology addresses the limitations of single-modality imaging by fusing complementary information from different spectral bands, enabling all-weather monitoring capabilities [6,15]. In remote sensing applications, this technology effectively overcomes illumination variations, adverse weather conditions, and complex background interference [4]. Early fusion methods relied on traditional mathematical models such as multi-scale transformation [16], sparse representation [17], and saliency-based approaches [18], which were constrained by manually designed feature extractors and predefined fusion rules.
The advent of deep learning has revolutionized multispectral object detection. Significant advances include Liu et al.’s [11] multispectral neural network for improved correlation learning between modalities, Wagner et al.’s [12] deep fusion CNN for visible–infrared image integration, and König et al.’s [13] fully convolutional region proposal networks. These approaches typically employ dual-stream architectures that process different modalities separately before feature fusion at various network levels.
Remote sensing applications present unique challenges including variable object sizes, diverse viewing angles, and unstable imaging conditions [1,4]. To address these issues, researchers have developed specialized solutions combining optical and SAR imagery. Notable contributions include Pang et al.’s [9] RTV-SIFT method for robust cross-modal image registration and Fang et al.’s [7] cross-modal attentive feature fusion technique, which adaptively weights different modality features to enhance detection performance in complex environments.
Recent research has demonstrated the significant potential of Transformer architectures in this domain. Qing et al.’s [19] cross-modal fusion Transformer effectively captures long-range dependencies between modalities, marking an important advancement in attention-based methods for multispectral object detection in remote sensing applications.

2.2. Feature Fusion Strategies

Feature fusion strategies, as the core of multispectral object detection, directly impact final detection performance and are particularly critical in remote sensing image analysis. Based on the stage where fusion occurs, existing methods can be categorized into early fusion, middle fusion, and late fusion [20]. Early fusion directly merges original inputs at the pixel level, offering high computational efficiency but potentially losing modality-specific information; middle fusion occurs after feature extraction, preserving more modal features, also known as feature-level fusion; late fusion integrates outputs from different modalities after detection results are generated [21]. In remote sensing applications, selecting appropriate fusion strategies requires consideration of the characteristics of different sensor data and specific application requirements.
Traditional fusion strategies include weighted averaging, maximum/minimum value selection, and principal component analysis [22]. However, these fixed rules struggle to adapt to complex and variable terrain scenes and imaging conditions in remote sensing images. Recently, deep learning-based adaptive fusion strategies have gained widespread attention in remote sensing. Li et al. [23] proposed a multi-granularity attention network that improved fusion effects of infrared and visible images by learning feature correlations at different levels. Wang et al. [24] developed the Res2Fusion architecture, using multi-receptive field aggregation blocks to generate multi-level features and designing non-local attention models for effective fusion, which demonstrates good adaptability to multi-scale characteristics of objects in remote sensing images.
Recent advances in feature fusion have explored sophisticated mechanisms for multi-modal integration. Xie et al. [25] proposed an oriented object detection framework that leverages contextual dependence mining and penalty-incentive allocation to improve detection performance in aerial images. Their contextual dependence mining network constructs multiple features containing contexts across various ranges with low computational burden, demonstrating that modeling contextual relationships between objects and their surroundings can significantly enhance detection accuracy, particularly for densely packed targets. This work highlights the importance of context-aware fusion strategies in remote sensing applications.
Interactive alignment and mutual-assistance learning approaches have emerged as promising directions for cross-modal object detection. Zhang et al. [26] proposed a mutual-assistance learning framework that enables different modalities to guide each other’s feature learning through progressive alignment mechanisms, demonstrating significant improvements in thermal–visible pedestrian detection. Liu et al. [11] developed an interactive feature alignment method that iteratively refines cross-modal representations by establishing correspondence between modality-specific features, achieving robust performance across varying environmental conditions. These approaches highlight the importance of bidirectional information exchange and progressive refinement in multi-modal fusion.
Cross-modal attention mechanisms provide new perspectives for feature fusion in remote sensing imagery. Zhang et al. [27] proposed a cross-stream and cross-scale adaptive fusion network that improved detection performance of objects in multi-modal images by establishing connections between different modules and scales. Li et al. [28] designed an attention-based generative adversarial network that achieved efficient fusion of infrared and visible images through adversarial training. In the remote sensing domain, Zhao et al. [29] developed an attention receptive pyramid network specifically for ship detection in SAR images, significantly improving detection accuracy by suppressing background interference. These attention-based methods can adaptively emphasize key information in different modalities, suppress background clutter and noise common in remote sensing images, and achieve more precise feature fusion, particularly suitable for object detection tasks in complex terrain scenes. Recent advances have also explored cross-modal attention mechanisms and cosine similarity-based channel resampling modules for multispectral fusion. While these approaches show promise, they often lack the deep iterative refinement capability that our recursive fusion strategy provides, which progressively enhances feature representations through multiple interaction rounds.

2.3. YOLO Series in Multispectral Object Detection

YOLO (You Only Look Once) series algorithms have achieved remarkable success in object detection with their efficient single-stage detection framework [30]. From YOLOv1 [31] to YOLOv10 [32], YOLOv11 [33], and YOLOv12 [34], this series of algorithms has continuously evolved, improving detection accuracy while maintaining efficient inference speed. In remote sensing image analysis, YOLO has attracted significant attention due to its real-time performance and high accuracy characteristics. Chang et al. [35] developed a ship detection method for SAR images based on YOLOv2, while Van Etten [36] proposed the YOLT framework specifically designed for satellite imagery, achieving rapid object detection in large-scale remote sensing images.
In multispectral object detection for remote sensing, YOLO applications primarily utilize dual-stream network structures, processing different modal inputs through parallel backbone networks. Sharma et al. [37] proposed YOLOrs for object detection in multi-modal remote sensing imagery, improving detection stability under different imaging conditions through mid-level feature fusion. For SAR ship detection, Ren et al. [6] developed YOLO-Lite, which reduced parameters from 47.1M to 7.64M while increasing frame rate to 103.5 FPS through feature enhancement networks and attention mechanisms, without sacrificing detection accuracy. Chen [38] and Guo et al. [39] incorporated attention mechanisms into YOLO frameworks, improving ship detection performance in optical and SAR images, respectively.
Current research trends indicate promising directions in combining Transformer with YOLO frameworks. Wang et al. [40] proposed SwinFuse, applying residual Swin Transformer fusion networks to multi-modal image fusion, combining CNN’s local feature extraction capability with Transformer’s global modeling capability. This hybrid architecture provides new technical approaches for multispectral object detection in remote sensing, particularly suitable for processing high-resolution remote sensing data. For small object detection in aerial imagery such as vehicles, Razakarivony et al. [41] developed the VEDAI dataset, providing a standard platform for evaluating different detection algorithms in remote sensing applications.

3. Methodology

This section presents the SDRFPT-Net algorithm, which addresses the fundamental challenges of multispectral object detection through three synergistic modules. Each module is specifically designed to tackle distinct aspects of the multispectral fusion problem:
Spectral Hierarchical Perception Architecture (SHPA) addresses the challenge of heterogeneous feature extraction from different spectral domains. Unlike conventional approaches that share parameters across modalities, SHPA employs independently parameterized dual-stream networks to preserve modality-specific characteristics while extracting hierarchical features at multiple scales. This design ensures that the unique properties of visible (texture, color) and infrared (thermal patterns) information are maintained throughout the feature extraction process.
Spectral Recursive Fusion Module (SRFM) tackles the limitation of shallow fusion strategies by implementing a parameter-efficient recursive mechanism. Instead of simple concatenation or addition, SRFM performs multiple rounds of cross-modal interaction through hybrid attention mechanisms, progressively refining the fused features. The recursive design with parameter sharing addresses the computational efficiency challenge while achieving deep feature integration.
Spectral Target Perception Enhancement Module (STPEM) specifically addresses the challenge of small object detection and background suppression in complex scenes. By generating target-aware masks and applying similarity-based feature weighting, STPEM enhances the representation of potential object regions while suppressing irrelevant background features.
These three modules work collaboratively within a unified framework: SHPA provides rich multi-scale features from both modalities → SRFM performs deep recursive fusion to create comprehensive multi-modal representations → STPEM enhances target regions in the fused features → Finally, the enhanced features are decoded by the detection head for accurate object localization. The overall architecture is illustrated in Figure 2.
The system’s data processing flow is as follows: First, the input visible and infrared images are processed separately through dual-stream feature extraction networks, generating feature maps of different scales. Then, these feature maps undergo deep interaction and fusion through the spectral self-adaptive recursive fusion module. Next, the fused features are further enhanced by the self-adaptive target perception enhancement module to strengthen the representation of the target regions. Finally, the enhanced multi-scale features are aggregated through feature aggregation and input to the detection head, generating the final detection results.
Compared to traditional single-modality object detection methods, this architecture can more effectively utilize the complementary information of RGB and infrared images, especially showing greater detection accuracy and robustness in challenging scenarios such as low light, adverse weather, and complex backgrounds.

3.1. Spectral Hierarchical Perception Architecture (SHPA)

The SHPA architecture, as the core design of this algorithm, effectively processes visible and infrared spectral domain information through a dual-stream structure, laying the foundation for hierarchical perception and fusion of multi-scale features. This architecture is based on YOLOv10’s excellent features and has been systematically improved for multi-modal perception.

3.1.1. Dual-Stream Separated Spectral Architecture Design

Compared to YOLOv10’s single backbone network feature extraction mechanism, the dual-stream separated spectral architecture proposed in this paper can effectively process RGB-IR dual-modal data’s heterogeneous properties, as shown in Figure 3.
This architecture expands a single feature extraction network into a dual-stream network, processing visible spectral and infrared spectral information separately. The two feature extraction streams share similar network structures but use independent parameters, and the feature extraction process of the dual-stream network can be formalized as follows:
F r g b = F r g b ( I r g b ; θ r g b )
F i r = F i r ( I i r ; θ i r )
where F r g b and F i r represent the extracted feature maps from RGB and infrared modalities, respectively; I r g b and I i r represent RGB and infrared input images, F r g b and F i r represent the corresponding feature extraction functions; and θ r g b and θ i r represent their respective network parameters.
The main advantages of the dual-stream architecture are as follows:
(1)
It can design specific extraction strategies for the characteristics of different spectral domains, thereby better adapting to the characteristics of data from each modality;
(2)
It preserves the unique information of each spectral domain, avoiding the potential loss of information that might occur when processing in a single network;
(3)
It captures the feature distributions of different spectral domains through independent parameters, improving the diversity of feature representations.
Compared to YOLOv10’s single feature extraction path, the dual-stream architecture shows greater robustness in complex environments, especially when the quality of information from one modality decreases (such as insufficient RGB information at night or reduced infrared contrast during the day); the system can still maintain detection performance by relying on stable information provided by the other modality.

3.1.2. Multi-Scale Spectral Feature Expansion

To comprehensively capture the multi-scale representation of targets, this paper designs a multi-scale spectral feature expansion mechanism. In each spectral stream, features form a multi-scale feature pyramid through progressive downsampling. For each spectral domain s { r g b , i r } , the feature expansion process can be represented as follows:
F i s = H i ( F i 1 s ; θ i s ) , i { 1 , 2 , 3 , 4 }
where F 0 s represents the input initial feature map, F i s represents the level i feature, H i represents the downsampling function, and θ i s is the corresponding parameter. Specifically, the spatial resolution and channel number of each level feature are as follows:
F 1 s B × 128 × H 4 × W 4 ( P 2 / 4 )
F 2 s B × 256 × H 8 × W 8 ( P 3 / 8 )
F 3 s B × 512 × H 16 × W 16 ( P 4 / 16 )
F 4 s B × 1024 × H 32 × W 32 ( P 5 / 32 )
where B represents the batch size, W represents the feature map width, H represents the feature map height, and P2/4–P5/32 represent feature layers at different scales, with numbers indicating the downsampling factor relative to the input image.

3.1.3. Feature Aggregation and Detection

After multi-scale expansion, images go through the Pre-Pro, SRFM, Post-Pro, and STPEM models for fusion, thereby obtaining high-quality multi-scale fusion features. These features need to be further aggregated and processed to generate the final object detection results, as shown in Figure 4. The feature aggregation process follows the mathematical formulations described in Equations (8) and (9), which define the FPN and PAN operations, respectively.
First, multi-scale fusion features are aggregated through the feature pyramid (FPN) and path aggregation network (PAN), enhancing information exchange between features of different scales, as illustrated in Figure 4 and formulated in Equations (8) and (9):
P i = F P N i ( F f u s e d )
M i = P A N i ( P i ) , if   i = 3 P A N i ( M i 1 , P i ) , if   i > 3  
where P i represents the FPN output of level i feature, M i represents the PAN output of level i feature, and F f u s e d represents the feature after fusion, containing complementary information from RGB and IR.
As shown in Figure 4, the FPN pathway (Equation (8)) processes the fused features Ffused F f u s e d through top-down connections, while the PAN pathway (Equation (9)) performs bottom-up aggregation starting from the P3 level. The visual representation in Figure 4 directly corresponds to these mathematical operations, where the arrows indicate the information flow directions defined by the equations.
FPN transmits semantic information from high levels to low levels, while PAN transmits spatial details from low levels to high levels, forming a powerful feature representation. This bidirectional feature flow mechanism ensures that features at each scale can incorporate both rich semantic information and fine spatial details.
Finally, the aggregated features pass through the v10Detect detection head for object detection:
D = Detect ( M 3 , M 4 , M 5 )  
where D e t e c t ( · ) represents the detection function, with outputs including object class, bounding box coordinates, and confidence information. v10Detect adopts a more efficient feature decoding method, including dynamic convolution and branch specialization design, further improving detection accuracy and efficiency. v10Detect employs branch specialization design, designating specialized branches for bounding box regression, feature processing, and classification tasks, further improving detection accuracy and efficiency.
Compared to YOLOv10, our feature aggregation and detection stage utilizes the advantages brought by modal fusion and target perception enhancement, allowing the detection head to perform object detection based on richer and more accurate feature representations. This is particularly important in low light, adverse weather, and complex background conditions, as single-modal information is often unreliable in these scenarios.

3.2. Spectral Recursive Fusion Module (SRFM)

The SRFM module achieves deep interaction and optimized integration of RGB-IR dual-modal features through innovative fusion mechanisms, significantly improving detection performance in complex environments. Unlike traditional fusion methods, SRFM combines hybrid attention mechanisms with recursive progressive fusion strategies organically, achieving deep multi-modal feature interaction while maintaining parameter efficiency, providing powerful feature representation capabilities for multispectral object detection.
As shown in Figure 5, SRFM receives dual-stream features from SHPA and outputs fused enhanced features after cyclic progressive fusion. This section will introduce the design principles, key components and workflow of the mechanism in detail.

3.2.1. Hybrid Attention Mechanism

As shown in Figure 6, the hybrid attention mechanism builds a comprehensive feature enhancement system by integrating three complementary mechanisms, including self-attention, cross-modal attention, and channel attention, capturing complex feature dependencies from spatial, modal relationship, and channel importance dimensions. This multi-dimensional feature enhancement design significantly improves the model’s processing capability for different scenes.
According to the data flow shown in the figure, the overall calculation process of the hybrid attention mechanism can be expressed as Equations (11) and (12), where the CrossAtt functions represent the cross-modal attention connections visible in Figure 6:
F o u t R G B = S e l f A t t ( F c h a n R G B ) + C r o s s A t t ( F c h a n R G B , F c h a n I R )
F o u t I R = S e l f A t t ( F c h a n I R ) + C r o s s A t t ( F c h a n I R , F c h a n R G B )
where F c h a n R G B = C h a n A t t ( F i n R G B ) and F c h a n I R = C h a n A t t ( F i n I R ) represent RGB and infrared features, respectively, after channel attention processing. The cross-modal connections in Figure 6 visually demonstrate how these features interact through the CrossAtt functions to achieve inter-modal information exchange.
As illustrated in Figure 6, the CrossAtt functions in Equations (11) and (12) are implemented through the cross-modal attention pathways shown in the lower detailed diagram. Specifically, F c h a n R G B serves as both input to its own self-attention and as cross-modal input to the IR attention branch, while F c h a n I R similarly participates in both self-attention and cross-modal attention with the RGB branch. The bidirectional arrows in the Cross-modal Attention section of Figure 6 represent these cross-modal interactions, where RGB features (blue path) cross-attend to IR features (orange path) and vice versa, corresponding to the C r o s s A t t ( F c h a n R G B , F c h a n I R )   and C r o s s A t t ( F c h a n I R , F c h a n R G B )   terms, respectively, in the equations.
Channel attention mechanism. The channel attention sub-module learns channel dependencies through global information modeling, providing all-around enhanced features for the spectral recursive progressive fusion strategy. Given input feature map F B × C × H × W , where B , C , H , W represent batch size, channel number, height, and width, respectively, the calculation process of channel attention can be expressed as follows:
F c h a n = C h a n A t t ( F ) = F · σ ( W 2 · R e L U ( W 1 · F a v g ) + W 2 · R e L U ( W 1 · F m a x ) )
where F a v g = A v g P o o l ( F ) B × C × 1 × 1 and F m a x = M a x P o o l ( F ) B × C × 1 × 1 represent global average pooling and global maximum pooling operations; and W 1 C r × C and W 2 C × C r are shared-weight fully connected layer parameter matrices, where r is the reduction rate. The computation process involves two parallel pathways: (1) W 1 · F a v g performs matrix multiplication between the first fully connected layer weights and the average-pooled features, producing intermediate representations of size B × C / r × 1 × 1 . (2) W 1 · F max similarly computes the product between W 1 and the max-pooled features. Both products are then activated by ReLU functions and further processed by W 2 to restore the original channel dimension. The σ (sigma) function represents the sigmoid activation that maps the final combined features to the (0,1) interval, generating channel-wise attention weights. Finally, the computed channel attention weights are multiplied with the input feature map on a channel-by-channel basis, where each feature channel is scaled by its corresponding attention weight value, effectively amplifying important channels while suppressing less relevant ones through this selective weighting mechanism.
Self-attention mechanism. The self-attention mechanism focuses on capturing spatial dependencies within a modality, allowing features to attend to related regions within the same modality, providing richer contextual information for the spectral hierarchical perception architecture. For input feature F c h a n , the calculation process of self-attention can be expressed as follows:
F s e l f = S e l f A t t ( F c h a n ) = S o f t m a x Q · K T d k · V
where Q = W Q · F , K = W K · F , and V = W V · F are query, key, and value matrices obtained through learnable parameter matrices W Q , W K , and W V ; and d k is the feature dimension, serving as a scaling factor to avoid gradient vanishing problems.
Cross-modal attention mechanism. The cross-modal attention mechanism is used to capture complementary information between different modalities, establishing connections between visible and infrared features, and is the core component for achieving spectral information exchange. The unique aspect of cross-modal attention is that it uses the query from one modality to interact with the keys and values from another modality, thereby enabling information flow between modalities. For RGB and IR features, the calculation of cross-modal attention can be expressed as follows:
F c r o s s R G B = C r o s s A t t ( F c h a n R G B , F c h a n I R ) = α · S o f t m a x Q R G B · ( K I R ) T d k · V I R
F c r o s s I R = C r o s s A t t ( F c h a n I R , F c h a n R G B ) = α · S o f t m a x Q I R · ( K R G B ) T d k · V R G B
where Q R G B = W Q · F c h a n R G B , K I R = W K · F c h a n I R , and V I R = W V · F c h a n I R are the cross-modal attention calculation from RGB to IR; Q I R = W Q · F c h a n I R , K R G B = W K · F c h a n R G B , and V R G B = W V · F c h a n R G B are the matrices for IR to RGB calculation; and α is a learnable scaling factor that controls the strength of cross-modal information fusion.
Finally, the outputs of self-attention and cross-modal attention are added to obtain the final enhanced features:
F o u t R G B = F s e l f R G B + F c r o s s R G B
F o u t I R = F s e l f I R + F c r o s s I R
Through this design, the hybrid attention mechanism can simultaneously attend to channel importance, spatial dependency relationships, and modal complementary information, building a more comprehensive and robust feature representation.

3.2.2. Recursive Progressive Fusion Strategy

This section presents the recursive progressive fusion strategy, which addresses the computational efficiency challenge in deep multi-modal fusion through three core mechanisms. First, the parameter cycling reuse structure enables deep feature interaction while maintaining computational efficiency by sharing the same parameter set across multiple refinement rounds. Second, the spectral feature progressive fusion operates in the spectral dimension to preserve and enhance modality-specific characteristics through normalized processing, hybrid attention calculation, and residual connections. Third, the multi-scale fusion mechanism applies recursive processing across different feature scales (P3/8, P4/16, P5/32) to achieve comprehensive feature optimization. This strategy achieves a “feature distillation” effect through three progressive refinement stages: initial fusion, feature reinforcement, and final refinement.
Multi-modal feature fusion is a key challenge in RGB-T object detection. Traditional multi-modal feature fusion methods typically enhance performance by stacking multiple Transformer blocks, but this approach leads to dramatic increases in parameter count and computational complexity. Inspired by the “review-consolidate” mechanism in human learning processes, this paper proposes a spectral hierarchical recursive progressive fusion strategy, achieving feature progressive refinement through repeatedly applying the same feature transformation operations, thereby enhancing fusion effects without increasing model parameters.
Parameter Cycling Reuse Structure. As shown in Figure 7, The core idea of the spectral hierarchical recursive progressive fusion strategy is to use the same set of parameters for multiple rounds of feature refinement. Each refinement builds on the results of the previous round, forming a continuous, progressive feature fusion process. This process can be expressed as follows:
[ F R G B t + 1 , F I R t + 1 ] = T ( F R G B t , F I R t ; θ )
where F R G B t and F I R t represent the visible and infrared features, respectively, after the t rounds of cycling, T represents the feature transformation function, and θ is the reused model parameter.
Through multiple cycles, the feature representation ability is continuously enhanced:
[ F R G B f i n a l , F I R f i n a l ] = T n ( F R G B 0 , F I R 0 ; θ )
where T n represents applying the transformation function T continuously n times, and F R G B 0 and F I R 0 are the initial features.
Compared to traditional methods, the cyclic weight reuse structure significantly reduces the model parameter count, while achieving deep feature interaction through multiple refinements. This design not only improves the model’s representation ability but also alleviates the risk of overfitting.
Spectral Feature Progressive Fusion. Spectral feature progressive fusion is the core characteristic of this strategy, progressively fusing different spectral domain features. This progressive fusion process operates in the spectral dimension, ensuring each spectral property is fully preserved and mutually enhanced. The fusion process includes the following key steps:
  • Spectral feature normalization: Normalization is performed separately on visible and infrared features, expressed as follows.
    F ^ R G B = LN ( F R G B )
    F ^ I R = LN ( F I R )
  • Hybrid attention calculation: The hybrid attention mechanism is applied to process normalized features, expressed as follows, where H y b r i d A t t e n t i o n ( · ) represents the hybrid attention calculation.
    [ F R G B , F I R ] = HybridAttention ( F ^ R G B , F ^ I R ; θ a t t n )
  • Spectral residual connection: The attention outputs are combined with original spectral features, expressed as follows.
    F R G B = F R G B + F R G B
    F I R = F I R + F I R
  • Spectral feature enhancement: Each spectral feature is further enhanced through multi-layer perceptron and residual connection, expressed as follows.
    F R G B = F R G B + MLP ( LN ( F R G B ) )
    F I R = F I R + MLP ( LN ( F I R ) )
    where L N ( · ) represents layer normalization operation, and M L P ( · ) represents multi-layer perceptron.
Progressive feature refinement process. The progressive feature refinement process can be viewed as a “feature distillation” mechanism, where each round of cycling makes the feature representation purer and more effective. In this research, we adopt a fixed three-round cycling structure, a design based on extensive experimental validation.
The refinement process can be divided into three stages:
  • First round of cycling: Initial fusion stage. Mainly captures basic intra-modal and inter-modal relationships, establishing initial feature interaction;
  • Second round of cycling: Feature reinforcement stage. Based on the already established initial relationships, this stage further strengthens important feature connections, suppressing noise and irrelevant information;
  • Third round of cycling: Feature refinement stage. Performs final optimization and fine-tuning on features, forming high-quality fusion representations.
This three-round progressive refinement process can be expressed as follows:
[ F R G B 3 , F I R 3 ] = T 3 ( F R G B 0 , F I R 0 ; θ )
The progressive refinement mechanism creates a “deep cascade” effect, achieving deep network feature representation capabilities within a fixed parameter space, which is fundamentally different from traditional “multi-layer stacking” approaches. Traditional methods require introducing new parameter sets for each additional layer, while our method achieves deeper effective network depth through parameter reuse while maintaining parameter efficiency.
Spectral Multi-scale Fusion Mechanism. The spectral multi-scale fusion mechanism is an important component of the recursive progressive fusion strategy, applying recursive progressive fusion on features of different scales to achieve comprehensive multi-scale feature optimization. This mechanism includes the following key designs:
  • Multi-scale feature selection: The fusion strategy is applied separately on three scales—P3/8, P4/16, and P5/32—ensuring thorough fusion of features at all three scales;
  • Inter-scale information flow: Information exchange between features of different scales is achieved through FPN and PAN structures;
The multi-scale fusion process can be expressed as follows:
F s f u s i o n = T s N ( F s i n i t ; θ s ) , s { P 3 / 8 , P 4 / 16 , P 5 / 32 }
where s represents the feature scale index, and T s represents the feature transformation function for the s-th scale. By applying recursive progressive fusion across multiple scales, the system comprehensively enhances the representation capability of features at different scales, providing a solid foundation for detecting targets of various sizes.
In summary, the recursive progressive fusion strategy achieves three key objectives through its innovative design: (1) parameter efficiency—achieving deep feature interaction through cyclic weight reuse rather than parameter multiplication; (2) progressive refinement—implementing a three-stage “feature distillation” process that gradually purifies and enhances multi-modal representations; and (3) multi-scale optimization—applying the recursive mechanism across different feature scales (P3/8, P4/16, P5/32) to ensure comprehensive feature enhancement. This strategy represents a paradigm shift from traditional “multi-layer stacking” approaches to a more efficient “iterative deepening” methodology, providing superior feature representation capabilities while maintaining computational efficiency for multispectral object detection tasks.

3.3. Spectral Target Perception Enhancement Module (STPEM)

As shown in Figure 8, the STPEM module focuses on enhancing target regions in features while reducing background interference. Through mask generation and feature enhancement mechanisms, this module significantly improves the model’s detection capability for small and low-contrast targets, providing more precise feature representation for object detection in complex environments.

3.3.1. Lightweight Mask Prediction

Lightweight mask prediction is the core component of STPEM. Given input feature F B × C × H × W , mask prediction first predicts target region masks through a lightweight convolutional network:
M = σ ( M p r e d ( F ) )  
where M p r e d represents the mask prediction network, and σ represents the sigmoid activation function. The mask prediction network adopts a two-layer convolutional structure:
M p r e d ( F ) = Conv 1 × 1 ( ReLU ( BN ( Conv 3 × 3 ( F ) ) ) )  
The first layer is a 3 × 3 convolution that reduces the number of channels from C to C/2, followed by batch normalization and ReLU activation; the second layer is a 1 × 1 convolution that reduces the number of channels from C/2 to 1, outputting a single-channel mask. Finally, the sigmoid function maps values to the [0, 1] range, representing the probability that each position contains a target.
The mask prediction network is essentially learning “what feature patterns might correspond to target regions.” For example, in RGB images, targets typically have distinct edges and texture features; in infrared images, targets often appear as regions with significant temperature differences from the background. The mask prediction network captures these feature patterns through convolutional operations to generate masks representing potential target regions.

3.3.2. Similarity Calculation and Adjustment

After mask generation, the module calculates the cosine similarity between features and masks to evaluate the correlation between each feature channel and the target region, thereby establishing explicit associations between feature channels and potential target regions:
F f l a t = Flatten ( F ) B × C × ( H × W )  
M f l a t = Flatten ( M ) B × 1 × ( H × W )  
M e x p a n d e d = Expand ( M f l a t , C ) B × C × ( H × W )  
S = CosineSimilarity ( F f l a t , M e x p a n d e d , dim = 2 ) B × C  
where the F l a t t e n ( · ) operation flattens the spatial dimensions of the features, the E x p a n d ( · ) operation expands the mask to the same number of channels as the features, and Cos ineSimilarity ( · ) calculates the cosine similarity between two vectors.
After calculating the similarity between each channel and the mask, further processing is performed through averaging operations and a learnable adjustment layer:
S a v g = Mean ( S , dim = 1 , keepdim = T r u e ) B × 1  
S a d j u s t e d = σ ( S a d j u s t ( S a v g ) ) B × 1 × 1 × 1  
where S a d j u s t is a 1 × 1 convolutional layer for adjusting similarity, and σ is the sigmoid activation function. This learnable similarity adjustment mechanism enables the module to adaptively adjust similarity calculations according to different scenes, improving the flexibility and adaptability of the module.

3.3.3. Feature Enhancement Mechanism

Finally, the enhanced feature F e n h a n c e d is achieved through similarity weighting:
F e n h a n c e d = F × S a d j u s t e d
The core idea of this weighting mechanism is as follows: If a feature has high similarity with the predicted target region, it is preserved or enhanced. If the similarity is low, the feature is suppressed. In this way, features of target regions are effectively enhanced, while features of background regions are suppressed, thereby improving the signal-to-noise ratio of the features.
The STPEM module significantly improves the performance of multispectral object detection by effectively identifying and enhancing potential target regions, showing excellent performance especially when processing complex background scenes.

4. Experiments

This section will detail the experimental results of SDRFPT-Net on the VEDAI, FLIR-aligned and LLVIP datasets, verifying the effectiveness of our proposed algorithm. First, we introduce the experimental setup and evaluation datasets; second, we compare SDRFPT-Net with current state-of-the-art multispectral object detection methods; finally, we analyze the contribution of each innovative module through comprehensive ablation experiments.

4.1. Datasets and Evaluation Metrics

4.1.1. Datasets

This study employs three widely used multispectral object detection benchmark datasets: VEDAI [41], FLIR-aligned [42], and LLVIP [43], and Figure 9 shows the object size distribution characteristics in the multispectral detection dataset.
VEDAI. A benchmark dataset for aerial remote sensing vehicle detection with 1210 images (1024 × 1024 pixels) at 12.5–25.0 cm resolution, captured from aircraft platforms. Contains eight vehicle classes across diverse terrain backgrounds (woods, cities, roads) with 3700+ annotated targets. Specifically designed for remote sensing applications, it features complex terrain textures and variable imaging perspectives common in earth observation tasks. It is particularly challenging due to small target sizes (0.7% of image pixels), significant scale variations, and complex ground backgrounds typical in aerial imagery acquisition. The evaluation uses 10-fold cross-validation with mAP and FPPI metrics.
FLIR-aligned. An aligned dataset derived from FLIR ADAS [44] containing 4129 training and 1013 testing image pairs of spatially aligned thermal infrared and visible light images. It features three object classes (person, car, bicycle) captured in diverse environments (urban roads, highways, residential areas) under various conditions (day, night, dusk). The dataset is valuable for evaluating multispectral detection algorithms in real driving scenarios.
LLVIP. A visible-infrared paired dataset for low-light visual tasks with 16,836 image pairs captured at night (6–10 PM) across 26 locations. All pairs are time–space aligned and contain annotated pedestrians. Targets difficult to identify in visible images are clearly visible in infrared. The images are registered via a semi-automatic method ensuring an identical field of view and are processed to a uniform 1080 × 720 resolution.

4.1.2. Metrics

To comprehensively evaluate the performance of object detection models, this study adopts the following standard evaluation metrics:
Precision (P). Precision is a key metric for measuring detection accuracy, and is defined as the ratio of correctly detected targets (true positives) to all detected targets (true positives and false positives). This metric reflects the accuracy of model target recognition, calculated as follows:
P = T P T P + F P
Recall (R). Recall measures the model’s ability to detect all targets, defined as the ratio of correctly detected targets (true positives) to all actually existing targets (true positives and false negatives). This metric reflects the completeness of the model’s capture of all targets in the image, calculated as follows:
R = T P T P + F N
Mean average precision at IoU = 0.50 (mAP50). Mean average precision at IoU = 0.50 (mAP50): mAP50 is the average precision calculated at an Intersection over Union (IoU) threshold of 0.50. This metric primarily measures the model’s performance on “simple” detection tasks, i.e., the detection accuracy when the overlap area between the predicted bounding box and the ground truth bounding box accounts for at least 50% of the total area.
Mean average precision across IoU = 0.50:0.95 (mAP50–95). mAP50-95 is a more comprehensive evaluation metric that calculates the average precision at different IoU thresholds (from 0.50 to 0.95, with a step size of 0.05), and then takes the average of these values. Compared to mAP50, mAP50-95 better reflects the model’s localization accuracy by considering a range of stricter IoU thresholds. A high mAP50-95 score indicates that the model can maintain good performance under stricter localization standards, which is particularly important for applications requiring high localization accuracy.

4.2. Experimental Setup

All experiments in this study were conducted on a server equipped with an NVIDIA RTX 4090 GPU (24 GB memory) and an Intel Core i7-13700 processor (24 cores), with 62 GB system memory. The experimental environment was based on a Linux 20.04 operating system, PyTorch 2.0.1 deep learning framework, CUDA 11.7, cuDNN 8.7.0, and Python 3.9.21.
During the training process, we used the SGD optimizer with a momentum parameter of 0.937 and a weight decay coefficient of 0.0005. The initial learning rate was set to 0.01, and a cosine annealing strategy was adopted to reduce the learning rate to 0.01 times its initial value by the end. The batch size was fixed at 4, input image dimensions were uniformly adjusted to 640 × 640 pixels, and the maximum number of training epochs was 300. An early stopping strategy was also implemented—automatically terminating the training process when there was no performance improvement for 30 consecutive epochs.

4.3. Comparison with State-of-the-Art Methods

This section presents comprehensive experimental validation of SDRFPT-Net across three benchmark datasets, each representing distinct application scenarios and challenges. The evaluation is structured as follows: Section 4.3.1 examines performance on the VEDAI dataset, which focuses on small vehicle detection in aerial remote sensing imagery with complex terrain backgrounds and extreme scale variations. Section 4.3.2 evaluates the FLIR-aligned dataset, representing automotive applications with diverse environmental conditions including day, night, and transitional lighting scenarios. Section 4.3.3 assesses the LLVIP dataset, specifically targeting pedestrian detection in low-light nighttime conditions where infrared modality dominance tests our fusion strategy’s effectiveness. Each evaluation includes quantitative performance metrics, qualitative visual comparisons, and detailed analysis of method-specific advantages and limitations.

4.3.1. Performance on the VEDAI Dataset

Table 1 presents the performance comparison between SDRFPT-Net and current state-of-the-art methods on the VEDAI dataset. The VEDAI dataset serves as a significant benchmark for evaluating small vehicle detection performance in aerial remote sensing imagery and is used to test the effectiveness of multispectral object detection algorithms under complex terrain backgrounds. To provide a more comprehensive evaluation, we have included recent state-of-the-art algorithms for comparison.
The experimental results demonstrate that the proposed SDRFPT-Net achieves excellent performance across all key evaluation metrics. Compared to single-modality baseline methods, SDRFPT-Net realizes significant performance improvements. Specifically, in terms of mAP50, SDRFPT-Net reaches 0.734, showing a 2.5% improvement over the second-best performing SuperYOLO (0.716). Under the more stringent mAP50:95 evaluation criterion, SDRFPT-Net achieves 0.450, outperforming the second-best method CFT (0.427) by 5.4%, indicating that our method not only enhances target detection rates but also maintains high-precision bounding box localization capabilities.
As shown in Figure 10, under complex terrain textures and multi-scale remote sensing imaging conditions, SDRFPT-Net can accurately detect all targets and provide precise bounding box localization. In comparison, YOLOv10-add exhibits notable missed detections for small targets, while CMAFF and CFT demonstrate issues with false detections and imprecise bounding box localization. These visualization results intuitively validate the detection advantages of SDRFPT-Net in complex remote sensing scenarios where spatial resolution, imaging angle, and environmental factors pose significant challenges for object detection tasks.
The superior performance of SDRFPT-Net on VEDAI can be attributed to several key factors. Modality complementarity analysis reveals that, in aerial remote sensing scenarios, visible modalities provide rich texture details for vehicle boundary delineation, while infrared captures thermal signatures that distinguish vehicles from similar-textured backgrounds. The 5.4% improvement in mAP50:95 particularly demonstrates enhanced localization precision, crucial for small vehicle detection where precise boundaries are challenging to define. Scale sensitivity analysis shows that our multi-scale fusion effectively handles the 147× size variation (20–300 pixels) in VEDAI, with P3 features capturing small distant vehicles and P5 features providing semantic context for larger vehicles. However, failure case analysis reveals limitations in dense parking scenarios where thermal signatures blend together, accounting for approximately 8% of missed detections.
Comprehensive analysis indicates that SDRFPT-Net significantly improves vehicle detection performance in aerial remote sensing imagery by effectively integrating complementary information from visible and infrared modalities, particularly excelling in detecting small-sized targets and targets in complex terrain backgrounds. This capability is crucial for remote sensing applications where targets often occupy only a fraction of the pixel space and must be distinguished from heterogeneous ground textures and shadows. This outstanding performance can be attributed to the synergistic effect of three innovative modules: the Spectral Hierarchical Perception Architecture (SHPA), Spectral Recursive Fusion Module (SRFM), and Spectral Target Perception Enhancement Module (STPEM). These modules collectively enhance the network’s ability to process multi-source remote sensing data and extract discriminative features across different spectral domains, addressing the unique challenges of earth observation imagery

4.3.2. Performance on the FLIR-Aligned Dataset

Table 2 shows the comparison results of SDRFPT-Net with other state-of-the-art methods on the FLIR-aligned dataset. This dataset is widely used as a benchmark for evaluating the performance of multispectral object detection systems under various environmental conditions. To ensure comprehensive evaluation against current advances in object detection, we have included additional state-of-the-art methods in our comparison.
The experimental results show that the proposed SDRFPT-Net outperforms existing methods in all key metrics, including precision, recall, and mAP. Compared to the best-performing single-modality method YOLOv10-infrared (with mAP50 of 0.727), SDRFPT-Net’s mAP50 showed an 8.0% improvement (from 0.727 to 0.785). This demonstrates that our proposed multi-modal fusion strategy can effectively integrate complementary information from different spectral domains.
Compared to other multispectral fusion methods, SDRFPT-Net achieves precision (P) and recall (R) of 0.854 and 0.700, respectively, significantly outperforming other methods. Particularly in terms of mAP50, SDRFPT-Net (0.785) improved by 11.5% compared to the second-best performing BA-CAMF Net (0.704), demonstrating the superiority of our proposed spectral dual-stream recursive fusion perception architecture in multispectral object detection tasks.
Notably, under the more stringent mAP50:95 evaluation criterion, SDRFPT-Net achieves 0.426, comparable to the single-modality baseline YOLOv10-infrared (0.424), while significantly outperforming other multi-modal fusion methods (with the highest being BA-CAMF Net’s 0.351). This indicates that SDRFPT-Net not only improves the target detection rate but also maintains high-precision bounding box localization capability.
The remarkable 11.5% mAP50 improvement on FLIR-aligned demonstrates SDRFPT-Net’s effectiveness in diverse driving scenarios. Cross-modal dependency analysis reveals that, in urban environments, visible modality excels at detecting vehicles through shape and texture cues, while infrared provides robust person detection in challenging lighting conditions. The recursive fusion mechanism proves particularly effective for pedestrian detection, where thermal signatures remain consistent regardless of clothing variations captured in visible images. Computational efficiency analysis shows that despite processing dual modalities, inference speed remains practical at 26.8 FPS. Environmental adaptability analysis indicates superior performance in transitional lighting conditions (dawn/dusk) where neither modality alone provides sufficient information, with our method achieving 12.3% better recall than single-modality approaches in these challenging scenarios.
As shown in Figure 11, in complex lighting and occlusion conditions, SDRFPT-Net (j) can accurately detect all targets with more precise bounding boxes. In contrast, YOLOv10-add (g) has some missed detections on small targets, while TFDet (h) and CMAFF (i) have some false detections and inaccurate bounding box issues. These visualization results intuitively demonstrate the detection advantages of SDRFPT-Net in complex scenes.

4.3.3. Performance on the LLVIP Dataset

Table 3 shows the comparison results of SDRFPT-Net with other state-of-the-art methods on the LLVIP dataset. The LLVIP dataset focuses on pedestrian detection in low-light environments and is an important benchmark for evaluating algorithm robustness in nighttime scenes. To provide a comprehensive evaluation against the latest developments in object detection, we have included additional state-of-the-art methods in our comparison.
As can be seen from Table 3, SDRFPT-Net also achieves excellent performance on the LLVIP dataset. In terms of mAP50, SDRFPT-Net reaches 0.963, showing a slight improvement (0.2%) compared to the closest-performing single-modality method YOLOv10-infrared and the multi-modal method TFDet (both at 0.961). Although this improvement is modest, achieving further improvement at an already near-saturated performance level is still significant. Under the more stringent mAP50:95 evaluation criterion, SDRFPT-Net reaches 0.706, significantly outperforming all comparison methods. Compared to the second-best performing method, YOLOv8-infrared (0.645), it improves by 9.5%, indicating that the proposed method has significant advantages in precise bounding box localization. This result confirms that SDRFPT-Net can not only detect target locations but also more accurately describe target boundaries.
Notably, under the low-light conditions of the LLVIP dataset, the infrared modality alone can achieve high performance (e.g., YOLOv8-infrared achieves an mAP50 of 0.961). In this case, SDRFPT-Net still achieved performance improvements through effective integration of complementary information from visible light, especially with significant improvement in mAP50:95 (from 0.645 to 0.706). This indicates that the proposed spectral recursive fusion mechanism can still effectively extract and integrate valuable features from the visible light modality, even when infrared information is dominant.
SDRFPT-Net’s recall reaches 0.911, higher than all comparison methods, indicating it has stronger target detection capability and can find pedestrian targets that might be missed by other methods, which is particularly important for practical application scenarios.
The significant 9.5% improvement in mAP50:95 on LLVIP validates SDRFPT-Net’s low-light detection capabilities. Modal dominance analysis reveals that, while infrared provides primary detection signals in nighttime scenarios, visible modality contributes crucial spatial context and boundary refinement. The recursive fusion mechanism effectively extracts valuable features from low-quality visible images, with ablation studies showing 4.2% performance degradation when visible modality is excluded, even in predominantly infrared-favorable conditions. Challenging scenario analysis demonstrates the model’s superior performance in complex urban lighting with mixed artificial illumination, where traditional methods struggle with variable visibility conditions. The high recall (0.911) indicates robust detection capability, particularly important for pedestrian safety applications where missed detections have critical consequences.
Figure 12 provides visualized detection results of various methods in typical nighttime scenes from the LLVIP dataset. Qualitative analysis shows that SDRFPT-Net can accurately locate all pedestrian targets in low-contrast environments with high bounding box matching accuracy. In contrast, other multi-modal fusion methods such as YOLOv10-add, TFDet, and CMAFF show varying degrees of detection instability in complex scenes, including missed detections, false detections, or bounding box localization deviations. These visualization results further confirm the advantages of SDRFPT-Net demonstrated in the quantitative evaluation.
Figure 12 shows the visualization detection results of various algorithms on the LLVIP dataset. In typical nighttime low-light scenes, SDRFPT-Net (j) can accurately detect all pedestrians with more precise bounding boxes. In contrast, YOLOv10-add (g), TFDet (h), and CMAFF (i) exhibit missed detections or inaccurate bounding box issues in some complex scenes. These visualization results further confirm the detection advantages of SDRFPT-Net in low-light environments.
Combining the experimental results from both the FLIR-aligned and LLVIP datasets, SDRFPT-Net demonstrates powerful detection capability and robustness under various environmental conditions. This is attributed to the collaborative work of our three innovative modules: The spectral hierarchical perception architecture provides a solid foundation for multi-modal feature extraction, and the spectral adaptive recursive fusion module achieves deep interaction and efficient fusion, while the spectral adaptive target perception enhancement module further improves target region feature representation. The organic combination of these three modules enables SDRFPT-Net to achieve excellent multispectral object detection performance while maintaining low computational complexity.

4.4. Ablation Studies

To verify the effectiveness of each innovative module in SDRFPT-Net, we conducted systematic ablation experiments on the FLIR-aligned dataset, which can more intuitively reflect the effectiveness of our algorithm. These experiments aim to evaluate the contribution of each component to the overall performance of the network and validate the rationality of our proposed design scheme.
To comprehensively demonstrate the training characteristics and convergence behavior of SDRFPT-Net, we first analyze the learning curves of the model on the FLIR-aligned dataset. Figure 13 illustrates the evolution of different evaluation metrics during the training process, which intuitively reflects the stability and effectiveness of our proposed method.
From the precision–confidence curves in Figure 13a, we can observe that as the confidence threshold increases, the precision of each category shows a stable upward trend, with the car category performing best, maintaining precision close to 1.0 in the high confidence interval. The recall–confidence curves in Figure 13b show that recall gradually decreases with increasing confidence, which conforms to the general pattern of object detection, indicating that the model can achieve a good balance between precision and recall at different confidence thresholds.
The precision–recall curve shown in Figure 13c is a key metric for evaluating detection performance. It can be seen that the PR curves of all categories maintain a large area under the curve, with the car category demonstrating excellent detection performance. Although the bicycle category is lower, it still maintains a reasonable performance level, which may be related to the smaller number of samples of this category in the dataset.
The F1–confidence curves in Figure 13d comprehensively reflect the overall performance of the model. The F1 score reaches its peak in the confidence range of 0.2–0.4, indicating that the model achieves the best balance between precision and recall within this confidence range.
These learning curves validate the robustness and effectiveness of SDRFPT-Net in multispectral object detection tasks, providing a performance baseline for subsequent ablation experiments. Next, we will analyze the specific contributions of each innovative module to the overall performance through systematic ablation studies.

4.4.1. Baseline Model Comparison

First, we established a baseline model, then gradually added each core component to evaluate the contribution of each module. Table 4 shows the performance comparison of different component combinations.
From the results in Table 4, it is evident that each component we proposed contributes significantly to detection performance. The model based on the Spectral Hierarchical Perception Architecture (SHPA) (A1) achieves an mAP50 of 0.701. After adding the Spectral Adaptive Recursive Fusion Module (SRFM) (A2), the mAP50 increases to 0.775, a relative improvement of 10.6%. With the further addition of the Spectral Adaptive Target Perception Enhancement Module (STPEM) (A3), mAP50 and mAP50:95 reach 0.785 and 0.426, respectively, with mAP50:95 showing a particularly significant 14.2% relative improvement over A2, indicating that STPEM greatly enhances the model’s localization accuracy under stricter detection standards.
Progressive Module Integration Analysis: The ablation results reveal a clear performance hierarchy, with each module contributing distinct advantages. SHPA alone (A1) establishes a solid foundation with 0.701 mAP50 by providing modality-specific feature extraction, avoiding the information loss common in shared-parameter approaches. The substantial 10.6% relative improvement from SRFM addition (A2) demonstrates the critical importance of deep cross-modal interaction over simple concatenation methods. Feature visualization analysis shows that SRFM enables more coherent feature representations, with enhanced target–background contrast compared to simple addition fusion. STPEM’s contribution (A3) is particularly significant for precise localization, evidenced by the 14.2% improvement in mAP50:95. This improvement primarily stems from enhanced small target detection, where STPEM’s mask prediction mechanism reduces false positives in cluttered backgrounds by 23% compared to the baseline configuration.

4.4.2. Ablation Experiments on Hybrid Attention Mechanism

To evaluate the effectiveness of various attention mechanisms in the hybrid attention mechanism, we designed a series of comparative experiments, with the results shown in Table 5.
The results show that different types of attention mechanisms have varying impacts on model performance. When used individually, the self-attention mechanism (B1) performs best with an mAP50 of 0.776 and mAP50:95 of 0.408, indicating that capturing spatial dependencies within modalities is critical for object detection. Although cross-modal attention (B2) and channel attention (B3) show slightly inferior performance when used alone, they provide feature enhancement capabilities in different dimensions.
In combinations of two attention mechanisms, the combination of self-attention and cross-modal attention (B4) performs best, with mAP50:95 reaching 0.424, approaching the performance of the complete model. The complete combination of three attention mechanisms (B7) achieves the best performance, confirming the rationality of the hybrid attention mechanism design, which can comprehensively capture the complex relationships in multi-modal data.
To gain a deeper understanding of the role of different attention mechanisms in multispectral object detection, we analyzed the visualization results of self-attention, cross-modal attention, and channel attention on the P3 feature layer.
From Figure 14, it can be observed that the feature maps of single attention mechanisms present different attention patterns:
  • Self-attention mechanism (B1): Mainly focuses on target contours and edge information, effectively capturing spatial contextual relationships, with strong response to target boundaries, helping to improve localization accuracy;
  • Cross-modal attention mechanism (B2): Presents overall attention to target areas, integrating complementary information from RGB and infrared modalities, but with relatively weak background suppression capability;
  • Channel attention mechanism (B3): Demonstrates selective enhancement of specific semantic information, highlighting important feature channels, with strong response to specific parts of targets, improving the discriminability of feature representation.
Furthermore, we conducted a comparative analysis of the visualization effects of dual-attention mechanisms versus the full attention mechanism on the P3 feature layer, as shown in Figure 15.
Through the visualization comparative analysis in Figure 15, we observe that the feature maps of dual-attention mechanisms present complex and differentiated feature representations:
  • Self-attention + Cross-modal attention (B4): The feature map simultaneously possesses excellent boundary localization capability and overall target region representation capability. The heatmap shows precise response to target regions with significant background suppression effect. This combination fully leverages the complementary advantages of self-attention in spatial modeling and cross-modal attention in multi-modal fusion, enabling it to reach 0.424 in mAP50:95, approaching the performance of the full attention mechanism.
  • Self-attention + Channel attention (B5): The feature map enhances the representation of specific semantic features while preserving target boundary information. The heatmap shows strong responses to key parts of targets, enabling the model to better distinguish different categories of targets, achieving 0.409 in mAP50:95, and outperforming any single attention mechanism.
  • Cross-modal attention + Channel attention (B6): The feature map enhances specific channel representation based on multi-modal fusion, but lacks the spatial context modeling capability of self-attention. The heatmap shows some response to target regions, but the boundaries are not clear enough, and the background suppression effect is relatively weak, which explains its relatively lower performance.
Although dual-attention mechanisms (especially self-attention + cross-modal attention) can improve feature representation capability to some extent, they cannot completely replace the comprehensive advantages of the full attention mechanism.
As shown in Figure 14d and Figure 15d, the full attention mechanism, through the synergistic effects of three attention mechanisms, shows the most precise and strong response to target regions in the heatmap, with clear boundaries and optimal background suppression effects, achieving an organic unification of spatial context modeling, multi-modal information fusion, and channel feature enhancement, and obtaining optimal performance in multispectral object detection tasks.
Based on the above experiments and visualization analysis, we verified the effectiveness of the proposed hybrid attention mechanism. The results show that, despite the advantages of single- and dual-attention mechanisms, the complete combination of three attention mechanisms achieves optimal performance across all evaluation metrics. This confirms the rationality of our proposed “spatial-modal-channel” multi-dimensional attention framework, which creates an efficient synergistic mechanism through self-attention capturing spatial contextual relationships, cross-modal attention fusing complementary information, and channel attention selectively enhancing key features. This multi-dimensional feature enhancement strategy provides a new feature fusion paradigm for multispectral object detection, offering valuable reference for research in related fields.
Attention Mechanism Synergy Analysis: The hybrid attention mechanism validation reveals complementary benefits of different attention types. Self-attention (B1) captures spatial context most effectively, explaining its superior individual performance (0.776 mAP50) through enhanced intra-modal spatial relationships. Cross-modal attention (B2) and channel attention (B3) provide specialized enhancements for inter-modality fusion and feature channel selection, respectively, though their individual performances are lower due to limited scope. The attention mechanism combination analysis demonstrates progressive improvements: dual combinations (B4–B6) show varying effectiveness depending on the specific pairing, with self-attention + cross-modal attention (B4) achieving the best dual performance (0.774 mAP50) by combining spatial modeling with cross-modal information exchange. The optimal combination (B7) demonstrates synergistic effects beyond simple addition, with the three mechanisms collectively addressing spatial modeling, cross-modal correlation, and feature importance weighting simultaneously. Computational overhead analysis shows that the hybrid attention increases inference time by only 8% while providing substantial accuracy gains and validating the efficiency of our design.

4.4.3. Ablation Experiments on Spectral Hierarchical Recursive Progressive Fusion Strategy

To verify the effectiveness of the spectral fusion strategy, we conducted ablation experiments from two aspects: fusion position selection, and recursive progression iterations.
Impact of introducing fusion positions. Multi-scale feature fusion is a key link in multispectral object detection, and effective fusion of features at different scales has a decisive impact on model performance. This section explores the impact of fusion positions and fusion strategies on detection performance through ablation experiments and visualization analysis.
To systematically study the impact of fusion positions on model performance, we designed a series of ablation experiments, as shown in Table 6. The experiments started from the baseline model (C1, using simple addition fusion at all scales), progressively applied our proposed innovative fusion modules (fusion mechanism combining SRFM and STPEM) at different scales, and finally evaluated the effect of comprehensive application of advanced fusion strategies.
As shown in Table 6, with the increase in application positions of innovative fusion modules, model performance progressively improves. The baseline model (C1) only uses simple addition fusion at all feature scales, with an mAP50 of 0.701. When applying SRFM and STPEM modules at the P3/8 scale (C2), performance significantly improves to an mAP50 of 0.769. With further application of advanced fusion at the P4/16 scale (C3), mAP50 increases to 0.776. Finally, when the complete fusion strategy is applied to all three scales (C4), performance reaches optimal levels with an mAP50 of 0.785 and mAP50:95 of 0.426.
To intuitively understand the effect differences between different fusion strategies, we visualized and compared feature maps of simple addition fusion and advanced fusion strategies (SRFM + STPEM) at three scales: P3/8, P4/16, and P5/32 (as shown in Figure 16).
To gain a deeper understanding of the performance differences between different fusion strategies, we conducted systematic visualization comparisons at different scales (P3/8, P4/16, P5/32).
  • At the P3/8 scale, simple addition fusion presents dispersed activation patterns with insufficient target–background differentiation, especially with suboptimal activation intensity for small vehicles, whereas the feature map generated by the SRFM + STPEM fusion strategy possesses more precise target localization capability and boundary representation, with significantly improved background suppression effect and activation intensity distribution more concentrated on target regions, effectively enhancing small target detection performance.
  • The comparison of P4/16 scale feature maps shows that, although simple addition fusion can capture medium target positions, activation is not prominent enough and background noise interference exists; in contrast, the advanced fusion strategy produces more concentrated activation areas with higher target–background contrast and clearer boundaries between vehicles. As an intermediate resolution feature map, P4 (40 × 40) demonstrates superior structured representation and background suppression capability under the advanced fusion strategy.
  • At the P5/32 scale, rough semantic information generated by simple addition fusion makes it difficult to distinguish main vehicle targets, whereas the advanced fusion strategy can better capture overall scene semantics, accurately represent main vehicle targets, and effectively suppress background interference. Although the P5 feature map has the lowest resolution (20 × 20), it has the largest receptive field, and the advanced fusion strategy fully leverages its advantages in large target detection and scene understanding.
Through comparative analysis, we observed three key synergistic effects of multi-scale fusion:
  • Complementarity enhancement: the advanced fusion strategy makes features at different scales complementary, with P3 focusing on details and small targets, P4 processing medium targets, and P5 capturing large-scale structures and semantic information;
  • Information flow optimization: features at different scales mutually enhance each other, with semantic information guiding small target detection and detail information precisely locating large target boundaries;
  • Noise suppression capability: the advanced fusion strategy demonstrates superior background noise suppression capability at all scales, effectively reducing false detections.
Impact of recursive iteration count. The recursive progression mechanism is a key strategy in our proposed SDRFPT-Net model, which can further enhance feature representation capability through multiple recursive progressive fusions. To explore the optimal number of recursive iterations, we designed a series of experiments to observe model performance changes by varying the number of iterations. Table 7 shows the impact of iteration count on model performance.
The experimental results show that the number of recursive iterations significantly affects model performance. When the iteration count is 1 (D1), the model achieves an mAP50 of 0.769, indicating that even a single round of iteration can provide effective feature fusion. As the iteration count increases to 2 (D2) and 3 (D3), performance continues to improve, reaching mAP50 values of 0.783 and 0.785, and mAP50:95 values of 0.418 and 0.426, respectively. However, when the iteration count further increases to 4 (D4) and 5 (D5), performance begins to decline, with the mAP50 of 5 iterations significantly dropping to 0.761.
To more intuitively understand the impact of iteration count on feature representation, we conducted visualization analysis of feature maps at three feature scales—P3, P4, and P5—with different iteration counts, as shown in Figure 17.
Through visualization, we observed the following feature evolution patterns:
  • P3 feature layer (high resolution): As the iteration count increases, the feature map gradually evolves from initial dispersed response (n = 1) to more focused target representation (n = 2, 3), with clearer boundaries and a stronger background suppression effect. However, when the iteration count reaches 4 and 5, over-smoothing phenomena begin to appear, with some loss of boundary details.
  • P4 feature layer (medium resolution): At n = 1, the feature map has a basic response to targets but is not focused enough. After 2–3 rounds of iteration, the activation intensity of target areas significantly increases, improving target differentiation. Continuing to increase the iteration count to 4–5 rounds, the feature response begins to diffuse, reducing precise localization capability.
  • P5 feature layer (low resolution): This layer demonstrates the most obvious evolution trend, gradually developing from a blurred response at n = 1 to a highly structured representation at n = 3 that can clearly distinguish main targets. However, obvious signs of overfitting appear at n = 4 and n = 5, with feature maps becoming overly smoothed and target representation degrading.
These observations reveal the working mechanism of recursive progressive fusion: moderate iteration count (n = 3) can achieve progressive optimization of features through multiple rounds of interactive fusion of complementary information from different modalities, enhancing target feature representation and suppressing background interference. However, excessive iteration count may lead to “over-fusion” of features, i.e., the model overfits specific patterns in the training data, losing generalization capability.
Combining quantitative and visualization analysis results, we determined n = 3 as the optimal iteration count, achieving the best balance between feature enhancement and computational efficiency. This finding is also consistent with similar observations in other research areas, such as the optimal unfolding steps in recurrent neural networks and the optimal iteration count in message-passing neural networks, where similar “performance saturation points” exist.
Through the above ablation experiments, we verified the effectiveness and optimal configuration of each core component of SDRFPT-Net. The results show that the spectral hierarchical perception architecture (SHPA), the complete combination of three attention mechanisms, the full-scale advanced fusion strategy, and three rounds of recursive progressive fusion collectively contribute to the model’s superior performance. The rationality of these design choices is not only validated through quantitative metrics but also intuitively explained through feature visualization, providing new ideas for multispectral object detection research.
Fusion Strategy Optimization Analysis: Multi-scale fusion ablation (Table 6) reveals progressive performance gains with broader application of advanced fusion strategies. The most significant improvement occurs when transitioning from simple addition to SRFM + STPEM at P3/8 scale (C1→C2: +9.7%), highlighting the importance of high-resolution feature enhancement for small object detection. Scale-specific analysis shows that P3/8 benefits most from advanced fusion due to its role in capturing fine details, while P5/32 improvements are more modest but crucial for semantic understanding. Recursive iteration analysis (Table 7) demonstrates an optimal balance at n = 3, where performance plateaus before degrading due to over-smoothing effects. Feature evolution visualization reveals that iterations 1–3 progressively refine target representations, but iterations 4–5 lead to feature homogenization and loss of discriminative details. Computational efficiency analysis shows that our three-iteration design achieves 97% of the performance of deeper iterations while maintaining practical inference speeds, making it suitable for real-time applications.

5. Discussion

SDRFPT-Net demonstrated superior performance across the VEDAI, FLIR-aligned, and LLVIP datasets, effectively integrating complementary information from visible and infrared domains with significant improvements over single-modality methods (8.0% in mAP50 on FLIR-aligned, and 9.5% in mAP50:95 on LLVIP). The proposed spectral recursive fusion mechanism represents a computationally efficient innovation through cyclic weight reuse with parameter sharing, with ablation experiments confirming three recursive iterations as optimal for the progressive “feature distillation” process. Feature visualization validated the synergistic effects across different scales (P3 capturing details and small targets, P4 providing better target–background differentiation, P5 retaining richer semantic information) and the complementary functions of hybrid attention components (self-attention for spatial context, cross-modal attention for inter-modality exchange, channel attention for semantic enhancement).
Comparative analysis with existing methods reveals key advantages of our approach. The recursive fusion mechanism addresses fundamental limitations of shallow fusion methods like simple addition (YOLOv10-add) or basic attention mechanisms (CMAFF), achieving substantial improvements: 2.5% mAP50 and 5.4% mAP50:95 over SuperYOLO on VEDAI, and 11.5% mAP50 over BA-CAMF Net on FLIR-aligned. The parameter-sharing design provides computational efficiency compared to conventional dual-stream approaches while maintaining superior accuracy across diverse environmental conditions.
Performance analysis demonstrates consistent advantages across different scenarios. On VEDAI’s aerial imagery with complex terrain backgrounds, the 5.4% mAP50:95 improvement over CFT validates enhanced localization accuracy for small targets. On FLIR-aligned’s automotive scenarios, the substantial 11.5% mAP50 improvement indicates strong practical applicability. On LLVIP’s low-light conditions, the 9.5% mAP50:95 improvement over YOLOv8-infrared demonstrates effective fusion, even when visible modality is severely degraded.
Despite these advantages, SDRFPT-Net faces limitations including performance challenges in densely arranged targets, slower inference speed compared to single-modality detectors, and fixed three-round recursive iteration lacking adaptive capabilities. Ablation studies revealed performance degradation beyond three iterations, indicating potential over-fusion effects that limit scalability.
Future research directions include optimizing dense target detection through specialized loss functions, improving inference speed via model pruning and quantization, designing adaptive recursive mechanisms, extending the framework to incorporate more spectral modalities, and exploring integration with Vision Transformers and other recent architectures.

6. Conclusions

This paper presents SDRFPT-Net, a novel multispectral object detection architecture integrating three key modules: Spectral Hierarchical Perception Architecture (SHPA) for modality-specific feature extraction, Spectral Recursive Fusion Module (SRFM) for efficient cross-modal interaction, and Spectral Target Perception Enhancement Module (STPEM) for background suppression.
Experimental validation on the VEDAI, FLIR-aligned, and LLVIP datasets demonstrates significant improvements, achieving 2.5% mAP50 and 5.4% mAP50:95 on VEDAI, 11.5% mAP50 on FLIR-aligned, and 9.5% mAP50:95 on LLVIP, compared to state-of-the-art methods. The recursive fusion mechanism achieves computational efficiency through parameter sharing while maintaining superior accuracy across diverse environmental conditions.
Systematic ablation studies confirm the effectiveness of each component and optimal configurations. The proposed method provides robust solutions for remote sensing applications, intelligent surveillance, and autonomous driving domains. Future work will focus on adaptive iteration mechanisms, lightweight architectures, and extension to additional spectral modalities.

Author Contributions

Conceptualization, P.Z. and X.S.; methodology, P.Z. and B.S.; validation, X.S., B.S. and R.G.; formal analysis, P.Z.; investigation, P.Z. and X.S.; resources, P.Z.; data curation, B.S.; software implementation, P.Z.; writing—original draft preparation, P.Z.; writing—review and editing, X.S.; visualization, P.Z. and Z.D.; supervision, X.S. and S.S.; project administration, X.S.; P.Z. was responsible for the primary algorithm design and experimental implementation. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Hunan Provincial Postgraduate Research Innovation Programme, grant number XJZH2024033.

Data Availability Statement

The data presented in this study are available on request from the corresponding author. The data are not publicly available due to the relevance of data to individual privacy.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Li, K.; Wan, G.; Cheng, G.; Meng, L.; Han, J. Object Detection in Optical Remote Sensing Images: A Survey and a New Benchmark. ISPRS J. Photogramm. Remote Sens. 2020, 159, 296–307. [Google Scholar] [CrossRef]
  2. Zhang, C.; Chen, B.Y.; Lam, W.H.K.; Ho, H.W.; Shi, X.; Yang, X.; Ma, W.; Wong, S.C.; Chow, A.H.F. Vehicle Re-Identification for Lane-Level Travel Time Estimations on Congested Urban Road Networks Using Video Images. IEEE Trans. Intell. Transp. Syst. 2022, 23, 12877–12893. [Google Scholar] [CrossRef]
  3. Feng, D.; Haase-Schutz, C.; Rosenbaum, L.; Hertlein, H.; Glaser, C.; Timm, F.; Wiesbeck, W.; Dietmayer, K. Deep Multi-Modal Object Detection and Semantic Segmentation for Autonomous Driving: Datasets, Methods, and Challenges. IEEE Trans. Intell. Transp. Syst. 2021, 22, 1341–1360. [Google Scholar] [CrossRef]
  4. Li, C.; Cong, R.; Hou, J.; Zhang, S.; Qian, Y.; Kwong, S. Nested Network with Two-Stream Pyramid for Salient Object Detection in Optical Remote Sensing Images. IEEE Trans. Geosci. Remote Sens. 2019, 57, 9156–9166. [Google Scholar] [CrossRef]
  5. Yun Liu, X.-Y.Z. SAMNet: Stereoscopically Attentive Multi-Scale Network for Lightweight Salient Object Detection. IEEE Trans. Image Process. 2021, 30, 3804–3814. [Google Scholar]
  6. Ren, X.; Bai, Y.; Liu, G.; Zhang, P. YOLO-Lite: An Efficient Lightweight Network for SAR Ship Detection. Remote Sens. 2023, 15, 3771. [Google Scholar] [CrossRef]
  7. Qingyun, F.; Zhaokui, W. Cross-Modality Attentive Feature Fusion for Object Detection in Multispectral Remote Sensing Imagery. Pattern Recognit. 2022, 130, 108786. [Google Scholar] [CrossRef]
  8. Zhang, T.; Wu, H.; Liu, Y.; Peng, L.; Yang, C.; Peng, Z. Infrared Small Target Detection Based on Non-Convex Optimization with Lp-Norm Constraint. Remote Sens. 2019, 11, 559. [Google Scholar] [CrossRef]
  9. Pang, S.; Ge, J.; Hu, L.; Guo, K.; Zheng, Y.; Zheng, C.; Zhang, W.; Liang, J. RTV-SIFT: Harnessing Structure Information for Robust Optical and SAR Image Registration. Remote Sens. 2023, 15, 4476. [Google Scholar] [CrossRef]
  10. Song, K.; Bao, Y.; Wang, H.; Huang, L.; Yan, Y. A Potential Vision-Based Measurements Technology: Information Flow Fusion Detection Method Using RGB-Thermal Infrared Images. IEEE Trans. Instrum. Meas. 2023, 72, 1–13. [Google Scholar] [CrossRef]
  11. Liu, J.; Zhang, S.; Wang, S.; Metaxas, D.N. Multispectral Deep Neural Networks for Pedestrian Detection. arXiv 2016, arXiv:1611.02644. [Google Scholar]
  12. Wagner, J.; Fischer, V.; Herman, M.; Behnke, S. Multispectral Pedestrian Detection Using Deep Fusion Convolutional Neural Networks. In Proceedings of the European Symposium on Artificial Neural Networks, Computational Intelligence and Machine Learning, Bruges, Belgium, 27–29 April 2016. [Google Scholar]
  13. Konig, D.; Adam, M.; Jarvers, C.; Layher, G.; Neumann, H.; Teutsch, M. Fully Convolutional Region Proposal Networks for Multispectral Person Detection. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Honolulu, HI, USA, 21–26 July 2017; pp. 243–250. [Google Scholar]
  14. Zhang, Y.; Yu, H.; He, Y.; Wang, X.; Yang, W. Illumination-Guided RGBT Object Detection with Inter- and Intra-Modality Fusion. IEEE Trans. Instrum. Meas. 2023, 72, 2508013. [Google Scholar] [CrossRef]
  15. Zhou, W.; Zhu, Y.; Lei, J.; Wan, J.; Yu, L. CCAFNet: Crossflow and Cross-Scale Adaptive Fusion Network for Detecting Salient Objects in RGB-D Images. IEEE Trans. Multimed. 2022, 24, 2192–2204. [Google Scholar] [CrossRef]
  16. Wang, Z.; Yang, F.; P, Z.; Chen, L.; Ji, L. Multi-sensor image enhanced fusion algorithm based on NSST and top-hat transformation. Optik 2015, 126, 4184–4190. [Google Scholar] [CrossRef]
  17. Zhang, Q.; Liu, Y.; Blum, R.S.; Han, J.; Tao, D. Sparse representation based multi-sensor image fusion for multi-focus and multi-modality images: A review. Inf. Fusion 2018, 40, 57–75. [Google Scholar] [CrossRef]
  18. Ma, J.; Zhou, Z.; Wang, B.; Zong, H. Infrared and Visible Image Fusion Based on Visual Saliency Map and Weighted Least Square Optimization. Infrared Phys. Technol. 2017, 82, 8–17. [Google Scholar] [CrossRef]
  19. Qingyun, F.; Dapeng, H.; Zhaokui, W. Cross-modality fusion transformer for multispectral object detection. arXiv 2021, arXiv:2111.00273. [Google Scholar]
  20. Chen, Y.-T.; Shi, J.; Ye, Z.; Mertz, C.; Ramanan, D.; Kong, S. Multimodal object detection via probabilistic ensembling. In European Conference on Computer Vision; Springer Nature: Cham, Switzerland, 2022. [Google Scholar]
  21. Zhang, H.; Xu, H.; Tian, X.; Jiang, J.; Ma, J. Image Fusion Meets Deep Learning: A Survey and Perspective. Inf. Fusion 2021, 76, 323–336. [Google Scholar] [CrossRef]
  22. Fu, Y.; Wu, X.-J.; Durrani, T. Image Fusion Based on Generative Adversarial Network Consistent with Perception. Inf. Fusion 2021, 72, 110–125. [Google Scholar] [CrossRef]
  23. Li, J.; Huo, H.; Li, C.; Wang, R.; Sui, C.; Liu, Z. Multigrained Attention Network for Infrared and Visible Image Fusion. IEEE Trans. Instrum. Meas. 2021, 70, 5002412. [Google Scholar] [CrossRef]
  24. Wang, Z.; Wu, Y.; Wang, J.; Xu, J.; Shao, W. Res2Fusion: Infrared and Visible Image Fusion Based on Dense Res2net and Double Nonlocal Attention Models. IEEE Trans. Instrum. Meas. 2022, 71, 5005012. [Google Scholar] [CrossRef]
  25. Xie, X.; Cheng, G.; Rao, C.; Lang, C.; Han, J. Oriented Object Detection via Contextual Dependence Mining and Penalty-Incentive Allocation. IEEE Trans. Geosci. Remote Sens. 2024, 62, 5618010. [Google Scholar] [CrossRef]
  26. Zhang, L.; Zhu, X.; Chen, X.; Yang, X.; Lei, Z.; Liu, Z. Weakly Aligned Cross-Modal Learning for Multispectral Pedestrian Detection. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Republic of Korea, 27 October–2 November 2019; pp. 5126–5136. [Google Scholar]
  27. Zhang, X.; Wang, J.; Wang, T.; Jiang, R. Hierarchical Feature Fusion with Mixed Convolution Attention for Single Image Dehazing. IEEE Trans. Circuits Syst. Video Technol. 2022, 32, 510–522. [Google Scholar] [CrossRef]
  28. Li, J.; Huo, H.; Li, C.; Wang, R.; Feng, Q. AttentionFGAN: Infrared and Visible Image Fusion Using Attention-Based Generative Adversarial Networks. IEEE Trans. Multimed. 2021, 23, 1383–1396. [Google Scholar] [CrossRef]
  29. Zhao, Y.; Zhao, L.; Xiong, B.; Kuang, G. Attention Receptive Pyramid Network for Ship Detection in SAR Images. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2020, 13, 2738–2756. [Google Scholar] [CrossRef]
  30. Redmon, J.; Farhadi, A. YOLO9000: Better, faster, stronger. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016. [Google Scholar]
  31. Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You Only Look Once: Unified, Real-Time Object Detection. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 779–788. [Google Scholar]
  32. Wang, A.; Chen, H.; Liu, L.; Chen, K.; Lin, Z.; Han, J.; Ding, G. YOLOv10: Real-time end-to-end object detection. arXiv 2024, arXiv:2405.14458. [Google Scholar]
  33. Khanam, R.; Hussain, M. YOLOv11: An Overview of the Key Architectural Enhancements. arXiv 2024, arXiv:2410.17725. [Google Scholar]
  34. Tian, Y.; Ye, Q.; Doermann, D. YOLOv12: Attention-centric real-time object detectors. arXiv 2025, arXiv:2502.12524. [Google Scholar]
  35. Chang, Y.-L.; Anagaw, A.; Chang, L.; Wang, Y.C.; Hsiao, C.-Y.; Lee, W.-H. Ship Detection Based on YOLOv2 for SAR Imagery. Remote Sens. 2019, 11, 786. [Google Scholar] [CrossRef]
  36. Van Etten, A. You Only Look Twice: Rapid Multi-Scale Object Detection In Satellite Imagery. arXiv 2018, arXiv:1805.09512. [Google Scholar]
  37. Sharma, M.; Dhanaraj, M.; Karnam, S.; Chachlakis, D.G.; Ptucha, R.; Markopoulos, P.P.; Saber, E. YOLOrs: Object detection in multimodal remote sensing imagery. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2021, 14, 1497–1508. [Google Scholar] [CrossRef]
  38. Chen, L. Improved YOLOv3 Based on Attention Mechanism for Fast and Accurate Ship Detection in Optical Remote Sensing Images. Remote Sens. 2021, 13, 660. [Google Scholar] [CrossRef]
  39. Guo, Y.; Chen, S.; Zhan, R.; Wang, W.; Zhang, J. SAR Ship Detection Based on YOLOv5 Using CBAM and BiFPN; College of Electronic Science and Engineering, National University of Defense Technology: Changsha, China, 2022. [Google Scholar]
  40. Wang, Z.; Chen, Y.; Shao, W.; Li, H.; Zhang, L. SwinFuse: A Residual Swin Transformer Fusion Network for Infrared and Visible Images. IEEE Trans. Instrum. Meas. 2022, 71, 1–12. [Google Scholar] [CrossRef]
  41. Razakarivony, S.; Jurie, F. Vehicle detection in aerial imagery: A small target detection benchmark. J. Vis. Commun. Image Represent. 2016, 34, 187–203. [Google Scholar] [CrossRef]
  42. Zhang, H.; Fromont, E.; Lefevre, S.; Avignon, B. Multispectral Fusion for Object Detection with Cyclic Fuse-and-Refine Blocks. In Proceedings of the IEEE International Conference on Image Processing (ICIP), Abu Dhabi, Saudi Arabia, 25–28 October 2020. [Google Scholar]
  43. Jia, X.; Zhu, C.; Li, M.; Tang, W.; Liu, S.; Zhou, W. LLVIP: A Visible-Infrared Paired Dataset for Low-Light Vision. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 11–17 October 2021. [Google Scholar]
  44. Flir, T. Free FLIR Thermal Dataset for Algorithm Training. 2018. Available online: https://oem.flir.com/solutions/automotive/adas-dataset-form/ (accessed on 3 July 2025).
Figure 1. Comparison of multispectral object detection advantages under different lighting conditions. The figure shows detection results for visible (top row) and infrared (bottom row) imaging in daytime (ac) and nighttime (d) scenes. It clearly demonstrates that visible images (top row) provide richer color and texture information for better detection in daylight, while infrared images (bottom row) provide clearer object contours by capturing thermal radiation, showing significant advantages in low-light conditions. This complementarity proves the necessity of multispectral fusion for all-weather object detection, especially in complex and variable environmental conditions.
Figure 1. Comparison of multispectral object detection advantages under different lighting conditions. The figure shows detection results for visible (top row) and infrared (bottom row) imaging in daytime (ac) and nighttime (d) scenes. It clearly demonstrates that visible images (top row) provide richer color and texture information for better detection in daylight, while infrared images (bottom row) provide clearer object contours by capturing thermal radiation, showing significant advantages in low-light conditions. This complementarity proves the necessity of multispectral fusion for all-weather object detection, especially in complex and variable environmental conditions.
Remotesensing 17 02312 g001
Figure 2. Overall architecture of SDRFPT-Net (Spectral Dual-stream Recursive Fusion Perception Target Network). The architecture employs a dual-stream design with parallel processing paths for visible and infrared input images. The upper stream processes RGB features through C2f, C2f, SPPF, and C2f modules, while the lower stream handles IR features through identical but independently parameterized modules. Pre-Pro modules perform preprocessing including normalization and reshaping. The SRFM modules (marked in blue) conduct recursive fusion at multiple scales (P3/8, P4/16, P5/32). Post-Pro and STPEM modules enhance target perception before final feature aggregation through FPN and PAN structures. The network consists of three key innovative modules: Spectral Hierarchical Perception Architecture (SHPA) for extracting modality-specific features, Spectral Recursive Fusion Module (SRFM) for deep cross-modal feature interaction, and Spectral Target Perception Enhancement Module (STPEM) for enhancing target region representation and suppressing background interference. The feature pyramid and detection head (V10 Detect) enable multi-scale object detection.
Figure 2. Overall architecture of SDRFPT-Net (Spectral Dual-stream Recursive Fusion Perception Target Network). The architecture employs a dual-stream design with parallel processing paths for visible and infrared input images. The upper stream processes RGB features through C2f, C2f, SPPF, and C2f modules, while the lower stream handles IR features through identical but independently parameterized modules. Pre-Pro modules perform preprocessing including normalization and reshaping. The SRFM modules (marked in blue) conduct recursive fusion at multiple scales (P3/8, P4/16, P5/32). Post-Pro and STPEM modules enhance target perception before final feature aggregation through FPN and PAN structures. The network consists of three key innovative modules: Spectral Hierarchical Perception Architecture (SHPA) for extracting modality-specific features, Spectral Recursive Fusion Module (SRFM) for deep cross-modal feature interaction, and Spectral Target Perception Enhancement Module (STPEM) for enhancing target region representation and suppressing background interference. The feature pyramid and detection head (V10 Detect) enable multi-scale object detection.
Remotesensing 17 02312 g002
Figure 3. Dual-stream separated spectral architecture design in SDRFPT-Net. The architecture expands a single feature extraction network into a dual-stream structure, where the upper stream processes visible spectral information, while the lower stream handles infrared spectral information. Although both processing paths share similar network structures, they employ independent parameter sets for optimization, allowing each stream to specifically learn the feature distribution and representation of its respective modality.
Figure 3. Dual-stream separated spectral architecture design in SDRFPT-Net. The architecture expands a single feature extraction network into a dual-stream structure, where the upper stream processes visible spectral information, while the lower stream handles infrared spectral information. Although both processing paths share similar network structures, they employ independent parameter sets for optimization, allowing each stream to specifically learn the feature distribution and representation of its respective modality.
Remotesensing 17 02312 g003
Figure 4. Multi-scale fusion feature aggregation and detection process in SDRFPT-Net. The figure shows features from three different scales (P3, P4, P5) that already contain fused information from visible and infrared modalities. The middle section presents two complementary information flow networks: the Feature Pyramid Network (FPN), and Path Aggregation Network (PAN). The FPN process (light yellow background) follows Equation (8) where Pi = FPNi (Ffused), implementing a top-down pathway for semantic feature enhancement, while the PAN process (light pink background) follows Equation (9) where Mi represents the aggregated output through bottom-up feature integration. The right-hand side of the figure shows the network layer components including C2f modules for feature extraction, Conv layers for convolution operations, Concat for feature concatenation, C2fCIB for enhanced feature processing, Upsample for feature upsampling, and SCDown for spatial compression and downsampling. This bidirectional feature flow mechanism ensures that features at each scale incorporate both fine spatial localization information and rich semantic representation.
Figure 4. Multi-scale fusion feature aggregation and detection process in SDRFPT-Net. The figure shows features from three different scales (P3, P4, P5) that already contain fused information from visible and infrared modalities. The middle section presents two complementary information flow networks: the Feature Pyramid Network (FPN), and Path Aggregation Network (PAN). The FPN process (light yellow background) follows Equation (8) where Pi = FPNi (Ffused), implementing a top-down pathway for semantic feature enhancement, while the PAN process (light pink background) follows Equation (9) where Mi represents the aggregated output through bottom-up feature integration. The right-hand side of the figure shows the network layer components including C2f modules for feature extraction, Conv layers for convolution operations, Concat for feature concatenation, C2fCIB for enhanced feature processing, Upsample for feature upsampling, and SCDown for spatial compression and downsampling. This bidirectional feature flow mechanism ensures that features at each scale incorporate both fine spatial localization information and rich semantic representation.
Remotesensing 17 02312 g004
Figure 5. Detailed architecture of the Spectral Recursive Fusion Module (SRFM). The framework is divided into two main parts: The upper light blue background area shows the overall recursive fusion process, labeled as ‘SRFM n = 3’, indicating a three-round recursive fusion strategy. This part receives RGB and IR dual-stream features from SHPA and processes them through three cascaded hybrid attention units with parameter sharing to improve computational efficiency. The lower part shows the detailed internal structure of the hybrid attention unit, including preprocessing components (Pre-Pro) such as AvgPool, Flatten, and Position Encoding; the hybrid attention mechanism implementation including Channel Attention, Self-Attention, and Cross-modal Attention; and the MLP module with Linear layers, GELU activation, and Dropout.
Figure 5. Detailed architecture of the Spectral Recursive Fusion Module (SRFM). The framework is divided into two main parts: The upper light blue background area shows the overall recursive fusion process, labeled as ‘SRFM n = 3’, indicating a three-round recursive fusion strategy. This part receives RGB and IR dual-stream features from SHPA and processes them through three cascaded hybrid attention units with parameter sharing to improve computational efficiency. The lower part shows the detailed internal structure of the hybrid attention unit, including preprocessing components (Pre-Pro) such as AvgPool, Flatten, and Position Encoding; the hybrid attention mechanism implementation including Channel Attention, Self-Attention, and Cross-modal Attention; and the MLP module with Linear layers, GELU activation, and Dropout.
Remotesensing 17 02312 g005
Figure 6. Detailed structure of the hybrid attention mechanism in SDRFPT-Net. The mechanism integrates three complementary attention computation methods to achieve multi-dimensional feature enhancement. The upper part shows the overall processing flow: RGB and IR features first undergo reshaping and enter the Channel Attention module, which focuses on learning the importance weights of different feature channels. After reshaping back, the features simultaneously enter both the Self-Attention and Cross-modal Attention modules, capturing intra-modal spatial dependencies and inter-modal complementary information. The Cross-modal Attention modules implement the CrossAtt functions shown in Equations (11) and (12), where RGB features cross-attend to IR features and vice versa, as illustrated by the bidirectional connections between the RGB and IR processing streams in the lower part of the figure. Finally, the outputs from both attention modules are added to generate enhanced RGB and IR feature representations.
Figure 6. Detailed structure of the hybrid attention mechanism in SDRFPT-Net. The mechanism integrates three complementary attention computation methods to achieve multi-dimensional feature enhancement. The upper part shows the overall processing flow: RGB and IR features first undergo reshaping and enter the Channel Attention module, which focuses on learning the importance weights of different feature channels. After reshaping back, the features simultaneously enter both the Self-Attention and Cross-modal Attention modules, capturing intra-modal spatial dependencies and inter-modal complementary information. The Cross-modal Attention modules implement the CrossAtt functions shown in Equations (11) and (12), where RGB features cross-attend to IR features and vice versa, as illustrated by the bidirectional connections between the RGB and IR processing streams in the lower part of the figure. Finally, the outputs from both attention modules are added to generate enhanced RGB and IR feature representations.
Remotesensing 17 02312 g006
Figure 7. Spectral recursive progressive fusion architecture in SDRFPT-Net. The light blue background area (labeled as ‘SRFM n = 3’) shows the parameter-sharing three-round recursive fusion process. The left side includes RGB and IR input features after preprocessing (Pre-Pro), which flow through three cascaded hybrid attention units. The key innovation is that these three processing units share the exact same parameter set (indicated by ‘Parameter Sharing’ connections), achieving deep recursive structure without increasing model complexity. Each processing unit contains normalization (Norm) components and MLP modules, forming a complete feature refinement path.
Figure 7. Spectral recursive progressive fusion architecture in SDRFPT-Net. The light blue background area (labeled as ‘SRFM n = 3’) shows the parameter-sharing three-round recursive fusion process. The left side includes RGB and IR input features after preprocessing (Pre-Pro), which flow through three cascaded hybrid attention units. The key innovation is that these three processing units share the exact same parameter set (indicated by ‘Parameter Sharing’ connections), achieving deep recursive structure without increasing model complexity. Each processing unit contains normalization (Norm) components and MLP modules, forming a complete feature refinement path.
Remotesensing 17 02312 g007
Figure 8. Spectral Target Perception Enhancement Module (STPEM) structure and data flow. The module aims to enhance target region representation while suppressing background interference to improve detection accuracy. The figure is divided into three main parts: the upper and middle parts show parallel processing paths for features from RGB and IR modalities. Both feature paths first go through post-processing (Post-Pro) modules, including feature normalization, reshaping, and upsampling, before entering the STPEM module for enhancement processing.
Figure 8. Spectral Target Perception Enhancement Module (STPEM) structure and data flow. The module aims to enhance target region representation while suppressing background interference to improve detection accuracy. The figure is divided into three main parts: the upper and middle parts show parallel processing paths for features from RGB and IR modalities. Both feature paths first go through post-processing (Post-Pro) modules, including feature normalization, reshaping, and upsampling, before entering the STPEM module for enhancement processing.
Remotesensing 17 02312 g008
Figure 9. Object size distribution characteristics in multispectral detection datasets. (a) VEDAI dataset: The scatter plot shows 3700+ annotated vehicle targets across eight classes. Each point represents one object instance plotted by width (x-axis, 0–300 pixels) versus height (y-axis, 0–300 pixels). Color coding indicates vehicle types: boats (green), buses (red), camping cars (blue), cars (yellow), motorcycles (cyan), pickups (magenta), tractors (black), and trucks (orange). The main cluster centers around 20–100 pixel width and 10–100 pixel height, with tractors showing the largest height variance (up to 250 pixels). The marginal histograms display frequency distributions for width (top) and height (right). (b) FLIR-aligned dataset: Three object classes are color-coded as blue (person, 2845 instances), purple (car, 1674 instances), and orange (bicycle, 267 instances). Persons form a distinctive vertical cluster (narrow width 20–80 pixels, extended height 50–400 pixels), cars display a triangular distribution pattern reflecting viewing angle variations, and bicycles occupy an intermediate size range. (c) LLVIP dataset: Shows 16,836 pedestrian instances with highly concentrated circular clustering (width 20–80 pixels, height 40–150 pixels), indicating consistent target scales in nighttime surveillance scenarios. The tight clustering reflects controlled imaging conditions and uniform pedestrian detection ranges.
Figure 9. Object size distribution characteristics in multispectral detection datasets. (a) VEDAI dataset: The scatter plot shows 3700+ annotated vehicle targets across eight classes. Each point represents one object instance plotted by width (x-axis, 0–300 pixels) versus height (y-axis, 0–300 pixels). Color coding indicates vehicle types: boats (green), buses (red), camping cars (blue), cars (yellow), motorcycles (cyan), pickups (magenta), tractors (black), and trucks (orange). The main cluster centers around 20–100 pixel width and 10–100 pixel height, with tractors showing the largest height variance (up to 250 pixels). The marginal histograms display frequency distributions for width (top) and height (right). (b) FLIR-aligned dataset: Three object classes are color-coded as blue (person, 2845 instances), purple (car, 1674 instances), and orange (bicycle, 267 instances). Persons form a distinctive vertical cluster (narrow width 20–80 pixels, extended height 50–400 pixels), cars display a triangular distribution pattern reflecting viewing angle variations, and bicycles occupy an intermediate size range. (c) LLVIP dataset: Shows 16,836 pedestrian instances with highly concentrated circular clustering (width 20–80 pixels, height 40–150 pixels), indicating consistent target scales in nighttime surveillance scenarios. The tight clustering reflects controlled imaging conditions and uniform pedestrian detection ranges.
Remotesensing 17 02312 g009
Figure 10. Comparison of detection performance for different multispectral object detection models on the VEDAI dataset. Images are organized by columns from left to right: Ground Truth (GT), YOLOv10-add, CMAFF, CFT, and our proposed SDRFPT-Net model. Each row displays typical remote sensing scenarios with different environmental conditions and object distributions, including roads, building clusters, and open areas, from an aerial perspective. The visualization results clearly demonstrate the advantages of SDRFPT-Net: under complex terrain textures and multi-scale remote sensing imaging conditions, SDRFPT-Net can accurately detect all targets with more precise bounding box localization. In contrast, other methods such as YOLOv10-add, CMAFF, and CFT exhibit certain limitations in small target detection and bounding box localization accuracy.
Figure 10. Comparison of detection performance for different multispectral object detection models on the VEDAI dataset. Images are organized by columns from left to right: Ground Truth (GT), YOLOv10-add, CMAFF, CFT, and our proposed SDRFPT-Net model. Each row displays typical remote sensing scenarios with different environmental conditions and object distributions, including roads, building clusters, and open areas, from an aerial perspective. The visualization results clearly demonstrate the advantages of SDRFPT-Net: under complex terrain textures and multi-scale remote sensing imaging conditions, SDRFPT-Net can accurately detect all targets with more precise bounding box localization. In contrast, other methods such as YOLOv10-add, CMAFF, and CFT exhibit certain limitations in small target detection and bounding box localization accuracy.
Remotesensing 17 02312 g010aRemotesensing 17 02312 g010b
Figure 11. Comparison of detection performance for different multispectral object detection models on various scenarios in the FLIR-aligned dataset. Images are organized by columns from left to right: Ground Truth (GT), YOLOv10-add, TFDet, CMAFF, and our proposed SDRFPT-Net model. Each row shows typical scenarios with different environmental conditions and object distributions, including close-range vehicles, multiple roadway targets, parking areas, narrow streets, and open roads. The visualization results clearly demonstrate the advantages of SDRFPT-Net: In the first row’s close-range vehicle scene, SDRFPT-Net’s bounding boxes almost perfectly match GT. In the second row’s complex multi-target scene, it successfully detects all pedestrians and bicycles without obvious misses. In the third row’s parking lot scene, it accurately identifies multiple closely parked vehicles with precise bounding box localization.
Figure 11. Comparison of detection performance for different multispectral object detection models on various scenarios in the FLIR-aligned dataset. Images are organized by columns from left to right: Ground Truth (GT), YOLOv10-add, TFDet, CMAFF, and our proposed SDRFPT-Net model. Each row shows typical scenarios with different environmental conditions and object distributions, including close-range vehicles, multiple roadway targets, parking areas, narrow streets, and open roads. The visualization results clearly demonstrate the advantages of SDRFPT-Net: In the first row’s close-range vehicle scene, SDRFPT-Net’s bounding boxes almost perfectly match GT. In the second row’s complex multi-target scene, it successfully detects all pedestrians and bicycles without obvious misses. In the third row’s parking lot scene, it accurately identifies multiple closely parked vehicles with precise bounding box localization.
Remotesensing 17 02312 g011aRemotesensing 17 02312 g011b
Figure 12. Comparison of pedestrian detection performance for different detection models on nighttime low-light scenes from the LLVIP dataset. Images are organized by columns from left to right: Ground Truth (GT), YOLOv10-add, TFDet, CMAFF, and our proposed SDRFPT-Net model. The rows display five typical nighttime scenes representing challenging situations with different lighting conditions, viewing angles, and target distances. In all low-light scenes, SDRFPT-Net demonstrates excellent pedestrian detection capability: The model accurately identifies distant pedestrians with precise bounding boxes in the first row’s street lighting scene, maintains stable detection performance despite strong light interference in the second and fourth rows, and successfully detects distant pedestrians that other methods tend to miss in the fifth row’s dark area.
Figure 12. Comparison of pedestrian detection performance for different detection models on nighttime low-light scenes from the LLVIP dataset. Images are organized by columns from left to right: Ground Truth (GT), YOLOv10-add, TFDet, CMAFF, and our proposed SDRFPT-Net model. The rows display five typical nighttime scenes representing challenging situations with different lighting conditions, viewing angles, and target distances. In all low-light scenes, SDRFPT-Net demonstrates excellent pedestrian detection capability: The model accurately identifies distant pedestrians with precise bounding boxes in the first row’s street lighting scene, maintains stable detection performance despite strong light interference in the second and fourth rows, and successfully detects distant pedestrians that other methods tend to miss in the fifth row’s dark area.
Remotesensing 17 02312 g012
Figure 13. Learning curves of SDRFPT-Net on the FLIR-aligned dataset. (a) The precision–confidence curves show the variation in detection precision for each category under different confidence thresholds. (b) The recall–confidence curves demonstrate the trend in recall with confidence threshold changes. (c) The precision–recall curves reflect the performance trade-off at different operating points. (d) The F1–confidence curves comprehensively evaluate the overall detection performance of the model.
Figure 13. Learning curves of SDRFPT-Net on the FLIR-aligned dataset. (a) The precision–confidence curves show the variation in detection precision for each category under different confidence thresholds. (b) The recall–confidence curves demonstrate the trend in recall with confidence threshold changes. (c) The precision–recall curves reflect the performance trade-off at different operating points. (d) The F1–confidence curves comprehensively evaluate the overall detection performance of the model.
Remotesensing 17 02312 g013
Figure 14. Comparative impact of different attention mechanisms on the P3 feature layer (high-resolution features) in SDRFPT-Net. The top row (ad) presents feature activation maps, while the bottom row (eh) shows the corresponding original image heatmap overlay effects, demonstrating the differences in feature attention patterns. Self-attention (a,e) focuses on target contours and edge information. Cross-modal attention (b,f) presents overall attention to target areas with complementary information from RGB and IR modalities. Channel attention (c,g) demonstrates selective enhancement of specific semantic information. Hybrid attention (d,h) combines the advantages of all three mechanisms for optimal feature representation.
Figure 14. Comparative impact of different attention mechanisms on the P3 feature layer (high-resolution features) in SDRFPT-Net. The top row (ad) presents feature activation maps, while the bottom row (eh) shows the corresponding original image heatmap overlay effects, demonstrating the differences in feature attention patterns. Self-attention (a,e) focuses on target contours and edge information. Cross-modal attention (b,f) presents overall attention to target areas with complementary information from RGB and IR modalities. Channel attention (c,g) demonstrates selective enhancement of specific semantic information. Hybrid attention (d,h) combines the advantages of all three mechanisms for optimal feature representation.
Remotesensing 17 02312 g014
Figure 15. Representational differences between dual-attention mechanism combinations and the complete triple-attention mechanism on the P3 feature layer. The top row (ad) shows feature activation maps, while the bottom row (eh) shows original image heatmap overlay effects, revealing the complementarity and synergistic effects of different attention combinations. Self + Cross-modal attention (a,e) simultaneously possesses excellent boundary localization and target region representation; Self + Channel attention (b,f) enhances specific semantic features while preserving boundary information; Cross + Channel attention (c,g) enhances channel representation based on multi-modal fusion but lacks spatial context. Hybrid attention (d,h) achieves the most comprehensive and effective feature representation through synergistic integration of all three mechanisms.
Figure 15. Representational differences between dual-attention mechanism combinations and the complete triple-attention mechanism on the P3 feature layer. The top row (ad) shows feature activation maps, while the bottom row (eh) shows original image heatmap overlay effects, revealing the complementarity and synergistic effects of different attention combinations. Self + Cross-modal attention (a,e) simultaneously possesses excellent boundary localization and target region representation; Self + Channel attention (b,f) enhances specific semantic features while preserving boundary information; Cross + Channel attention (c,g) enhances channel representation based on multi-modal fusion but lacks spatial context. Hybrid attention (d,h) achieves the most comprehensive and effective feature representation through synergistic integration of all three mechanisms.
Remotesensing 17 02312 g015
Figure 16. Comparison between simple addition fusion and innovative fusion strategies (SRFM + STPEM) on three feature scale layers. The upper row (ac) presents traditional simple addition fusion at different scales: The P3/8 high-resolution layer (a) shows dispersed activation with insufficient target–background differentiation. The P4/16 medium-resolution layer (b) has some response to vehicle areas but with blurred boundaries. The P5/32 low-resolution layer (c) only has a rough response to the central vehicle. The lower row (df) shows feature maps of the innovative fusion strategy: The P3/8 layer (d) provides clearer vehicle contour representation with precise edge localization. The P4/16 layer (e) shows more concentrated target area activation. The P5/32 layer (f) preserves richer scene semantic information while enhancing central target representation.
Figure 16. Comparison between simple addition fusion and innovative fusion strategies (SRFM + STPEM) on three feature scale layers. The upper row (ac) presents traditional simple addition fusion at different scales: The P3/8 high-resolution layer (a) shows dispersed activation with insufficient target–background differentiation. The P4/16 medium-resolution layer (b) has some response to vehicle areas but with blurred boundaries. The P5/32 low-resolution layer (c) only has a rough response to the central vehicle. The lower row (df) shows feature maps of the innovative fusion strategy: The P3/8 layer (d) provides clearer vehicle contour representation with precise edge localization. The P4/16 layer (e) shows more concentrated target area activation. The P5/32 layer (f) preserves richer scene semantic information while enhancing central target representation.
Remotesensing 17 02312 g016
Figure 17. Impact of iteration counts (n = 1 to n = 5) in the recursive progressive fusion strategy on three feature scale layers of SDRFPT-Net. By comparing the evolution within the same row, changes in features with recursive depth can be observed. By comparing different rows, response characteristics at different scales can be understood. The P3 high-resolution layer (ae) shows feature representation gradually evolving from initial dispersed response (n = 1) to more focused target contours (n = 2, 3), with clearer boundaries and stronger background suppression, but experiencing over-smoothing at n = 4, 5. The P4 medium-resolution layer (fj) shows optimal target–background differentiation at n = 3, followed by feature response diffusion at n = 4, 5. The P5 low-resolution layer (ko) presents the most significant changes, achieving highly structured representation at n = 3 that clearly distinguishes main scene elements, while showing obvious degradation at n = 4 and n = 5.
Figure 17. Impact of iteration counts (n = 1 to n = 5) in the recursive progressive fusion strategy on three feature scale layers of SDRFPT-Net. By comparing the evolution within the same row, changes in features with recursive depth can be observed. By comparing different rows, response characteristics at different scales can be understood. The P3 high-resolution layer (ae) shows feature representation gradually evolving from initial dispersed response (n = 1) to more focused target contours (n = 2, 3), with clearer boundaries and stronger background suppression, but experiencing over-smoothing at n = 4, 5. The P4 medium-resolution layer (fj) shows optimal target–background differentiation at n = 3, followed by feature response diffusion at n = 4, 5. The P5 low-resolution layer (ko) presents the most significant changes, achieving highly structured representation at n = 3 that clearly distinguishes main scene elements, while showing obvious degradation at n = 4 and n = 5.
Remotesensing 17 02312 g017
Table 1. Performance comparison of SDRFPT-Net with state-of-the-art methods on the VEDAI dataset. The table presents precision (P), recall (R), mean average precision at IoU threshold of 0.5 (mAP50), and mean average precision across IoU thresholds from 0.5 to 0.95 (mAP50:95). V-I indicates the fusion of visible and infrared modalities. The best results are highlighted in bold.
Table 1. Performance comparison of SDRFPT-Net with state-of-the-art methods on the VEDAI dataset. The table presents precision (P), recall (R), mean average precision at IoU threshold of 0.5 (mAP50), and mean average precision across IoU thresholds from 0.5 to 0.95 (mAP50:95). V-I indicates the fusion of visible and infrared modalities. The best results are highlighted in bold.
MethodsModalityPRmAP50mAP50:95
YOLOv5Visible0.7040.6040.6750.398
Infrared0.5210.5140.4980.280
YOLOv8Visible0.7270.4310.5370.333
Infrared0.5200.5230.4940.291
YOLOv10Visible0.6100.5230.5870.329
Infrared0.4210.4630.4470.244
DETRVisible0.6820.5480.6210.356
Infrared0.5470.4890.5320.298
YOLOv10-addV-I0.5000.5270.5370.276
CFTV-I0.7010.6270.6720.427
CMAFFV-I0.6160.5080.4520.275
SuperYOLOV-I0.7900.6780.7160.425
SDRFPT-Net (ours)V-I0.7960.6830.7340.450
Table 2. Performance comparison of SDRFPT-Net with state-of-the-art methods on the FLIR-aligned dataset. The table presents precision (P), recall (R), mean average precision at IoU threshold of 0.5 (mAP50), and mean average precision across IoU thresholds from 0.5 to 0.95 (mAP50:95). The best results are highlighted in bold.
Table 2. Performance comparison of SDRFPT-Net with state-of-the-art methods on the FLIR-aligned dataset. The table presents precision (P), recall (R), mean average precision at IoU threshold of 0.5 (mAP50), and mean average precision across IoU thresholds from 0.5 to 0.95 (mAP50:95). The best results are highlighted in bold.
MethodsModalityPRmAP50mAP50:95
YOLOv5Visible0.5310.3950.4410.202
Infrared0.6250.4680.5390.272
YOLOv8Visible0.5320.3960.4480.218
Infrared0.5590.5140.5490.288
YOLOv10Visible0.7270.5380.6200.305
Infrared0.7730.6180.7270.424
DETRVisible0.6980.5210.5950.289
Infrared0.7410.5920.6810.378
YOLOv10-addV-I0.7480.6230.7010.354
CMA-DetV-I0.8120.4680.5180.237
TFDetV-I0.8270.6060.6530.346
CMAFFV-I0.7920.5500.5580.302
BA-CAMF NetV-I0.7980.6320.7040.351
SDRFPT-Net (ours)V-I0.8540.7000.7850.426
Table 3. Performance comparison of SDRFPT-Net with state-of-the-art methods on the LLVIP dataset. The table presents precision (P), recall (R), mean average precision at IoU threshold of 0.5 (mAP50), and mean average precision across IoU thresholds from 0.5 to 0.95 (mAP50:95). V-I indicates the fusion of visible and infrared modalities. The best results are highlighted in bold.
Table 3. Performance comparison of SDRFPT-Net with state-of-the-art methods on the LLVIP dataset. The table presents precision (P), recall (R), mean average precision at IoU threshold of 0.5 (mAP50), and mean average precision across IoU thresholds from 0.5 to 0.95 (mAP50:95). V-I indicates the fusion of visible and infrared modalities. The best results are highlighted in bold.
MethodsModalityPRmAP50mAP50:95
YOLOv5Visible0.9060.8200.8950.504
Infrared0.9620.8980.9600.631
YOLOv8Visible0.9330.8290.8960.513
Infrared0.9560.9010.9610.645
YOLOv10Visible0.9140.8330.8920.512
Infrared0.9620.9090.9610.637
DETRVisible0.8890.7950.8630.485
Infrared0.9480.8850.9340.598
YOLOv10-addV-I0.9610.8930.9570.628
TFDetV-I0.9600.8960.9600.594
CMAFFV-I0.9580.8990.9150.574
BA-CAMF NetV-I0.8660.8280.8870.511
SDRFPT-Net (ours)V-I0.9630.9110.9630.706
Table 4. Impact of different attention combinations on detection performance. The table compares the effects of Self Attention, Cross-modal Attention, and Channel Attention in various combinations. The best results are highlighted in bold.
Table 4. Impact of different attention combinations on detection performance. The table compares the effects of Self Attention, Cross-modal Attention, and Channel Attention in various combinations. The best results are highlighted in bold.
IDSHPASRFMSTPEMmAP50mAP50:95
A1 0.7010.354
A2 0.7750.373
A30.7850.426
Table 5. Impact of different attention combinations on detection performance. The table compares the effects of self-attention, cross-modal attention, and channel attention in various combinations. The best results are highlighted in bold.
Table 5. Impact of different attention combinations on detection performance. The table compares the effects of self-attention, cross-modal attention, and channel attention in various combinations. The best results are highlighted in bold.
IDSelf-AttentionCross-AttentionChannel-AttentionmAP50mAP50:95
B1 0.7760.408
B2 0.7490.372
B3 0.7300.384
B4 0.7740.424
B5 0.7630.409
B6 0.7290.362
B70.7850.426
Table 6. Ablation experiments on fusion positions. The table shows the impact of applying advanced fusion modules at different feature scales. P3/8, P4/16, and P5/32 represent feature maps at different scales, with numbers indicating the downsampling factor relative to the input image. The best results are highlighted in bold.
Table 6. Ablation experiments on fusion positions. The table shows the impact of applying advanced fusion modules at different feature scales. P3/8, P4/16, and P5/32 represent feature maps at different scales, with numbers indicating the downsampling factor relative to the input image. The best results are highlighted in bold.
IDP3/8P4/16P5/32mAP50mAP50:95
C1AddAddAdd0.7010.354
C2SRFM + STPEMAddAdd0.7690.404
C3SRFM + STPEMSRFM + STPEMAdd0.7760.410
C4SRFM + STPEMSRFM + STPEMSRFM + STPEM0.7850.426
Table 7. Impact of different iteration counts of the recursive progressive fusion strategy on model detection performance. The experiment compares performance with recursive depths from 1 to 5 iterations (D1–D5), evaluating detection accuracy using mAP50 and mAP50:95 metrics. The best results are highlighted in bold.
Table 7. Impact of different iteration counts of the recursive progressive fusion strategy on model detection performance. The experiment compares performance with recursive depths from 1 to 5 iterations (D1–D5), evaluating detection accuracy using mAP50 and mAP50:95 metrics. The best results are highlighted in bold.
IDTimesmAP50mAP50:95
D110.7690.395
D220.7830.418
D330.7850.426
D440.7830.400
D550.7610.417
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhou, P.; Sun, X.; Sun, B.; Guo, R.; Dang, Z.; Su, S. SDRFPT-Net: A Spectral Dual-Stream Recursive Fusion Network for Multispectral Object Detection. Remote Sens. 2025, 17, 2312. https://doi.org/10.3390/rs17132312

AMA Style

Zhou P, Sun X, Sun B, Guo R, Dang Z, Su S. SDRFPT-Net: A Spectral Dual-Stream Recursive Fusion Network for Multispectral Object Detection. Remote Sensing. 2025; 17(13):2312. https://doi.org/10.3390/rs17132312

Chicago/Turabian Style

Zhou, Peida, Xiaoyong Sun, Bei Sun, Runze Guo, Zhaoyang Dang, and Shaojing Su. 2025. "SDRFPT-Net: A Spectral Dual-Stream Recursive Fusion Network for Multispectral Object Detection" Remote Sensing 17, no. 13: 2312. https://doi.org/10.3390/rs17132312

APA Style

Zhou, P., Sun, X., Sun, B., Guo, R., Dang, Z., & Su, S. (2025). SDRFPT-Net: A Spectral Dual-Stream Recursive Fusion Network for Multispectral Object Detection. Remote Sensing, 17(13), 2312. https://doi.org/10.3390/rs17132312

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop