Next Article in Journal
Multi-UAV-Assisted Task Offloading and Trajectory Optimization for Edge Computing via NOMA
Previous Article in Journal
Digital Twin-Driven Intrusion Detection for Industrial SCADA: A Cyber-Physical Case Study
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Modality Alignment and Fusion-Based Method for Around-the-Clock Remote Sensing Object Detection

1
School of Computer Science and Engineering, North China Institute of Aerospace Engineering, Langfang 065000, China
2
Hebei Collaborative Innovation Center of Micro Nano Satellites, North China Institute of Aerospace Engineering, Langfang 065000, China
3
School of Information Engineering, Capital Normal University, Beijing 100048, China
4
School of Optoelectronic Engineering, Xidian University, Xi’an 710071, China
5
School of Electronics and Information Engineering, Hebei University of Technology, Tianjin 300401, China
*
Authors to whom correspondence should be addressed.
Sensors 2025, 25(16), 4964; https://doi.org/10.3390/s25164964
Submission received: 19 June 2025 / Revised: 31 July 2025 / Accepted: 7 August 2025 / Published: 11 August 2025
(This article belongs to the Section Remote Sensors)

Abstract

Cross-modal remote sensing object detection holds significant potential for around-the-clock applications. However, the modality differences between cross-modal data and the degradation of feature quality under adverse weather conditions limit detection performance. To address these challenges, this paper presents a novel cross-modal remote sensing object detection framework designed to overcome two critical challenges in around-the-clock applications: (1) significant modality disparities between visible light, infrared, and synthetic aperture radar data, and (2) severe feature degradation under adverse weather conditions including fog, and nighttime scenarios. Our primary contributions are as follows: First, we develop a multi-scale feature extraction module that employs a hierarchical convolutional architecture to capture both fine-grained details and contextual information, effectively compensating for missing or blurred features in degraded visible-light images. Second, we introduce an innovative feature interaction module that utilizes cross-attention mechanisms to establish long-range dependencies across modalities while dynamically suppressing noise interference through adaptive feature selection. Third, we propose a feature correction fusion module that performs spatial alignment of object boundaries and channel-wise optimization of global feature consistency, enabling robust fusion of complementary information from different modalities. The proposed framework is validated on visible light, infrared, and SAR modalities. Extensive experiments on three challenging datasets (LLVIP, OGSOD, and Drone Vehicle) demonstrate our framework’s superior performance, achieving state-of-the-art mean average precision scores of 66.3%, 58.6%, and 71.7%, respectively, representing significant improvements over existing methods in scenarios with modality differences or extreme weather conditions. The proposed solution not only advances the technical frontier of cross-modal object detection but also provides practical value for mission-critical applications such as 24/7 surveillance systems, military reconnaissance, and emergency response operations where reliable around-the-clock detection is essential.

1. Introduction

Remote sensing object detection has emerged as a fundamental technology in computer vision, with widespread applications spanning intelligent surveillance systems [1], wildlife conservation [2], and military reconnaissance operations [3]. The evolution of convolutional neural networks (CNNs) has significantly advanced detection methodologies [1,2,3,4], yet most existing approaches remain constrained to visible-light imagery or high-visibility conditions [5,6]. This limitation becomes particularly problematic when dealing with low-visibility scenarios (e.g., nighttime) or adverse weather conditions (e.g., haze, fog) [5,6], where single-modal detectors often fail to maintain robust performance. Consequently, researchers have increasingly turned to multi-modal solutions that combine visible light with complementary modalities such as infrared imagery or synthetic aperture radar (SAR) [7,8], aiming to achieve reliable around-the-clock detection through comprehensive feature fusion [9,10].
Prior work in cross-modal remote sensing object detection has primarily addressed two technical challenges through various approaches. For modality alignment, Sun et al. [11] proposed UA-CMDet to enhance spatial and semantic alignment through joint optimization, though its fusion strategy showed limited flexibility in complex scenarios. Song et al. [12] developed CMADet with cross-modal multi-scale attention, yet struggled with complete feature alignment in dynamic scenes. Regarding feature fusion, Yuan et al. [13] introduced C2Former using Transformer’s cross-attention mechanism, while Wang et al. [14] improved YOLO’s fusion strategy in YOLOFIV, both achieving notable but incomplete progress [13,14,15,16]. Other notable contributions include Qingyun et al.’s CFT framework [15] for unified feature space mapping and Bao et al.’s DDCI [16] with deep dual cross-interaction, though these methods still face limitations in noise suppression and dynamic scene adaptation [15,16]. Recent advances like GAFF [17], ProbEN [18], and CSSA [19] have further enriched the field through novel fusion architectures and probabilistic modeling, while knowledge distillation techniques (KD [20], DKD [21]) and local descriptor methods (LD [22], CoLD [23]) have provided alternative optimization pathways [20,21,22,23]. Attention mechanisms have also played a pivotal role, from SENet’s channel-wise modulation [24] to transformer-based global context modeling [25,26,27,28], particularly benefiting cross-modal interpretation tasks [25].
Despite these advancements, two critical limitations persist in current research:
(1)
Existing methods often address either modality differences or feature degradation in isolation, failing to fully exploit spatial-semantic complementary information across visible light, infrared, and SAR modalities [29,30];
(2)
Most solutions lack robust mechanisms for handling severe feature quality deterioration under extreme weather conditions, where visible light images suffer from illumination insufficiency while infrared images are affected by thermal noise [30]. These limitations become particularly apparent in mission-critical applications requiring operational capability, where even state-of-the-art methods like CCFINet [31] and CrossFormer [32] show compromised accuracy during modality transitions or weather extremes [31,32].
To overcome these challenges, this paper makes four key contributions:
(1)
We propose a comprehensive cross-modal detection framework that simultaneously addresses modality differences and feature degradation through synergistic module design.
(2)
We develop a multi-scale feature extraction module (MSFEM) with a multi-branch convolutional architecture that captures diverse receptive fields, effectively compensating for lost visible-light features in adverse conditions.
(3)
We introduce a feature interaction module (FIM) employing cross-attention mechanisms to model long-range inter-modal dependencies while implementing dynamic feature selection for noise suppression.
(4)
We design a feature correction fusion module (FCFM) that performs spatial boundary alignment and channel-wise consistency optimization through novel correction mechanisms.
As illustrated in Figure 1, our complete architecture demonstrates superior performance across three challenging datasets (LLVIP, OGSOD, Drone Vehicle), achieving mAP improvements of 2.2–13.6% over existing methods while maintaining efficient computational characteristics, a critical advantage for real-time applications. The network has a dual-stream architecture, with primary modules including the multi-scale feature extraction module, the feature interaction module, and the feature correction fusion module. where h i and ψ i denote the convolutional modules for the RGB and IR modalities at the i -th layer, respectively. F R i and F I i represent the feature maps of the corresponding modalities at the i -th layer. P 3 , P 4 , and P 5 serve as the inputs to the feature pyramid of the last three stages.
The remainder of this paper is organized as follows: Section 2 details our methodology, including mathematical formulations and module designs. Section 3 presents experimental results and comparative analyses. Section 4 concludes with implications and future research directions.

2. Methodology

2.1. Problem Formulation and Network Overview

Building upon the challenges identified in Section 1, particularly the modality disparities between visible light (RGB), infrared (IR), and synthetic aperture radar (SAR) data, along with feature degradation under adverse weather conditions, we propose a dual-stream architecture with three specialized modules. As shown in Figure 1, our framework processes aligned image pairs through parallel feature extraction streams, followed by multi-scale enhancement and cross-modal fusion.
The complete processing pipeline can be formulated as follows:
F o u t = F C F M ( F I M M S F E M ( F r g b ) , M S F E M ( F i r ) ) ,
where F r g b and F i r denote the input feature maps from RGB and IR/SAR modalities, respectively, and F o u t represents the final fused features for detection.
In designing this architecture, we aimed to create a coherent and comprehensive solution that addresses significant modality disparities and ensures robust performance under unfavorable lighting conditions. The multi-scale feature extraction module (MSFEM) captures diverse receptive fields using a multi-branch structure with convolutional operations of different scales, compensating for lost visible-light features. The feature interaction module (FIM) then employs cross-attention mechanisms to model long-range inter-modal dependencies, dynamically fusing features from different modalities to leverage their strengths while suppressing noise. Finally, the feature correction fusion module (FCFM) performs spatial boundary alignment and channel-wise consistency optimization, ensuring the accurate and consistent fusion of features. Together, these modules form a robust and reliable framework for cross-modal object detection, capable of handling multi-modal remote sensing data under various environmental conditions.

2.2. Multi-Scale Feature Extraction Module

Remote sensing images typically exhibit a large number of small objects that are densely distributed. Traditional convolutional neural networks (CNNs), limited by their fixed receptive fields, struggle to capture rich object feature information. To address this issue and enhance the network’s perception of small objects, this paper proposes a multi-scale feature extraction module (MSFEM). By employing convolutional operations of different scales within a multi-branch structure, the multi-scale feature extraction module can extract object feature information from multiple receptive field sizes. This enables the network to better understand the relationships between pixels in the feature map, thereby improving detection accuracy. Figure 2 illustrates the detailed structure of the multi-scale feature extraction module. The multi-scale feature extraction module consists of four feature extraction branches. The main branch includes convolutional modules of 1 × 1, 3 × 3, and 5 × 5, while the residual branch contains only a 1 × 1 convolutional module. It is worth noting that each convolutional module comprises three components: a convolutional layer, batch normalization, and a SiLU activation function.
The feature extraction process of the multi-scale feature extraction module is described as follows. First, the input feature map (H × W × C) is divided into a residual branch and a main branch. In the main branch, a 1 × 1 convolutional operation is initially performed to maintain the size of the feature map. Subsequently, the main branch is further divided into three convolutional branches. The first convolutional branch employs a single 1 × 1 convolutional module for feature extraction. The second convolutional branch applies a 3 × 3 convolutional module followed by a 1 × 1 convolutional module. The third convolutional branch utilizes a 5 × 5 convolutional module followed by a 1 × 1 convolutional module. Next, the three convolutional branches are concatenated along the channel dimension and passed through a 1 × 1 convolutional module, resulting in an output feature map size of H × W × C. In the residual branch, a single 1 × 1 convolutional module is used, yielding an output feature map size of H × W × C. The output feature maps from the main branch and the residual branch are then concatenated along the channel dimension, producing a feature map size of H × W × 2C. To match the channel number of the original input feature map, a 1 × 1 convolutional module is employed for dimensionality reduction, resulting in the output feature map of the multi-scale feature extraction module. At this point, the output feature map of the multi-scale feature extraction module has the same size as the input feature map, namely H × W × C.

2.3. Feature Interaction Module

To fully exploit the complementarity between different modalities and capture the long-range global dependencies between modalities, this paper proposes a feature interaction module. As shown in Figure 3, the feature interaction module takes F r g b and F i r as inputs and outputs the corresponding enhanced features F ^ r g b and F ^ i r after feature interaction.
Generally, the input features of the feature interaction module are first converted into patch embedding sequences and then fed into the cross-attention sub-module to perform cross-modal interactions by modeling the long-range relationships between the two modalities. Subsequently, the fused patch embeddings are, respectively, input into two feed-forward networks, and then reshaped and rearranged to obtain the interacted features. Finally, the enhanced features are obtained by the element-wise addition of the interacted features and the input features.
Initially, two 1 × 1 convolutional layers are employed to reduce the channel number of F r g b , F i r R H × W × C resulting in F r g b , F i r R H × W × C , where (H, W) represents the spatial resolution of the feature maps, C is the original channel number, and C is the reduced channel number. Subsequently, the multi-head cross-attention and the two feed-forward networks require sequences as inputs. Therefore, F r g b and F i r are reshaped into two flattened 2D patch sequences x r g b , x i r R N × d p a t c h , where d p a t c h = P 2 × C is the dimension of each flattened patch, with a resolution of (P, P), and N = H W / P 2 is the number of patches. Thereafter, through two fully connected layers, x r g b and x i r are encoded into patch embeddings x r g b e m b , x i r e m b R N × d , where d is the dimension of each patch embedding. Finally, x r g b e m b , x i r e m b are further processed through two normalization layers.
As shown in Figure 3, the patch embeddings are fed into the multi-head cross-attention sub-module. Specifically, x r g b e m b , x i r e m b are linearly projected to generate their respective queries, keys, and values, denoted as Q r g b , K r g b , V r g b R N × d and Q i r , K i r , V i r R N × d . Subsequently, scaled dot-product attention is applied to each head, which can be expressed as follows:
h e a d r g b , i = S o f t m a x Q r g b , i ,     K i r , i d k V i r , i ,
h e a d i r , i = S o f t m a x ( Q i r , i ,   K r g b , i d k ) V r g b , i ,
where h e a d r g b , i and h e a d i r , i represent the i -th head of RGB and IR, respectively. The output of each head is concatenated, and then fed into a series of operations including dropouts, residual connections, normalization, and feed-forward. The final output is the interacted patch embeddings x ˙ r g b e m b , x ˙ i r e m b R N × d .
Subsequently, two fully connected layers are employed to decode x ˙ r g b e m b , x ˙ i r e m b back into two interacted flattened patch sequences x ˙ r g b , x ˙ i r R N × d p a t c h These sequences are then reshaped and rearranged to obtain the interacted features with C channels. Finally, two 1 × 1 convolutional layers are used to restore the channel number, generating the interacted features F ˙ r g b , F ˙ i r R H × W × C .
Finally, the interacted features are element-wise added to the original input features to obtain the enhanced features, which can be expressed as follows:
F ^ r g b = F r g b + F ˙ r g b ,
F ^ i r = F i r + F ˙ i r ,

2.4. Feature Correction Fusion Module

The detailed information of the proposed feature correction fusion module is shown in Figure 4. It adaptively corrects and filters the complementary information between different modalities through spatial and channel correlations, achieving better cross-modal feature extraction and fusion. The order of feature correction affects the accuracy of the proposed model, as shown in Section 3.6. In this section, the specific structure is introduced in the order of spatial correction followed by channel correction.
(a)
Spatial correction first processes concatenated features:
The spatial feature correction module involves aligning features from different modalities to enhance detection accuracy, as shown in Figure 4. Let the visible light image be denoted as R G B I N , and the corresponding infrared image is denoted as I R I N , where infrared can be replaced by other complementary modalities such as SAR. First, the features from different modalities are concatenated. Subsequently, an MLP layer is employed to perform layer-wise nonlinear transformations and feature learning on the concatenated features, resulting in the feature map F S R H × W × 2 C . The MLP layer consists of three dilated convolutions, ReLU activation functions, and a 1 × 1 convolution layer, as shown in Equation (7). Next, the Sigmoid function is used to change the output range of F S to [0, 1], and it is further divided into two spatial weight maps, ω r g b R H × W × C and ω i r R H × W × C , as shown in Equation (8).
F S = M L P ( 2 C , C ) ( C o m c a t ( R G B I N , I R I N ) ) ,
M L P = σ ( C o n v ( C o n c a t ( r a t e 2 Y C , r a t e 5 Y C , r a t e 8 Y C ) ) ) ,
ω r g b S , ω i r S = S p i l t ( σ ( F S ) ) ,
where rate2, rate5, and rate8 denote the dilated convolution operations with dilation rates of 2, 5, and 8, respectively; σ represents the Sigmoid function.
By multiplying the obtained spatial correction feature map ω i r with the corresponding input I R I N , we achieve the spatial correction of the input R G B I N . Subsequently, this corrected feature is added to R G B I N to obtain the spatially corrected feature map R G B C O R S . Furthermore, a trainable parameter α ( 0 α 1 ) is introduced to control the dynamic weights of spatial and channel corrections. During the model training process, this parameter automatically adjusts based on the contributions of the two correction offsets to detection accuracy, thereby dynamically learning the optimal fusion ratio of the two sets of correction features. This enhances the model’s adaptability and robustness. The correction process can be represented by the following equation:
R G B C O R S = R G B I N + α ω i r S · I R I N ,
I R C O R S = I R I N + α ω r g b S · R G B I N ,
where R G B C O R S and I R C O R S represent the visible light and infrared feature maps that have undergone spatial correction, respectively; ω r g b S and ω i r S denote the spatial correction offset weights; and α signifies a learnable parameter during the spatial correction phase.
(b)
Channel correction then performs global optimization:
Channel-level feature correction is utilized to further correct global information after spatial-level feature correction has been applied to the local information, in order to optimize the overall consistency of the features. The features R G B C O R S and I R C O R S , which have undergone spatial-level feature correction, are first concatenated to obtain X C R H × W × 2 C . For complex cross-modal features, a comprehensive pooling strategy is employed, including global average pooling to capture overall average information, global max pooling to capture globally salient features, and global standard deviation pooling to focus on feature variations and distribution information. This combination of various pooling operations can provide a more comprehensive and rich feature representation, enabling the model to adapt to different types of data and distributions. Subsequently, the global feature vectors obtained from these three pooling operations are concatenated to form a feature map Y C of size R 6 C . Next, an MLP layer is applied to obtain F C R 2 C , and a Sigmoid activation function and slicing operation are used to obtain channel weight maps ω r g b C R C and ω i r C R C that contain global feature information. The aforementioned process can be represented as follows:
X C = C o n c a t ( R G B C O R S , I R C O R S ) ,
Y C = C o n c a t ( A v g ( X C ) , M a x ( X C ) , S t d ( X C ) ) ,
F C = M L P ( 6 C , 2 C ) ( Y C ) ,
M L P = L i n e a r ( C m i d , C o u t ) ( R e L U ( L i n e a r ( C i n , C o u t ) ( · ) ) ) ,
ω r g b C , ω i r C = S p i l t ( σ ( F C ) ) ,
where L i n e a r ( C i n , C o u t ) ( · ) represents a linear layer with input channels C i n and output channels C o u t ,   R e L U ( · ) denotes the ReLU activation function, and σ ( · ) represents the Sigmoid function.
In Equation (14), the MLP is composed of two linear layers and a rectifying linear unit; A v g ( · ) , M a x ( · ) , and S t d ( · ) represent global average pooling, global max pooling, and global standard deviation pooling, respectively; and σ denotes the Sigmoid function.
Similarly to spatial correction, channel feature correction can be represented in the following form:
R G B C O R S C = R G B C O R S + ( 1 α ) · ω I R C · I R C O R S ,
I R C O R S C = I R C O R S + ( 1 α ) · ω R G B C · R G B C O R S ,
where R G B C O R S and I R C O R S denote the visible light and infrared feature maps that have been spatially corrected, respectively. The terms R G B C O R S C and I R C O R S C refer to the feature maps of visible light and infrared that have undergone both spatial and channel corrections. The parameter 1 α represents a learnable parameter in the channel correction phase, where α in Equation (16) is the same as in Equation (9), governing the relative significance of the spatial and channel correction phases. The parameters ω I R C and ω R G B C signify the weights assigned to the channel correction offsets for the infrared and visible light modalities, respectively.
Finally, the features that have undergone both spatial and channel corrections are added element-wise to obtain the ultimately corrected and fused features, which can be represented as follows:
O u t p u t = R G B C O R S C + I R C O R S C ,

3. Experiments

3.1. Datasets

All experiments were evaluated on three datasets: LLVIP [33], OGSOD, and Drone Vehicle.
(a)
LLVIP. The LLVIP dataset is a highly challenging multi-spectral dataset for pedestrian detection under low-light conditions. Collected in low-light environments, this dataset makes accurate pedestrian detection in the RGB modality extremely challenging. The dataset comprises 12,025 pairs of aligned RGB-T images in the training set and 3463 pairs of images in the validation set, with each image having a resolution of 1024 × 1280.
(b)
OGSOD. The OGSOD dataset is a recently released optical-SAR paired dataset for cross-modal remote sensing object detection. It includes a training set with 14,665 image pairs and a test set with 3666 image pairs, containing over 48,000 instance annotations in total. All images have a size of 256 × 256. Three categories are annotated, including bridges, tanks, and ports.
(c)
Drone Vehicle. The Drone Vehicle dataset is a large-scale aerial optical-infrared dataset captured by drones, containing 28,439 image pairs and 953,087 vehicle annotation instances. This dataset covers a variety of scenes, including urban roads, residential areas, parking lots, and varying lighting conditions from day to night. The dataset is divided into five vehicle categories (cars, buses, trucks, vans, and cargo vehicles) and provides rich oriented bounding box annotations. To facilitate processing, the white borders of the images were removed, and the image size was uniformly adjusted to 640 × 512.

3.2. Implementation Details

This paper employs the Stochastic Gradient Descent (SGD) optimizer with an initial learning rate of 1 × 10−2, a momentum of 0.937, and a weight decay of 0.0005. A cosine annealing learning rate scheduler is applied, reducing the learning rate to 1 × 10−6 over 200 epochs. All models were trained for 200 epochs on a single NVIDIA RTX 3090 GPU with a batch size of 16. To achieve better performance, the YOLO model pre-trained on the COCO dataset [34] was utilized for weight initialization. Regarding data augmentation, the mosaic method, which combines four training images into one, was employed.

3.3. Evaluation Metrics

All models were evaluated using three object detection metrics introduced by MS-COCO: mean Average Precision (mAP), mAP at IoU = 0.5 (mAP50), and mAP at IoU = 0.75 (mAP75).
m A P = 1 n i = 0 n A P i = 1 n i = 0 n 0 1 P i ( r ) d r ,
where
A P i = 0 1 P i r d r = 0 1 P r e c i s i o n   d R e c a l l = 0 1 T P T P + F P d T P T P + F N
TP means true positive, which is when a predicted box by detectors and the ground truth (GT) meet the intersection over union (IoU) threshold; otherwise, it will be considered as a false positive (FP). False negative (FN) means there is a true object, but the detector does not find it. Equation (20) indicates that AP is the integral of the Precision–Recall Curve (PRC) for each category. mAP50 computes the mean of all the AP values for all categories at IoU = 0.50 in Equation (19). Similarly, mAP75 calculates the mean at IoU = 0.75. mAP is the primary challenge metric, which can be formulated as the mean at IoU = 0.50:0.05:0.95. Obviously, it is much stricter than the other two metrics.

3.4. Comparison with Existing Methods

To validate the performance advantages of the proposed cross-modal remote sensing object detection model under around-the-clock conditions and in complex scenarios, this section conducts comparative experiments on the LLVIP, OGSOD, and Drone Vehicle datasets against uni-modal and state-of-the-art cross-modal methods. These datasets cover a variety of scenes including traffic cameras, satellite images, and drone perspectives, ensuring the comprehensiveness and objectivity of the model evaluation due to their diverse viewpoints and environmental conditions. Overall, our method achieves significant improvements in mAP compared to both uni-modal methods and existing cross-modal approaches, with specific gains of 4.4% on LLVIP (66.3% vs. 61.9%), 12.3% on OGSOD (58.6% vs. 46.3%), and 8.3% on Drone Vehicle (71.7% vs. 63.4%).
On the LLVIP dataset, as shown in Table 1, the proposed cross-modal remote sensing object detection model demonstrates significant performance improvements compared to both uni-modal object detection models and existing cross-modal approaches. Specifically, compared to uni-modal methods, the cross-modal remote sensing object detection model achieved an improvement in mAP50 ranging from 2.7% to 12%. When compared to advanced methods such as CCFINet and CFT, the cross-modal remote sensing object detection model surpassed them with an mAP50 of 97.9% (compared to 97.6% and 97.5% for CCFINet and CFT, respectively), and achieved an overall mAP of 66.3%.
On the OGSOD dataset, as illustrated in Table 2, the cross-modal remote sensing object detection model further solidifies its leading position in cross-modal object detection. Under the RGB+SAR modality, the model achieved an mAP50 of 94.5% and an overall mAP of 58.6%, surpassing advanced methods such as CoLD and GI Imitation, which achieved mAP50 scores of 87.6% and 87.1%, respectively.
On the Drone Vehicle dataset, as demonstrated in Table 3, the proposed method was evaluated and compared with the state-of-the-art approaches. The model exhibited notable performance improvements under conditions with smaller objects and more complex environments. Specifically, compared to uni-modal methods, the cross-modal remote sensing object detection model improved the mAP50 by 15.6% to 32.18%. When compared to advanced methods such as COMO and CFT, the cross-modal remote sensing object detection model outperformed them with an mAP50 of 86.3% (compared to 86.1% and 84.3% for COMO and CFT, respectively), and achieved an overall mAP of 71.7%.
We compare the model size, computational complexity, and inference speed of our method against baselines on the Drone Vehicle dataset, as shown in Table 4. The metrics include parameter count (Params, M), FLOPs (@640, G), and FPS (Hz), measured on a single NVIDIA RTX 3090 GPU. Our model achieves low FLOPs of 14.36 G, the lowest among compared methods (e.g., 18.45 G for DaFF, 176.00 G for GM-DETR), and a high FPS of 226.2 Hz, surpassing most baselines (e.g., 217.4 Hz for ICAFusion, 208.3 Hz for CMADet). While our parameter count (68.43 M) is higher than several baselines (e.g., 4.83 M for SuperYOLO, 20.15 M for ICAFusion), the additional parameters in MSFEM’s multi-branch convolutions (3 × 3, 5 × 5, 7 × 7), FIM’s eight attention heads, and FCFM’s fusion layers contribute to a significant mAP improvement (e.g., 71.7% vs. 63.4% for the dual-stream baseline on Drone Vehicle).

3.5. Visualization Results

In addition to the comparisons with existing methods, this paper also presents the detection results of the cross-modal remote sensing object detection model on the three datasets in Figure 5, Figure 6 and Figure 7 to visually demonstrate the effectiveness of the proposed method. These figures illustrate the model’s ability to accurately detect objects under various conditions, highlighting its robustness and the advantages of integrating information from multiple modalities.
(a). Detection results on the LLVIP dataset: Figure 5 shows the detection outcomes for LLVIP, with rows from top to bottom representing CFT, ICAFusion, CrossFormer, AMFusion, and the cross-modal remote sensing object detection network. It is evident that the first three methods exhibit various degrees of missed detections due to insufficient extraction of complementary features. In contrast, the cross-modal remote sensing object detection network significantly improves the detection results. This confirms the effectiveness of the method proposed in this paper for cross-modal object detection images.
(b). Detection results on the Drone Vehicle dataset: Figure 6 displays the performance of the proposed cross-modal remote sensing object detection model alongside other methods, arranged from top to bottom as CFT, COMO, ICAFusion, GM-DETR, and the cross-modal remote sensing object detection network. Due to the impact of lighting and extreme weather conditions, several methods fail to fully exploit the advantages between infrared and visible light, leading to varying degrees of missed and false detections. In contrast, this paper’s method enhances the interaction between visible light and infrared features, effectively integrating complementary information, significantly reducing missed and false detections, and thus outperforming the comparative methods. This also demonstrates the effectiveness of the approach proposed in this paper.
(c). Detection results on the OGSOD dataset: Figure 7 illustrates the visual detection performance of several methods on the OGSOD dataset. From top to bottom, these are LD, CoLD, and the cross-modal remote sensing object detection method. Due to the noise and scattering effects in SAR images, there is inconsistency and redundancy between modalities, leading to varying degrees of missed and false detections in the results of other methods. The feature interaction module and the feature correction fusion module proposed in this paper aim to reduce modal inconsistency and redundancy, enabling the network to achieve better classification and more true positives (TPs).
(d). Visualization results of the multi-scale feature extraction module: As shown in Figure 8, this paper demonstrates the performance of the Multi-Scale Feature Extraction Module (MSFEM) in cross-modal remote sensing object detection through feature map visualization. To verify the effectiveness of this module, feature maps were visualized in Figure 8 and compared with the performance of the C3 module. The third column of Figure 8 shows the feature maps of the C3 module, indicating that it fails to adequately extract the edges and texture features of the objects, especially in multi-scale object scenarios. In contrast, the MSFEM accurately extracts key features and clearly delineates the contours of the objects through multi-scale feature fusion, particularly under conditions of low visibility or high noise, demonstrating its effectiveness in cross-modal object detection. Under daylight conditions, the MSFEM shows a significant ability to capture detailed features and textures, which is crucial for object detection in well-lit environments. The feature maps generated by the MSFEM exhibit higher clarity and better-defined object boundaries compared to those from the C3 module, which often struggles with noise and less defined edges. Under low-light (nighttime) conditions, the MSFEM’s multi-scale feature fusion is particularly effective. It compensates for the loss of visible-light features by leveraging information from multiple scales, thereby enhancing the overall feature quality. The feature maps from the MSFEM under low-light conditions show better contrast and less noise, which is essential for accurate object detection in such challenging environments.
(e). Visualization results of the feature interaction module: As shown in Figure 9, this paper demonstrates the performance of the feature interaction module in cross-modal remote sensing object detection through feature map visualization. To verify the effectiveness of this module, feature maps were visualized in Figure 9 and compared with the performance of the CFT module. Where x r g b and x i r represent the feature maps of visible light and infrared images, respectively; y r g b C F T and y i r C F T denote the complementary feature maps by the CFT module; y r g b F I M and y i r F I M indicate the complementary feature maps by the feature interaction module. From Figure 9, it can be observed that while the CFT module ( y r g b C F T and y i r C F T ) can extract basic complementary features between visible light and infrared light, it has limitations when modeling long-distance dependencies between modalities. Specifically, the feature maps of the CFT module lack clarity in distinguishing object edges from background noise, leading to insufficient precision in complementary feature extraction. Additionally, in complex scenarios, the CFT module is susceptible to noise and background interference, further reducing the efficiency and accuracy of feature extraction. In contrast, this module effectively models the long-distance relationships between visible light and infrared modalities by converting features into patch-embedded sequences and utilizing a cross-attention mechanism, significantly enhancing the quality of feature maps. The cross-attention mechanism dynamically captures complementary information between visible light and infrared, enhancing the clarity of feature maps at object edges and contours. Furthermore, the module optimizes the semantic information of feature maps through a feed-forward network, further highlighting object features. Through the synergistic effect of these mechanisms, the module effectively extracts complementary features between visible light and infrared light, suppresses background noise and modal redundancy, and significantly improves the efficiency and accuracy of feature extraction.
(f). Visualization results of the feature correction fusion module: To verify the effectiveness of the feature correction fusion module (FCFM), this paper visualizes the feature maps in Figure 10 and compares its performance with the element-wise addition method. Figure 10 shows a comparison of feature maps between the element-wise addition and the FCFM. The first row displays the visible light and infrared images under well-lit night conditions, while the second row shows the same under low-light (extremely dark) conditions. The third column presents the feature maps from element-wise addition, and the fourth column shows the feature maps from the FCFM. Under well-lit night conditions, the feature maps from element-wise addition can extract basic edges and texture features but struggle with spatial inconsistencies and feature redundancies between visible light and infrared modalities. This leads to less clarity in distinguishing object edges from background noise, resulting in insufficient feature extraction accuracy. In contrast, the FCFM significantly improves the quality of feature maps through spatial and channel-level feature correction mechanisms. Spatial feature correction aligns visible light and infrared features, effectively eliminating local spatial differences and enhancing the clarity of object features. Channel-level feature correction optimizes the semantic information of feature maps through global optimization, further highlighting object features. These mechanisms work synergistically to enhance feature extraction accuracy and robustness under well-lit night conditions. Under low-light (extremely dark) conditions, the element-wise addition method is particularly prone to noise and background interference in complex night scenes, further reducing detection accuracy. The feature maps from element-wise addition lack clarity and are often overwhelmed by background noise. The FCFM, however, demonstrates significant improvements in feature map quality under extremely low-visibility conditions. Spatial feature correction effectively aligns visible light and infrared features, reducing local spatial differences and enhancing object feature clarity. Channel-level feature correction optimizes the semantic information of feature maps through global optimization, further highlighting object features. Through these mechanisms, the FCFM effectively suppresses background noise and modal redundancy, significantly improving detection accuracy in complex night scenes.

3.6. Ablation Study

(1) In this section, various ablation studies were conducted to examine the effects of the multi-scale feature extraction module, the feature interaction module, and the feature correction fusion module on the LLVIP, OGSOD, and Drone Vehicle datasets. In Table 5, the detection performance on different datasets (LLVIP, OGSOD, and Drone Vehicle) is compared. The best records and improvements are indicated by bold and upward arrows (⬆), respectively. On the LLVIP dataset, the mAP values for RGB-only and IR-only are 50.0% and 61.9%, respectively, with IR-only detection results surpassing RGB-only, a trend that is also observed in the Drone Vehicle dataset. However, on the OGSOD dataset, the performance of RGB-only is slightly better than that of SAR-only (mAP: 6.1%⬆). This may be due to the presence of numerous low-light scenarios in LLVIP and Drone Vehicle, leading to the loss of effective object areas, whereas the OGSOD dataset contains relatively more SAR noise and lacks features that are as easily recognizable and identifiable as those in RGB and IR. Furthermore, comparing the mAP of IR-only and dual-stream baselines, the simple dual-stream network fails to fully exploit the inherent complementarity between different modalities. Moreover, these rudimentary approaches may increase the difficulty of network learning and exacerbate the imbalance between modalities, leading to a decrease in performance. From Table 5, it can be seen that after adopting the proposed MSFEM (multi-scale feature extraction module), FIM (feature interaction module), and FCFM (feature correction fusion module), the detection model’s performance has improved on all three datasets. Particularly for the OGSOD dataset, the evaluation metric mAP75 increased by 23.99%, and the mAP improved by 12.3%.
Table 5 demonstrates that the MSFEM enhances modality alignment and feature compensation, improving mAP by up to 4.9% (Drone Vehicle). The FIM addresses noise suppression and inter-modal dependencies, boosting mAP75 by 3.2% on OGSOD. The FCFM improves dynamic scene adaptation, with a 2.0% mAP gain on Drone Vehicle. These results validate our framework’s ability to address the limitations of prior work in modality alignment, feature degradation, and dynamic scenes.
(2) To further explore the impact of the feature correction fusion module on the cross-modal remote sensing object detection model, this paper conducted ablation experiments on the order of spatial and channel dimension feature corrections. “Channel → Spatial” indicates that channel dimension correction is performed first, followed by spatial dimension correction, while “Spatial → Channel” is the reverse order. As shown in Table 6, compared to performing channel dimension feature correction first, performing spatial feature correction first can increase the mAP by 2.8%. The paper suggests that this is because the intuitive local appearance differences between different modalities are more significant than the abstract global semantic differences. Therefore, to achieve the best experimental results, it is recommended to first eliminate the larger local spatial differences through spatial dimension correction and then eliminate the smaller global semantic differences through channel dimension correction.

4. Conclusions

This paper has presented a comprehensive solution to the critical challenges in cross-modal remote sensing object detection, specifically addressing (1) modality disparities between visible light, infrared, and SAR data, and (2) feature degradation under adverse weather conditions. Our framework’s three core innovations—the multi-scale feature extraction module (MSFEM) for robust feature representation, the feature interaction module (FIM) for cross-modal alignment, and the feature correction fusion module (FCFM) for dynamic feature optimization—collectively achieve state-of-the-art performance across multiple benchmark datasets. Experimental results demonstrate significant improvements, including a 4.4% increase in mAP on the LLVIP dataset (66.3% vs. 61.9% for IR-only baselines), a 12.3% higher mAP on OGSOD (58.6%), and a 71.7% mAP on the challenging Drone Vehicle dataset, validating the framework’s effectiveness in handling both modality differences and weather-induced feature deterioration. The proposed architecture’s modular design offers practical advantages for real-world deployment in 24/7 surveillance systems, military reconnaissance, and emergency response scenarios where reliability under varying conditions is paramount.
Despite these advancements, three limitations suggest promising directions for future research: (1) Computational efficiency—while our model achieves real-time performance on standard GPUs, further optimization through lightweight architectures or neural architecture search could enable deployment on edge devices with strict power constraints. (2) Extreme small object detection—performance on sub-10-pixel objects in dense scenes (common in satellite imagery) could be enhanced through dedicated high-resolution feature preservation techniques. (3) Generalized multi-modal learning—extending the framework to incorporate additional modalities like LiDAR or hyperspectral data while maintaining parameter efficiency requires novel fusion strategies. Potential solutions include developing attention-based modality gating mechanisms or investigating dynamic network architectures that adaptively activate relevant modality pathways. These improvements would further solidify the framework’s position as a versatile solution for mission-critical around-the-clock detection systems. Additionally, future work will further explore the robustness of the model, particularly in practical RGB-IR/SAR misregistration scenarios (e.g., ±1–10 pixel shifts, small rotations, and scale jitter) and partial modality dropout (i.e., the availability of only a single modality at test time). Additionally, we will analyze the performance of the model as a function of the level of perturbation and provide qualitative examples. These investigations will help to better understand the adaptability and stability of the model under different conditions. Meanwhile, we plan to combine three modalities (e.g., visible light, infrared, and SAR) to further demonstrate the robustness and versatility of our method.

Author Contributions

Conceptualization, Y.Q. and S.Y.; methodology, Y.Q. and S.Y.; validation, J.C. and M.Z.; formal analysis, S.Y.; investigation, S.Y.; resources, Y.Q.; data curation, J.Z. and J.C.; writing—original draft preparation, Y.Q. and S.Y.; writing—review and editing, H.Z. and X.L.; visualization, S.Y.; supervision, H.Z. and X.L.; project administration, Y.Q. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the Hebei Province Central Guiding Local Science and Technology Development Fund Project under grant number 236Z4901G.

Data Availability Statement

The LLVIP dataset (https://drive.google.com/file/d/1VTlT3Y7e1h-Zsne4zahjx5q0TK2ClMVv/view, accessed on 10 August 2025), OGSOD dataset (OGSOD-1.0.zip—Google Drive), and Drone Vehicle dataset (VisDrone/DroneVehicle: Drone-based RGB-Infrared Cross-Modality Vehicle Detection via Uncertainty-Aware Learning (github.com, accessed on 10 Auust 2025)) were utilized.

Acknowledgments

During the preparation of this manuscript/study, we utilized a language model to generate the initial draft of the literature review. By inputting relevant keywords and research directions, we obtained a preliminary literature review text. This enabled us to swiftly grasp the current state of research in our field and saved us a considerable amount of time. Additionally, a text-editing tool was employed during the manuscript writing process for text polishing and grammar checking. It assisted us in ensuring the smoothness and accuracy of the text by identifying and correcting grammatical and spelling errors. The authors have reviewed and edited the content as needed and take full responsibility for the final publication.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards real-time object detection with region proposal networks. IEEE Trans. Pattern Anal. Mach. Intell. 2016, 39, 1137–1149. [Google Scholar] [CrossRef] [PubMed]
  2. Jenkins, M.; Franklin, K.A.; Nicoll, M.A.; Cole, N.C.; Ruhomaun, K.; Tatayah, V.; Mackiewicz, M. Improving object detection for time-lapse imagery using temporal features in wildlife monitoring. Sensors 2024, 24, 8002. [Google Scholar] [CrossRef] [PubMed]
  3. Xie, M.; Tang, Q.; Tian, Y.; Feng, X.; Shi, H.; Hao, W. DCN-YOLO: A Small-Object Detection Paradigm for Remote Sensing Imagery Leveraging Dilated Convolutional Networks. Sensors 2025, 25, 2241. [Google Scholar] [CrossRef]
  4. Duan, K.; Bai, S.; Xie, L.; Qi, H.; Huang, Q.; Tian, Q. CenterNet: Keypoint triplets for object detection. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019. [Google Scholar]
  5. Kieu, M.; Bagdanov, A.D.; Bertini, M.; Del Bimbo, A. Domain adaptation for privacy-preserving pedestrian detection in thermal imagery. In Proceedings of the International Conference on Image Analysis and Processing, Trento, Italy, 9–13 September 2019. [Google Scholar]
  6. Li, C.; Song, D.; Tong, R.; Tang, M. Multispectral Pedestrian Detection via Simultaneous Detection and Segmentation. In Proceedings of the British Machine Vision Conference 2018, Newcastle, UK, 3–6 September 2018. [Google Scholar]
  7. Kieu, M.; Bagdanov, A.D.; Bertini, M.; Del Bimbo, A. Task-conditioned domain adaptation for pedestrian detection in thermal imagery. In Proceedings of the European Conference on Computer Vision (ECCV), Glasgow, UK, 23–28 August 2020. [Google Scholar]
  8. Wang, Q.; Chi, Y.; Shen, T.; Song, J.; Zhang, Z.; Zhu, Y. Improving RGB-Infrared Object Detection by Reducing Cross-Modality Redundancy. Remote Sens. 2022, 14, 2020. [Google Scholar] [CrossRef]
  9. Zhang, L.; Liu, Z.; Zhang, S.; Yang, X.; Qiao, H.; Huang, K.; Hussain, A. Cross-modality interactive attention network for multispectral pedestrian detection. Inf. Fusion 2019, 50, 20–29. [Google Scholar] [CrossRef]
  10. Guan, D.; Cao, Y.; Yang, J.; Cao, Y.; Yang, M.Y. Fusion of multispectral data through illumination-aware deep neural networks for pedestrian detection. Inf. Fusion 2019, 50, 148–157. [Google Scholar] [CrossRef]
  11. Sun, Y.; Cao, B.; Zhu, P.; Hu, Q. Drone-based RGB-infrared cross-modality vehicle detection via uncertainty-aware learning. IEEE Trans. Circuits Syst. Video Technol. 2022, 32, 6700–6713. [Google Scholar] [CrossRef]
  12. Song, K.; Xue, X.; Wen, H.; Ji, Y.; Yan, Y.; Meng, Q. Misaligned visible-thermal object detection: A drone-based benchmark and baseline. IEEE Trans. Intell. Veh. 2024, 9, 7449–7460. [Google Scholar] [CrossRef]
  13. Yuan, M.; Wei, X. C2Former: Calibrated and complementary transformer for RGB-infrared object detection. IEEE Trans. Geosci. Remote Sens. 2024, 62, 5403712. [Google Scholar] [CrossRef]
  14. Wang, H.; Wang, C.; Fu, Q.; Si, B.; Zhang, D.; Kou, R.; Yu, Y.; Feng, C. YOLOFIV: Object detection algorithm for around-the-clock aerial remote sensing images by fusing infrared and visible features. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2024, 17, 15269–15287. [Google Scholar] [CrossRef]
  15. Qingyun, F.; Dapeng, H.; Zhaokui, W. Cross-Modality Fusion Transformer for Multispectral Object Detection. arXiv 2021, arXiv:2111.00273. [Google Scholar]
  16. Zhang, H.; Fromont, E.; Lefevre, S.; Avignon, B. Multispectral fusion for object detection with cyclic fuse-and-refine blocks. In Proceedings of the 2020 IEEE International Conference on Image Processing (ICIP), Abu Dhabi, United Arab Emirates, 25–28 October 2020. [Google Scholar]
  17. Chen, Y.T.; Shi, J.; Ye, Z.; Mertz, C.; Ramanan, D.; Kong, S. Multimodal object detection via probabilistic ensembling. In Proceedings of the European Conference on Computer Vision (ECCV), Tel Aviv, Israel, 23–27 October 2022. [Google Scholar]
  18. Cao, Y.; Bin, J.; Hamari, J.; Blasch, E.; Liu, Z. Multimodal object detection by channel switching and spatial attention. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 18–24 June 2023. [Google Scholar]
  19. Bao, W.; Huang, M.; Hu, J.; Xiang, X. Dual-Dynamic Cross-Modal Interaction Network for Multimodal Remote Sensing Object Detection. IEEE Trans. Geosci. Remote Sens. 2025, 63, 5401013. [Google Scholar] [CrossRef]
  20. Bae, J.H.; Yeo, D.; Yim, J.; Kim, N.S.; Pyo, C.S.; Kim, J. Densely distilled flow-based knowledge transfer in teacher-student framework for image classification. IEEE Trans. Image Process. 2020, 29, 5698–5710. [Google Scholar] [CrossRef] [PubMed]
  21. Zhao, B.; Cui, Q.; Song, R.; Qiu, Y.; Liang, J. Decoupled knowledge distillation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18–24 June 2022. [Google Scholar]
  22. Zheng, Z.; Ye, R.; Wang, P.; Ren, D.; Zuo, W.; Hou, Q.; Cheng, M.M. Localization distillation for dense object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18–24 June 2022. [Google Scholar]
  23. Wang, C.; Ruan, R.; Zhao, Z.; Li, C.; Tang, J. Category-oriented localization distillation for SAR object detection and a unified benchmark. IEEE Trans. Geosci. Remote Sens. 2023, 61, 5211314. [Google Scholar] [CrossRef]
  24. Hu, J.; Shen, L.; Sun, G. Squeeze-and-Excitation Networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–23 June 2018. [Google Scholar]
  25. Vaswani, A.; Shazeer, N.M.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is All You Need. In Proceedings of the 31st International Conference on Neural Information Processing Systems (NeurIPS), Long Beach, CA, USA, 4–9 December 2017. [Google Scholar]
  26. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale. In Proceedings of the 9th International Conference on Learning Representations (ICLR), Virtual Event, Austria, 3–7 May 2021. [Google Scholar]
  27. Liu, Z.; Lin, Y.; Cao, Y.; Hu, H.; Wei, Y.; Zhang, Z.; Guo, B. Swin Transformer: Hierarchical Vision Transformer Using Shifted Windows. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, QC, Canada, 11–17 October 2021. [Google Scholar]
  28. Lou, M.; Zhang, S.; Zhou, H.Y.; Yang, S.; Wu, C.; Yu, Y. TransXNet: Learning Both Global and Local Dynamics with a Dual Dynamic Token Mixer for Visual Recognition. IEEE Trans. Neural Netw. Learn. Syst. 2025, 36, 11534–11547. [Google Scholar] [CrossRef]
  29. Zhou, K.; Chen, L.; Cao, X. Improving Multispectral Pedestrian Detection by Addressing Modality Imbalance Problems. In Proceedings of the 16th European Conference on Computer Vision (ECCV), Glasgow, UK, 23–28 August 2020; Springer International Publishing: Cham, Switzerland, 2020. [Google Scholar]
  30. Yang, X.; Qian, Y.; Zhu, H.; Wang, C.; Yang, M. BAANet: Learning Bi-directional Adaptive Attention Gates for Multispectral Pedestrian Detection. In Proceedings of the 2022 International Conference on Robotics and Automation (ICRA), Philadelphia, PA, USA, 23–27 May 2022. [Google Scholar]
  31. Yan, C.; Zhang, H.; Li, X.; Yang, Y.; Yuan, D. Cross-modality complementary information fusion for multispectral pedestrian detection. Neural Comput. Appl. 2023, 35, 10361–10386. [Google Scholar] [CrossRef]
  32. Lee, S.; Park, J.; Park, J. CrossFormer: Cross-guided attention for multi-modal object detection. Pattern Recognit. Lett. 2024, 179, 144–150. [Google Scholar] [CrossRef]
  33. Jia, X.; Zhu, C.; Li, M.; Tang, W.; Zhou, W. LLVIP: A visible-infrared paired dataset for low-light vision. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Republic of Korea, 11–17 October 2021. [Google Scholar]
  34. Lin, T.-Y.; Maire, M.; Belongie, S.; Hays, J.; Perona, P.; Ramanan, D.; Dollár, P.; Zitnick, C.L. Microsoft COCO: Common Objects in Context. In Proceedings of the 13th European Conference on Computer Vision (ECCV), Zurich, Switzerland, 6–12 September 2014. [Google Scholar]
  35. Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.Y.; Berg, A.C. SSD: Single Shot Multibox Detector. In Proceedings of the 14th European Conference on Computer Vision (ECCV), Amsterdam, The Netherlands, 11–14 September 2016. [Google Scholar]
  36. Jocher, G. YOLOv5 by Ultralytics. 2020. Available online: https://github.com/ultralytics/yolov5 (accessed on 28 February 2023).
  37. Qi, Y.; Yang, S.; Jia, Z.; Song, Y.; Zhu, J.; Liu, X.; Zheng, H. An Investigation of Infrared Small Target Detection by Using the SPT–YOLO Technique. Technologies 2025, 13, 40. [Google Scholar] [CrossRef]
  38. Zhang, H.; Fromont, E.; Lefèvre, S.; Avignon, B. Guided Attentive Feature Fusion for Multispectral Pedestrian Detection. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision (WACV), Waikoloa, HI, USA, 5–9 January 2021. [Google Scholar]
  39. Tang, L.; Xiang, X.; Zhang, H.; Gong, M.; Ma, J. DIVFusion: Darkness-free Infrared and Visible Image Fusion. Inf. Fusion 2023, 91, 477–493. [Google Scholar] [CrossRef]
  40. Shen, J.; Chen, Y.; Liu, Y.; Zuo, X.; Fan, H.; Yang, W. ICAFusion: Iterative Cross-Attention Guided Feature Fusion for Multispectral Object Detection. Pattern Recognit. 2024, 145, 109913. [Google Scholar] [CrossRef]
  41. Pei, S.; Lin, J.; Liu, W.; Zhao, T.; Lin, C.W. Beyond Night Visibility: Adaptive Multi-scale Fusion of Infrared and Visible Images. arXiv 2024, arXiv:2403.01083. [Google Scholar]
  42. Li, Z.; Pan, H.; Zhang, K.; Wang, Y.; Yu, F. MambaDFuse: A Mamba-based Dual-phase Model for Multi-modality Image Fusion. arXiv 2024, arXiv:2404.08406. [Google Scholar]
  43. Yang, Z.; Liu, S.; Hu, H.; Wang, L.; Lin, S. RepPoints: Point Set Representation for Object Detection. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Republic of Korea, 27 October–2 November 2019. [Google Scholar]
  44. Zhang, S.; Chi, C.; Yao, Y.; Lei, Z.; Li, S.Z. Bridging the Gap Between Anchor-Based and Anchor-Free Detection via Adaptive Training Sample Selection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 14–19 June 2020. [Google Scholar]
  45. He, Y.; Zhu, C.; Wang, J.; Savvides, M.; Zhang, X. Bounding Box Regression with Uncertainty for Accurate Object Detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019. [Google Scholar]
  46. Sun, P.; Zhang, R.; Jiang, Y.; Kong, T.; Xu, C.; Zhan, W.; Tomizuka, M.; Li, L.; Yuan, Z.; Wang, C.; et al. Sparse R-CNN: End-to-End Object Detection with Learnable Proposals. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021. [Google Scholar]
  47. Zand, M.; Etemad, A.; Greenspan, M. ObjectBox: From Centers to Boxes for Anchor-Free Object Detection. In Proceedings of the European Conference on Computer Vision (ECCV), Tel Aviv, Israel, 23–27 October 2022; Springer Nature: Cham, Switzerland, 2022. [Google Scholar]
  48. Wang, C.Y.; Bochkovskiy, A.; Liao, H.Y.M. YOLOv7: Trainable Bag-of-Freebies Sets New State-of-the-Art for Real-Time Object Detectors. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Vancouver, BC, Canada, 17–24 June 2023. [Google Scholar]
  49. Han, J.; Ding, J.; Li, J.; Xia, G.S. Align Deep Features for Oriented Object Detection. IEEE Trans. Geosci. Remote Sens. 2021, 60, 5602511. [Google Scholar] [CrossRef]
  50. Ding, J.; Xue, N.; Long, Y.; Xia, G.S.; Lu, Q. Learning RoI Transformer for Oriented Object Detection in Aerial Images. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019. [Google Scholar]
  51. Xie, X.; Cheng, G.; Wang, J.; Yao, X.; Han, J. Oriented R-CNN for Object Detection. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Republic of Korea, 11–17 October 2021. [Google Scholar]
  52. Althoupety, A.; Wang, L.Y.; Feng, W.C.; Rekabdar, B. DAFF: Dual Attentive Feature Fusion for Multispectral Pedestrian Detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 17–24 June 2024. [Google Scholar]
  53. Liu, C.; Ma, X.; Yang, X.; Zhang, Y.; Dong, Y. COMO: Cross-Mamba Interaction and Offset-Guided Fusion for Multimodal Object Detection. Inf. Fusion 2026, 125, 103414. [Google Scholar] [CrossRef]
  54. Xiao, Y.; Meng, F.; Wu, Q.; Xu, L.; He, M.; Li, H. GM-DETR: Generalized Multispectral Detection Transformer with Efficient Fusion Encoder for Visible-Infrared Detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 17–24 June 2024. [Google Scholar]
Figure 1. Architecture of the cross-modal remote sensing object detection network.
Figure 1. Architecture of the cross-modal remote sensing object detection network.
Sensors 25 04964 g001
Figure 2. Multi-scale feature extraction module.
Figure 2. Multi-scale feature extraction module.
Sensors 25 04964 g002
Figure 3. Feature interaction module.
Figure 3. Feature interaction module.
Sensors 25 04964 g003
Figure 4. Feature correction fusion module.
Figure 4. Feature correction fusion module.
Sensors 25 04964 g004
Figure 5. Visualization of detection results on LLVIP. We compare cross-modal remote object detection with four other methods. Blue bounding boxes denote TPs; red bounding boxes denote FNs.
Figure 5. Visualization of detection results on LLVIP. We compare cross-modal remote object detection with four other methods. Blue bounding boxes denote TPs; red bounding boxes denote FNs.
Sensors 25 04964 g005
Figure 6. Visualization of detection results on Drone Vehicle. We compare cross-modal remote object detection with four other methods. Blue bounding boxes denote TPs; red bounding boxes denote FNs; red bounding boxes denote FPs.
Figure 6. Visualization of detection results on Drone Vehicle. We compare cross-modal remote object detection with four other methods. Blue bounding boxes denote TPs; red bounding boxes denote FNs; red bounding boxes denote FPs.
Sensors 25 04964 g006
Figure 7. Visualization of detection results on Drone Vehicle. We compare cross-modal remote object detection with four other methods. Green, orange and blue bounding boxes denote TPs; red bounding boxes denote FNs; yellow bounding boxes denote FPs.
Figure 7. Visualization of detection results on Drone Vehicle. We compare cross-modal remote object detection with four other methods. Green, orange and blue bounding boxes denote TPs; red bounding boxes denote FNs; yellow bounding boxes denote FPs.
Sensors 25 04964 g007
Figure 8. A comparison of feature map visualizations between the C3 (CSP Bottleneck with 3 convolutions) module and the multi-scale feature extraction module, with RGB and IR scenes divided into day and night. The third column displays the visualization results of the C3 module, while the fourth column shows the visualization results of the multi-scale feature extraction module.
Figure 8. A comparison of feature map visualizations between the C3 (CSP Bottleneck with 3 convolutions) module and the multi-scale feature extraction module, with RGB and IR scenes divided into day and night. The third column displays the visualization results of the C3 module, while the fourth column shows the visualization results of the multi-scale feature extraction module.
Sensors 25 04964 g008
Figure 9. The visualization results of the feature interaction module. Where x r g b and x i r represent the visualization results of visible light and infrared images, respectively; y r g b C F T and y i r C F T denote the complementary feature visualization results of infrared and visible light by the CFT module; y r g b F I M and y i r F I M indicate the complementary feature visualization results of infrared and visible light by the feature interaction module.
Figure 9. The visualization results of the feature interaction module. Where x r g b and x i r represent the visualization results of visible light and infrared images, respectively; y r g b C F T and y i r C F T denote the complementary feature visualization results of infrared and visible light by the CFT module; y r g b F I M and y i r F I M indicate the complementary feature visualization results of infrared and visible light by the feature interaction module.
Sensors 25 04964 g009
Figure 10. Presents the comparative feature map visualization results between element-wise addition and the feature correction fusion module. The first and second columns represent the infrared and visible light images of night scenes, respectively; the third column shows the visualization results of the feature maps from Add (element-wise addition); the fourth column displays the visualization results of the feature maps from the feature correction fusion module.
Figure 10. Presents the comparative feature map visualization results between element-wise addition and the feature correction fusion module. The first and second columns represent the infrared and visible light images of night scenes, respectively; the third column shows the visualization results of the feature maps from Add (element-wise addition); the fourth column displays the visualization results of the feature maps from the feature correction fusion module.
Sensors 25 04964 g010
Table 1. Comparison of performances on LLVIP datasets measured. Bold red indicates the best; bold blue indicates the second best.
Table 1. Comparison of performances on LLVIP datasets measured. Bold red indicates the best; bold blue indicates the second best.
ModalityMethodmAP50mAP75mAP
IRFaster R-CNN92.648.850.7
SSD [35]90.257.953.5
YOLOv389.753.452.8
YOLOv5 [36]94.672.261.9
YOLOv8 [37]95.2-62.1
RGBFaster R-CNN88.845.747.5
SSD82.631.839.8
YOLOv385.937.943.4
YOLOv590.851.950.0
YOLOv891.9-54.0
RGB+IRHalfway Fusion [38]91.460.155.1
GAFF94.060.255.8
CFT97.572.963.6
ProbEN93.450.251.5
CSSA94.366.659.2
DIVFusion [39]89.859.952.0
CCFINet97.672.664.1
ICAFusion [40]96.971.562.2
CrossFormer97.476.366.1
AMFusion [41]96.570.160.3
MambaDFuse [42]95.666.459.3
Ours97.976.566.3
Note: Methods without citations are already introduced in the introduction or related work sections. Methods with citations are referenced here to provide proper attribution to the original work.
Table 2. Comparison of performances on OGSOD datasets measured. Bold red indicates the best; bold blue indicates the second best.
Table 2. Comparison of performances on OGSOD datasets measured. Bold red indicates the best; bold blue indicates the second best.
ModalityMethodYearOil TankBridgeHarbormAP50mAP
SARRetinaNet201717.373.395.362.036.7
YOLOv3201832.476.097.068.539.5
ATSS [43]201926.278.096.366.838.6
YOLOv5202057.787.297.980.946.3
RepPoints [44]202030.470.895.465.537.9
Generalized Focal [45]202033.472.896.567.641.8
Sparse R-CNN [46]202128.773.894.265.638.7
Object BOX [47]202251.082.496.576.640.1
YOLOv7 [48]202259.779.898.179.245.1
RGB+SARKD202060.388.498.882.648.4
GI Imitation202169.292.999.187.155.9
DKD202262.562.598.783.449.8
LD202265.765.798.384.551.9
CoLD202369.869.899.587.656.7
Ours202594.999.399.794.558.6
Note: Methods without citations are already introduced in the introduction or related work sections. Methods with citations are referenced here to provide proper attribution to the original work.
Table 3. Comparison of performances on Drone Vehicle datasets measured. Bold red indicates the best; bold blue indicates the second best.
Table 3. Comparison of performances on Drone Vehicle datasets measured. Bold red indicates the best; bold blue indicates the second best.
ModalityMethodCarTruckFreight CarBusVanmAP0.5mAP
IRYOLOv590.059.560.889.553.870.7-
YOLOV887.9654.5517.0586.1624.9054.12-
S2A-Net [49]89.751.050.289.044.064.867.5
Faster R-CNN89.453.548.387.042.664.2-
ROI Transformer [50]89.651.053.488.944.565.570.3
Oriented R-CNN [51]89.653.953.989.241.065.567.0
RGBYOLOv578.655.343.887.146.0-62.1
YOLOv870.1254.5518.282.0522.3848.57-
S2A-Net79.950.036.282.837.557.361.0
Faster R-CNN79.049.037.277.037.0-55.9
ROI Transformer61.655.142.285.544.861.661.6
Oriented R-CNN80.355.442.186.846.962.360.8
RGB+IRCFT98.575.068.582.397.384.361.9
DDCI91.078.966.190.765.578.4-
CMADet98.270.466.478.396.882.059.5
DaFF [52]92.258.958.271.994.475.145.5
YOLOFIV95.8964.2334.5791.5637.2964.71-
C2Former90.268.364.489.858.574.270.0
ICAFusion96.146.434.057.192.265.144.0
COMO [53]98.678.971.584.197.486.165.5
UA-CMDet88.673.056.088.354.872.264.0
GM-DETR [54]92.475.364.980.890.880.855.9
Ours98.783.175.897.397.686.371.7
Note: Methods without citations are already introduced in the introduction or related work sections. Methods with citations are referenced here to provide proper attribution to the original work.
Table 4. Model size, computation cost, and detection speed statistics for different models on the Drone Vehicle dataset.
Table 4. Model size, computation cost, and detection speed statistics for different models on the Drone Vehicle dataset.
MethodsParams (M)Flops@640 (G)FPS (Hz)
CFT44.7617.9291.74
SuperYOLO4.8317.9889.4
GHOST7.0620.36125.6
MFPT47.6534.5551.2
ICAFusion20.1514.93217.4
GM-DETR70.00176.0045.6
DaFF45.4218.4585.2
CMADet33.3316.86208.3
Ours68.4314.36226.2
Table 5. Comparisons of performances with different datasets in terms of mAP50, mAP75, and mAP.
Table 5. Comparisons of performances with different datasets in terms of mAP50, mAP75, and mAP.
DatasetModalityMethodmAP50mAP75mAP
LLVIPIRBaseline94.672.261.9
RGB90.851.950.0
RGB+IR+ Two Stream95.871.462.3
+ MSFEM96.373.563.5
+ MSFEM + FIM97.175.364.7
+ MSFEM + FIM + FCFM97.9 (3.3⬆)76.5 (4.3⬆)66.3 (4.4⬆)
OGSODRGBBaseline80.950.3146.3
SAR78.735.540.2
RGB+SAR+ Two Stream81.458.345.63
+ MSFEM86.265.049.78
+ MSFEM + FIM88.968.253.4
+ MSFEM + FIM + FCFM94.5 (13.6⬆)74.3 (23.99⬆)58.6 (12.3⬆)
Drone VehicleIRBaseline80.858.760.2
RGB74.646.946.7
RGB+IR+ Two Stream81.364.663.4
+ MSFEM84.169.5868.3
+ MSFEM + FIM85.372.4869.7
+ MSFEM + FIM + FCFM86.3 (5.5⬆)76.4 (17.7⬆)71.7 (8.1⬆)
Table 6. Ablation study of spatial vs. channel correction sequence.
Table 6. Ablation study of spatial vs. channel correction sequence.
FCFMmAP50mAP75mAP
Channel → Spatial96.473.363.5
Spatial → Channel97.976.566.3
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Qi, Y.; Yang, S.; Chen, J.; Zhang, M.; Zhu, J.; Liu, X.; Zheng, H. A Modality Alignment and Fusion-Based Method for Around-the-Clock Remote Sensing Object Detection. Sensors 2025, 25, 4964. https://doi.org/10.3390/s25164964

AMA Style

Qi Y, Yang S, Chen J, Zhang M, Zhu J, Liu X, Zheng H. A Modality Alignment and Fusion-Based Method for Around-the-Clock Remote Sensing Object Detection. Sensors. 2025; 25(16):4964. https://doi.org/10.3390/s25164964

Chicago/Turabian Style

Qi, Yongjun, Shaohua Yang, Jiahao Chen, Meng Zhang, Jie Zhu, Xin Liu, and Hongxing Zheng. 2025. "A Modality Alignment and Fusion-Based Method for Around-the-Clock Remote Sensing Object Detection" Sensors 25, no. 16: 4964. https://doi.org/10.3390/s25164964

APA Style

Qi, Y., Yang, S., Chen, J., Zhang, M., Zhu, J., Liu, X., & Zheng, H. (2025). A Modality Alignment and Fusion-Based Method for Around-the-Clock Remote Sensing Object Detection. Sensors, 25(16), 4964. https://doi.org/10.3390/s25164964

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop