Previous Article in Journal
DRADG: A Dynamic Risk-Adaptive Data Governance Framework for Modern Digital Ecosystems
Previous Article in Special Issue
Enhancing Weakly Supervised Video Anomaly Detection with Object-Centric Features
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

DCAM-DETR: Dual Cross-Attention Mamba Detection Transformer for RGB–Infrared Anti-UAV Detection

1
School of Public Security, Guangxi Police College, Nanning 530028, China
2
School of Cyberspace Security (School of Cryptology), Hainan University, Haikou 570228, China
*
Author to whom correspondence should be addressed.
Information 2026, 17(1), 103; https://doi.org/10.3390/info17010103
Submission received: 16 December 2025 / Revised: 6 January 2026 / Accepted: 14 January 2026 / Published: 19 January 2026
(This article belongs to the Special Issue Computer Vision for Security Applications, 2nd Edition)

Abstract

The proliferation of unmanned aerial vehicles (UAVs) poses escalating security threats across critical infrastructures, necessitating robust real-time detection systems. Existing vision-based methods predominantly rely on single-modality data and exhibit significant performance degradation under challenging scenarios. To address these limitations, we propose DCAM-DETR, a novel multimodal detection framework that fuses RGB and thermal infrared modalities through an enhanced RT-DETR architecture integrated with state space models. Our approach introduces four innovations: (1) a MobileMamba backbone leveraging selective state space models for efficient long-range dependency modeling with linear complexity O ( n ) ; (2) Cross-Dimensional Attention (CDA) and Cross-Path Attention (CPA) modules capturing intermodal correlations across spatial and channel dimensions; (3) an Adaptive Feature Fusion Module (AFFM) dynamically calibrating multimodal feature contributions; and (4) a Dual-Attention Decoupling Module (DADM) enhancing detection head discrimination for small targets. Experiments on Anti-UAV300 demonstrate state-of-the-art performance with 94.7% mAP@0.5 and 78.3% mAP@0.5:0.95 at 42 FPS. Extended evaluations on FLIR-ADAS and KAIST datasets validate the generalization capacity across diverse scenarios.

Graphical Abstract

1. Introduction

The exponential proliferation of unmanned aerial vehicles (UAVs) has fundamentally transformed numerous civilian and commercial domains, encompassing aerial cinematography, precision agriculture, logistics delivery, infrastructure inspection, and emergency response operations [1,2]. However, this technological democratization has concurrently engendered unprecedented security vulnerabilities, as malicious actors increasingly exploit UAVs for unauthorized surveillance, contraband smuggling, critical infrastructure disruption, and potential terrorist activities [3,4]. The Federal Aviation Administration (FAA) documented over 2800 UAV-related security incidents in 2023 alone, representing a 47% increase from the previous year [5]. Consequently, developing robust, efficient, and real-time anti-UAV detection systems has emerged as an imperative research priority for safeguarding public safety and national security. This challenge aligns with broader efforts in safety-critical IoT sensing and intelligent transportation systems, where AI models must support real-time, safety-related decisions in connected systems [6].
Conventional anti-UAV detection methodologies predominantly employ radar systems, acoustic sensors, or radio frequency (RF) analysis [7]. While these approaches demonstrate efficacy in specific operational contexts, they suffer from inherent limitations including prohibitive deployment costs, susceptibility to electromagnetic interference, limited spatial resolution, and difficulty in discriminating UAVs from avian species or other airborne objects [8]. Vision-based detection methods leveraging deep learning have emerged as compelling alternatives, offering cost-effectiveness, rich semantic information extraction, and precise spatial localization capabilities [9,10]. Nevertheless, UAV detection in unconstrained real-world environments presents formidable challenges: UAVs typically manifest as diminutive targets with limited discriminative visual features, frequently operating against complex heterogeneous backgrounds under varying illumination conditions, atmospheric perturbations, and partial occlusions [11,12].
Single-modality detection approaches exhibit fundamental limitations in addressing these challenges comprehensively. RGB cameras, while providing rich texture and chromatic information under adequate illumination, suffer severe performance degradation in low-light, nighttime, or adverse weather conditions [13]. Conversely, thermal infrared sensors capture thermal radiation signatures that remain robust across varying illumination conditions but lack sufficient textural detail for precise localization and classification [14]. This complementary nature of RGB and infrared modalities motivates multimodal fusion strategies that leverage the synergistic advantages of both sensing paradigms [15,16].
Recent advances in multimodal fusion for object detection have explored diverse architectural paradigms. Early fusion approaches employ simple concatenation or element-wise operations [17], which inadequately capture complex intermodal dependencies. Intermediate fusion methods leverage attention mechanisms [18,19] or adversarial learning [20] to enhance fusion quality, yet frequently incur substantial computational overhead or exhibit limited receptive field coverage. Late fusion strategies combine detection outputs from modality-specific detectors [21], potentially sacrificing fine-grained complementary information. Furthermore, most existing detection frameworks are constructed upon convolutional neural networks (CNNs) or standard vision transformers [22,23], which either lack global context modeling capability or impose quadratic computational complexity with respect to input resolution, limiting their applicability to high-resolution imagery or real-time deployment scenarios [24,25].
Despite these advances, current fusion-based UAV detectors exhibit several critical limitations. First, most methods employ fixed fusion strategies that apply uniform weights regardless of scene conditions, failing to adapt to varying modality reliability across different environments. Second, existing cross-modal attention mechanisms typically operate on a single dimension (spatial or channel), inadequately capturing the complex intermodal dependencies that span both dimensions. Third, transformer-based fusion approaches incur quadratic computational complexity O ( n 2 ) , limiting their applicability to real-time scenarios. Fourth, detection heads in current frameworks lack specialized mechanisms for small target enhancement, which is crucial for UAV detection where targets often occupy minimal pixel areas.
The recently proposed RT-DETR [26] framework has demonstrated remarkable performance in real-time object detection by synergistically combining the efficiency of YOLO-series detectors with the end-to-end detection paradigm of DETR [27]. Concurrently, state space models (SSMs), particularly the Mamba architecture [28], have emerged as a transformative paradigm for sequence modeling, offering linear computational complexity while preserving robust long-range dependency modeling capabilities. Vision Mamba variants [29,30] have successfully adapted SSMs to computer vision tasks, demonstrating competitive performance with vision transformers while achieving substantially improved computational efficiency. However, these architectural advances have not been systematically investigated for multimodal fusion in challenging detection scenarios such as anti-UAV applications.
To bridge this critical gap, we propose DCAM-DETR, a novel multimodal detection framework that synergistically integrates state space models with the RT-DETR architecture for robust real-time anti-UAV detection. Our approach introduces four principal innovations specifically designed for multimodal small target detection. First, we replace the conventional CNN backbone with a MobileMamba architecture that processes RGB and infrared streams through parallel Mamba encoders, enabling efficient global context aggregation with linear complexity O ( n ) compared to the quadratic complexity O ( n 2 ) of standard self-attention mechanisms. Second, we design Cross-Dimensional Attention (CDA) and Cross-Path Attention (CPA) modules that explicitly model intermodal correlations across both spatial and channel dimensions, facilitating fine-grained feature alignment and complementary information extraction through parallel attention streams.
Third, we introduce an Adaptive Feature Fusion Module (AFFM) that dynamically calibrates fusion weights based on scene characteristics and modality reliability through learned gating mechanisms, ensuring optimal information integration across diverse environmental conditions. Fourth, we propose a Dual-Attention Decoupling Module (DADM) in the detection head that employs hierarchical dilated convolutions with parallel spatial and channel attention streams to enhance feature discrimination for small target detection. These components operate synergistically to achieve superior detection performance while maintaining real-time inference capability.
The principal contributions of this work are summarized as follows:
  • We propose DCAM-DETR, a novel multimodal detection framework that integrates Mamba-based state space models with the RT-DETR architecture for efficient anti-UAV detection, achieving linear computational complexity while maintaining robust global context modeling capability through selective scan mechanisms.
  • We design Cross-Dimensional Attention (CDA) and Cross-Path Attention (CPA) modules that explicitly capture intermodal correlations across spatial and channel dimensions through parallel attention streams, enabling fine-grained multimodal feature alignment and complementary information extraction.
  • We introduce an Adaptive Feature Fusion Module (AFFM) that dynamically weights multimodal features based on scene-adaptive gating mechanisms, and a Dual-Attention Decoupling Module (DADM) that enhances detection head performance through hierarchical dilated convolutions and attention decomposition.
  • Comprehensive experiments on Anti-UAV300, FLIR-ADAS, and KAIST datasets demonstrate that DCAM-DETR achieves state-of-the-art performance with 94.7% mAP@0.5 on Anti-UAV300, outperforming existing methods by substantial margins while maintaining real-time inference speed of 42 FPS.
The remainder of this paper is organized as follows. Section 2 reviews related work on UAV detection, multimodal fusion, and state space models. Section 3 presents the proposed DCAM-DETR framework in detail. Section 4 describes experimental setup and presents comprehensive results. Section 5 provides in-depth analysis and discussion. Finally, Section 6 concludes the paper with future research directions.

2. Related Work

2.1. Vision-Based UAV Detection

Vision-based UAV detection has garnered substantial research attention owing to its cost-effectiveness and rich semantic information extraction capabilities. Early methodologies employed classical computer vision techniques including background subtraction, optical flow estimation, and hand-crafted feature descriptors [12,31], which exhibited limited robustness against complex backgrounds and varying environmental conditions. The advent of deep learning has fundamentally transformed this domain, with CNN-based detectors such as Faster R-CNN [32], the YOLO series [33,34,35,36,37,38,39], and RetinaNet [40] being extensively adopted for UAV detection tasks.
However, UAV detection presents distinctive challenges that conventional object detectors inadequately address. UAVs typically manifest as small targets occupying minimal pixel areas, exhibit limited discriminative visual features, and demonstrate high-speed motion characteristics [9,10]. Recent investigations have explored specialized architectures for UAV detection. Huang et al. [10] introduced the Anti-UAV410 benchmark with thermal infrared imagery and proposed customized tracking schemes incorporating temporal consistency constraints. Jiang et al. [9] presented the Anti-UAV dataset and systematically evaluated various tracking algorithms under diverse environmental conditions. Zhao et al. [41] proposed multi-scale feature aggregation networks specifically designed for small UAV detection. Nevertheless, most existing methods focus on single-modality detection, fundamentally limiting their robustness across diverse environmental conditions. Our work addresses this limitation by leveraging multimodal fusion with efficient state space models.

2.2. Multimodal Fusion for Object Detection

Multimodal fusion, particularly RGB–infrared fusion, has demonstrated significant advantages in challenging detection scenarios by exploiting complementary information from heterogeneous sensors [13,42]. Fusion strategies can be categorized into early fusion, intermediate fusion, and late fusion paradigms based on the integration stage within the detection pipeline [43].
Early fusion approaches concatenate or combine raw inputs or low-level features from different modalities [17,44]. While computationally efficient, these methods inadequately capture complex intermodal dependencies and may introduce noise from modality-specific artifacts. Intermediate fusion methods integrate features at multiple abstraction levels, leveraging attention mechanisms [19,45,46], cross-modal transformers [47,48], or learnable fusion strategies [49,50] to adaptively combine multimodal information. Chen et al. [15] proposed probabilistic ensembling for multimodal object detection, while Wang et al. [16] introduced Mask-Guided Mamba Fusion for enhanced UAV detection in complex environments. Liu et al. [14] developed target-aware dual adversarial learning for infrared-visible fusion.
Late fusion strategies combine detection outputs from modality-specific detectors through ensemble techniques [21]. While preserving modality-specific representations, these approaches potentially sacrifice fine-grained complementary information available at intermediate feature levels. Recent works have explored transformer-based architectures for multimodal fusion [14,51], achieving improved performance but often suffering from quadratic computational complexity that limits real-time applicability. Our approach leverages Mamba’s linear complexity while achieving superior fusion quality through cross-dimensional attention mechanisms.

2.3. State Space Models and Mamba

State space models (SSMs) have emerged as efficient alternatives to transformers for sequence modeling, offering linear computational complexity while maintaining strong long-range dependency modeling capabilities [52,53]. The foundational structured state space sequence model (S4) [53] demonstrated that properly parameterized SSMs can effectively capture long-range dependencies in sequential data. The theoretical foundations of state space modeling for handling long-term correlations have been extensively studied in various domains, including network systems with generalized nonlinear stochastic processes [54]. The recently proposed Mamba architecture [28] introduces selective state spaces with input-dependent dynamics, enabling the model to dynamically adjust its behavior based on input content.
The core innovation of Mamba lies in its selective scan mechanism, which allows the model to focus on relevant information while filtering out irrelevant details through input-dependent parameterization of the state space matrices. This selectivity enables Mamba to achieve performance comparable to transformers while maintaining linear computational complexity with respect to sequence length. The continuous-time state space formulation is discretized for practical implementation:
A ¯ = exp ( Δ A ) , B ¯ = ( Δ A ) 1 ( exp ( Δ A ) I ) · Δ B
where Δ represents the discretization step size, and A , B are the state space matrices.
Vision Mamba [29] and VMamba [30] have successfully adapted Mamba to computer vision tasks, demonstrating competitive performance with vision transformers on image classification, object detection, and semantic segmentation. These works typically employ bidirectional or cross-scanning strategies to capture spatial relationships in 2D images. LocalMamba [55] introduced local scanning windows to enhance local feature extraction. MambaVision [56] proposed hybrid architectures combining Mamba with convolutional layers. However, existing vision Mamba architectures have not been systematically explored for multimodal fusion scenarios. Our MobileMamba backbone extends these ideas by introducing parallel Mamba encoders for different modalities with cross-modal interaction mechanisms. Table 1 summarizes the key differences between DCAM-DETR and existing Mamba-based methods.

2.4. Detection Transformers

DETR [27] pioneered end-to-end object detection using transformers, eliminating hand-crafted components such as anchor generation and non-maximum suppression through bipartite matching and set prediction. Subsequent works addressed DETR’s limitations in convergence speed and computational efficiency. Deformable DETR [57] introduced deformable attention mechanisms that attend to a sparse set of sampling points, significantly improving convergence and performance. Conditional DETR [58] and DAB-DETR [59] enhanced query formulation for improved detection accuracy.
RT-DETR [26] represents a significant advancement in real-time detection transformers, introducing an efficient hybrid encoder that combines intra-scale feature interaction with cross-scale feature fusion, and uncertainty-minimal query selection for improved detection quality. RT-DETR achieves real-time performance while maintaining high accuracy, making it suitable for practical deployment scenarios. Our work builds upon RT-DETR’s efficient architecture while extending it to multimodal scenarios through the integration of MobileMamba backbone, cross-dimensional attention modules, and enhanced detection head components.

3. Methodology

3.1. Overall Architecture

Figure 1 illustrates the comprehensive architecture of our proposed DCAM-DETR framework. The detection pipeline comprises four principal components: (1) a MobileMamba backbone that processes RGB and infrared inputs through parallel state space model encoders with hierarchical feature extraction; (2) an Efficient Transformer Encoder enhanced with Cross-Dimensional Attention (CDA) and Cross-Path Attention (CPA) modules for multimodal feature fusion; (3) an Adaptive Feature Fusion Module (AFFM) that dynamically calibrates features from different scales and modalities; and (4) a detection head equipped with Dual-Attention Decoupling Module (DADM) for enhanced small target detection through the CCFF (Cross-scale Channel Feature Fusion) structure.
Given a spatially aligned RGB image I rgb R H × W × 3 and a thermal infrared image I ir R H × W × 1 , the MobileMamba backbone first extracts hierarchical feature representations at multiple scales. The spatial alignment is achieved through hardware calibration and software preprocessing: RGB and thermal cameras are mounted on a rigid platform with factory-calibrated intrinsic and extrinsic parameters, and we apply homography transformation followed by bilinear interpolation to resize both modalities to the same resolution ( 640 × 640 ). For each modality m { rgb , ir } , the backbone produces multi-scale features { F m ( i ) } i = 3 5 corresponding to feature pyramid levels P3, P4, and P5 with spatial resolutions of H 8 × W 8 , H 16 × W 16 , and H 32 × W 32 , respectively. These features are subsequently processed through the Efficient Transformer Encoder with CDA and CPA modules to capture intermodal correlations. The AFFM adaptively fuses multimodal features based on scene characteristics, producing unified representations for the detection head. Finally, the DADM-enhanced detection head generates bounding box predictions and class probabilities through the Uncertainty-minimal Query Selection mechanism.

3.2. MobileMamba Backbone

The MobileMamba backbone serves as the feature extraction foundation, processing RGB and infrared inputs through parallel Mamba encoders with shared architectural design but independent parameters. Unlike conventional CNN backbones that rely on local receptive fields with limited global context, or transformer-based backbones that incur quadratic computational complexity, MobileMamba leverages state space models to capture global dependencies with linear complexity O ( n ) , making it particularly suitable for detecting small targets that require extensive contextual information.

3.2.1. Selective State Space Model Formulation

The theoretical foundation of our backbone is the continuous-time state space model, which maps an input sequence x ( t ) R to an output sequence y ( t ) R through a latent state h ( t ) R N :
d h ( t ) d t = A h ( t ) + B x ( t ) , y ( t ) = C h ( t ) + D x ( t )
where A R N × N is the state transition matrix, B R N × 1 is the input projection matrix, C R 1 × N is the output projection matrix, and D R is the feedthrough coefficient. For practical implementation, this continuous system is discretized using the zero-order hold (ZOH) method with step size Δ :
A ¯ = exp ( Δ A ) , B ¯ = ( Δ A ) 1 ( exp ( Δ A ) I ) · Δ B
The discretized state space model operates as a recurrence:
h k = A ¯ h k 1 + B ¯ x k , y k = C h k + D x k
Intuitively, the discretization converts continuous-time dynamics into a form suitable for digital computation. The matrix A ¯ captures how the hidden state evolves over one time step, while B ¯ determines how new inputs are incorporated. The step size Δ controls temporal resolution.
The key innovation of Mamba is the selective mechanism that makes B , C , and Δ input-dependent:
B = Linear B ( x ) , C = Linear C ( x ) , Δ = softplus ( Linear Δ ( x ) + Broadcast ( p Δ ) )
where p Δ is a learnable parameter. This selectivity enables the model to dynamically decide which information to store in the hidden state (controlled by B ), which stored information to output (controlled by C ), and how quickly to update the state (controlled by Δ ), effectively filtering irrelevant information while preserving salient features.

3.2.2. SS2D Block for 2D Visual Processing

To adapt the 1D selective state space model for 2D visual data, we employ the SS2D (Selective Scan 2D) block illustrated in Figure 2. The SS2D block extends the selective scan mechanism to 2D spatial data through a cross-scan strategy that processes the feature map along four directions: left-to-right, right-to-left, top-to-bottom, and bottom-to-top.
Given an input feature map X R H × W × C , the scan expanding operation flattens the 2D spatial dimensions into four 1D sequences corresponding to different scanning directions:
X ( d ) = ScanExpand ( d ) ( X ) , d { lr , rl , tb , bt }
where X ( d ) R ( H · W ) × C represents the flattened sequence for direction d. Each sequence is independently processed through the S6 Block:
Y ( d ) = S 6 Block ( X ( d ) )
The outputs from all four directions are merged through a learnable fusion mechanism:
Y = ScanMerge ( Y ( lr ) , Y ( rl ) , Y ( tb ) , Y ( bt ) )
The cross-scan strategy addresses a fundamental challenge: 1D state space models process sequences sequentially, but images have 2D spatial structure. By scanning in four directions, each pixel aggregates information from all spatial directions, enabling comprehensive spatial context modeling.
This bidirectional cross-scanning strategy ensures comprehensive spatial context aggregation while maintaining linear computational complexity O ( H W C ) , compared to the quadratic complexity O ( H 2 W 2 C ) of standard self-attention mechanisms.

3.2.3. S6 Block Architecture

The S6 (Selective State Space Sequence) Block, illustrated in Figure 3, constitutes the core computational unit of our MobileMamba backbone. The S6 Block incorporates layer normalization, linear projections, depthwise convolution, SiLU activation, the SS2D selective scan operation, and residual connections for stable gradient flow.
The complete S6 Block operation is formulated as:
Z 1 = LN ( X ) Z 2 = SS2DBlock ( Z 1 ) + X Z 3 = LN ( Z 2 ) Y = FFN ( Z 3 ) + Z 2
The internal SS2D Block processing follows:
U = Linear ( Z 1 ) V = SiLU ( DWConv 3 × 3 ( Linear ( Z 1 ) ) ) W = SS2D ( V ) O = Linear ( LN ( W ) U )
where ⊙ denotes element-wise multiplication, and the depthwise convolution introduces local inductive bias complementing the global context modeling of the selective scan operation.
The MobileMamba backbone employs a hierarchical architecture with four stages, progressively downsampling spatial resolution while increasing channel dimensions. For RGB and infrared modalities, we utilize parallel Mamba encoders with shared architecture but independent parameters, enabling modality-specific feature learning while maintaining architectural consistency.

3.3. Cross-Dimensional Attention Modules

Effective multimodal fusion requires capturing complex intermodal dependencies that span both spatial and channel dimensions. Traditional fusion approaches, such as simple concatenation or element-wise operations, fail to model the intricate correlations between heterogeneous modalities. To address this limitation, we propose two complementary attention modules: Cross-Dimensional Attention (CDA) for spatial correlation modeling and Cross-Path Attention (CPA) for channel-wise dependency capture. These modules are integrated into the Efficient Transformer Encoder, operating in parallel to provide comprehensive multimodal feature alignment.

3.3.1. Cross-Dimensional Attention (CDA)

The CDA module is designed to capture spatial correlations between RGB and infrared features by learning position-dependent attention weights that identify corresponding regions across modalities. The key insight is that spatially aligned multimodal images share semantic correspondences at specific locations, and explicitly modeling these spatial relationships enhances feature alignment quality.
Given RGB features F rgb R H × W × C , CDA employs a multi-branch architecture combining spatial self-attention, frequency-domain analysis, and local feature extraction, as illustrated in the left panel of Figure 4. The spatial self-attention branch computes position-wise correlations through the scaled dot-product attention mechanism:
A spatial = Softmax Q s K s d k V s , where Q s , K s , V s = ϕ split ( W exp F rgb )
where W exp R 3 C × C is the channel expansion projection, ϕ split ( · ) partitions the expanded features into query, key, and value components, and d k = C is the scaling factor for numerical stability.
To capture complementary frequency-domain information that may be obscured in the spatial domain, we incorporate a frequency projection branch based on the discrete Fourier transform:
F freq = F 1 W freq F ( F rgb )
where F ( · ) and F 1 ( · ) denote the 2D FFT and inverse FFT operations, and W freq represents learnable frequency-domain filters. This branch captures periodic patterns and global structural information that complement the local spatial attention.
The final CDA output integrates spatial attention, frequency features, and local context through adaptive gating:
F CDA = W o ( F sp + F freq ) σ ( W s F rgb ) + V local σ ( W c ( F sp F freq ) )
where F sp is the spatially projected attention output, V local = DWConv 3 × 3 ( W 1 F rgb ) captures local context through depthwise separable convolution, σ ( · ) denotes the sigmoid activation, and ⊕ represents global average pooling followed by channel projection.

3.3.2. Cross-Path Attention (CPA)

While CDA focuses on spatial correlations, the CPA module captures channel-wise dependencies that encode semantic relationships between feature channels across modalities. Different channels in deep features correspond to different semantic concepts, and modeling their intermodal correlations enables more effective information exchange.
Operating on infrared features F ir R H × W × C , CPA computes channel attention by treating each channel as a token and applying self-attention along the channel dimension:
A channel = Softmax Q c K c H W V c , where Q c , K c , V c R H W × C
where the query, key, and value matrices are obtained by reshaping the linearly projected infrared features from R H × W × 3 C to R H W × C for each component. The scaling factor H W normalizes the attention scores based on the spatial dimension.
The CPA output combines channel-attended features with spatially-gated local information:
F CPA = W o F cp σ ( W c F ir ) + V ir σ ( W sp F cp )
where F cp is the channel-projected attention output, V ir = DWConv 3 × 3 ( W 1 F ir ) provides local infrared context, and W sp R 1 × C generates spatial attention weights through channel-wise projection.
The dual attention design ensures comprehensive multimodal feature alignment: CDA identifies where to attend by modeling spatial correspondences, while CPA determines what to attend by capturing semantic channel correlations. Together, they provide complementary perspectives for effective RGB–infrared fusion.

3.3.3. Why Cross-Dimensional Attention Helps for Small UAVs

Small UAVs present unique detection challenges due to their limited pixel coverage and lack of discriminative local features. Our cross-dimensional attention design specifically addresses these challenges through three mechanisms. First, CDA’s spatial attention aggregates contextual information from surrounding regions, providing discriminative cues from the broader scene context when local features are insufficient. Second, CPA’s channel attention identifies which feature channels from each modality are most informative for small targets—for small UAVs, thermal signatures in IR often provide stronger signals than RGB texture, and CPA learns to emphasize these channels adaptively. Third, the parallel attention streams capture correlations at different granularities, enabling the model to leverage both fine-grained local features and coarse global patterns essential for small target detection.

3.4. Adaptive Feature Fusion Module (AFFM)

A fundamental challenge in multimodal fusion is the varying reliability of different modalities across environmental conditions. In low-light scenarios, infrared features provide more discriminative information, while in well-illuminated scenes with rich texture, RGB features may be more informative. Fixed fusion strategies that apply uniform weights fail to adapt to these variations, leading to suboptimal performance. To address this challenge, we propose the Adaptive Feature Fusion Module (AFFM), illustrated in Figure 5, that dynamically calibrates multimodal feature contributions through learned scene-dependent gating mechanisms.
Given RGB features F g R H × W × C and infrared features F l R H × W × C at a specific pyramid level, AFFM computes content-aware gating weights that reflect the local reliability of each modality. The gating mechanism is formulated as a learnable soft attention over modality contributions:
W m = σ BN W g m F m , m { g , l }
where W g m R C × C × 1 × 1 denotes the 1 × 1 convolutional kernel for modality m, ∗ represents convolution, and σ ( · ) is the sigmoid activation that constrains weights to [ 0 , 1 ] .
The fusion process decomposes multimodal information into complementary and distinctive components. The complementary fusion captures shared information through weighted combination:
F comp = W g F g + ( 1 W l ) F l W l F l + ( 1 W g ) F g
where 1 denotes an all-ones tensor of appropriate dimensions. This formulation ensures that when one modality is unreliable (low gate value), the fusion relies more heavily on the other modality.
The distinctive fusion captures modality-specific information through weighted subtraction, which highlights unique features present in only one modality:
F dist = W g F g W l F l W l F l W g F g
The final fused representation aggregates both components through channel concatenation and convolutional refinement:
Z = ϕ refine F comp ; F dist = ReLU BN W r F comp ; F dist
where W r R C × 4 C × 3 × 3 is the refinement convolution kernel that reduces the concatenated 4 C channels back to C while capturing local spatial context. This design enables AFFM to adaptively emphasize reliable modality information while preserving distinctive features from both modalities.

3.5. Dual-Attention Decoupling Module (DADM)

Small target detection presents unique challenges that standard detection heads inadequately address. UAVs often occupy minimal pixel areas and lack sufficient discriminative features when considered in isolation. However, they become more distinguishable when contextual information from surrounding regions is incorporated. Furthermore, the detection head must simultaneously handle classification and localization tasks, which benefit from different feature characteristics. To address these challenges, we propose the Dual-Attention Decoupling Module (DADM), shown in Figure 6, that enhances detection capability through hierarchical multi-scale context aggregation and decoupled spatial-channel attention processing.
Given input features F R H × W × C , DADM first constructs a multi-scale feature pyramid through cascaded dilated convolutions with progressively increasing receptive fields. The dilated convolution operation with dilation rate r is defined as:
( W r F ) ( p ) = k W ( k ) · F ( p + r · k )
where p denotes the spatial position, k indexes the kernel elements, and r controls the spacing between kernel elements. This formulation enables exponentially expanding receptive fields without increasing parameters.
The hierarchical feature extraction applies cascaded dilated convolutions with rates r = [ 1 , 2 , 4 , 4 ] :
F i = W 1 1 F i = 1 W i r i F i 1 i { 2 , 3 , 4 }
where W i R C × C × 3 × 3 are learnable convolution kernels. The multi-scale features are aggregated through element-wise summation to preserve information from all receptive field scales:
F ms = i = 1 4 F i = F 1 + F 2 + F 3 + F 4
The Dual-Attention Module (DAM) computes spatial attention through a factorized query–key interaction that reduces computational complexity from O ( H 2 W 2 ) to O ( H W ) . The attention mechanism is formulated as:
A DAM = ϕ H ( W q F ms ) ϕ W ( W k F ms )
where W q , W k R 1 × C × 1 × 1 are pointwise convolution kernels, ϕ H : R H × W R H × 1 and ϕ W : R H × W R 1 × W are learnable projection functions implemented as fully connected layers after spatial reshaping, and ⊗ denotes outer product.
The final DADM output combines attention-weighted features with the original multi-scale representation through a residual connection:
F DADM = W 3 σ ( W a A DAM ) F ms + W p F ms
where W 3 R C × C × 3 × 3 and W p R C × C × 1 × 1 are convolution kernels for the attention and residual branches, respectively, and W a projects the attention map to match the feature channel dimension. This design enables DADM to focus on discriminative spatial regions while preserving rich multi-scale contextual information essential for small target detection.

3.6. Loss Function

Following the RT-DETR framework, we employ a composite loss function combining classification loss, bounding box regression loss, and Generalized IoU (GIoU) loss:
L total = λ cls L cls + λ bbox L bbox + λ giou L giou
The classification loss employs focal loss [40] to address class imbalance:
L cls = α t ( 1 p t ) γ log ( p t )
where p t is the predicted probability for the ground-truth class, α t is the balancing factor, and γ is the focusing parameter.
The bounding box regression loss uses L1 loss:
L bbox = i = 1 4 | b i b ^ i |
where b i and b ^ i are the ground-truth and predicted bounding box coordinates, respectively.
The GIoU loss [60] provides scale-invariant localization supervision:
L giou = 1 GIoU ( b , b ^ ) = 1 IoU ( b , b ^ ) | C ( b b ^ ) | | C |
where C is the smallest enclosing box containing both b and b ^ .
The loss weights are set to λ cls = 2.0 , λ bbox = 5.0 , and λ giou = 2.0 following standard practice.

4. Experiments

4.1. Datasets

We evaluate DCAM-DETR on three multimodal datasets spanning different application domains to comprehensively assess detection performance and generalization capability.
Anti-UAV300 Dataset. Our primary evaluation is conducted on the Anti-UAV300 dataset [9], a challenging multimodal benchmark specifically designed for UAV detection. The dataset comprises 300 video sequences with spatially aligned RGB and thermal infrared imagery, captured under diverse environmental conditions including varying illumination (daytime, dusk, nighttime), weather changes (clear, cloudy, foggy), and complex backgrounds (urban, rural, sky). The dataset contains approximately 45,000 frame pairs with bounding box annotations for UAVs of various sizes (small, medium, large) and types (quadcopter, fixed-wing, helicopter). We follow the standard train/validation/test split of 200/50/50 sequences, ensuring no overlap in capture sessions between splits.
FLIR-ADAS Dataset. To evaluate generalization capability across application domains, we additionally test on the FLIR-ADAS dataset [61], which contains 10,228 thermal images and corresponding RGB images captured from automotive scenarios. The dataset includes annotations for vehicles, pedestrians, bicycles, and other road users. While not specifically designed for UAV detection, this dataset provides diverse multimodal data for validating our fusion approach in different environmental contexts.
KAIST Multispectral Dataset. We further validate on the KAIST Multispectral Pedestrian Detection Benchmark [62], containing 95,328 RGB-thermal image pairs captured in various traffic scenarios across daytime and nighttime conditions. This dataset tests our method’s robustness across different application domains and illumination conditions.

4.2. Implementation Details

Our model is implemented in PyTorch 2.0 and trained on 4 NVIDIA A100 GPUs with 80 GB memory each. We employ the AdamW optimizer [63] with an initial learning rate of 1 × 10 4 , weight decay of 1 × 10 4 , and effective batch size of 16 (4 per GPU). The learning rate follows a cosine annealing schedule [64] over 100 epochs with 5 epochs of linear warmup. Input images are resized to 640 × 640 resolution with standard data augmentation including random horizontal flipping (probability 0.5), color jittering (brightness, contrast, saturation, hue), mosaic augmentation [34], and mixup [65] with probability 0.15.
The MobileMamba backbone is initialized with ImageNet-1K pre-trained weights when available. The Efficient Transformer Encoder uses 6 encoder layers with 8 attention heads and hidden dimension 256. The detection head employs 6 decoder layers with 300 object queries. For the DADM module, dilated convolution rates are set to r { 1 , 2 , 4 , 4 } for the four hierarchical levels. All experiments use mixed-precision training (FP16) for computational efficiency.

4.3. Evaluation Metrics

We adopt standard object detection metrics following the COCO evaluation protocol for comprehensive performance assessment. The primary metrics include Mean Average Precision (mAP) at different Intersection over Union (IoU) thresholds, inference speed, and model complexity measures.
Intersection over Union (IoU). IoU quantifies the overlap between the predicted bounding box b ^ and the ground-truth bounding box b, serving as the fundamental criterion for determining detection correctness:
IoU ( b , b ^ ) = | b b ^ | | b b ^ | = Area of Intersection Area of Union
A detection is considered a true positive if IoU ( b , b ^ ) τ , where τ is the IoU threshold.
Precision and Recall. For a given IoU threshold and confidence threshold, precision measures the proportion of correct detections among all detections, while recall measures the proportion of ground-truth objects that are successfully detected:
Precision = TP TP + FP , Recall = TP TP + FN
where TP, FP, and FN denote true positives, false positives, and false negatives, respectively.
Average Precision (AP). AP summarizes the precision–recall curve by computing the area under the curve. We employ the 101-point interpolation method:
AP = 1 101 r { 0 , 0.01 , . . . , 1 } p interp ( r ) , where p interp ( r ) = max r ˜ r p ( r ˜ )
where p ( r ) is the precision at recall level r, and p interp ( r ) is the interpolated precision.
Mean Average Precision (mAP). mAP averages AP across all object classes. For single-class UAV detection, mAP equals AP. We report two standard metrics:
  • mAP@0.5: AP computed at IoU threshold τ = 0.5 , which is lenient and focuses on detection capability.
  • mAP@0.5:0.95: AP averaged across IoU thresholds from 0.5 to 0.95 with step 0.05, which requires precise localization:
    mAP@0.5:0.95 = 1 10 τ { 0.5 , 0.55 , . . . , 0.95 } AP τ
Inference Speed (FPS). Frames per second measures real-time detection capability, computed as the inverse of average inference time per image. All FPS measurements are conducted on a single NVIDIA RTX 3090 GPU with batch size 1 and input resolution 640 × 640 .
Model Complexity. We report model parameters in millions (M) and floating-point operations (FLOPs) in giga-operations (G) to assess computational requirements for deployment.

4.4. Comparison with State-of-the-Art Methods

We conduct comprehensive comparisons with state-of-the-art object detection methods spanning both single-modality approaches (RGB-only and infrared-only) and multimodal fusion methods. For single-modality baselines, we include representative detectors from the YOLO series (YOLOv5 through YOLOv10) and the detection transformer family (RT-DETR). For multimodal methods, we compare against established fusion approaches including DenseFuse, U2Fusion, TarDAL, M3FD, CFT, and TransFuse. All baseline methods are retrained on the Anti-UAV300 dataset using their official implementations with optimized hyperparameters to ensure fair comparison. Table 2 presents the quantitative comparison results.
As demonstrated in Table 2, DCAM-DETR achieves superior performance across all evaluation metrics. Compared to the best single-modality method (RT-DETR on RGB), our approach improves mAP@0.5 by 9.0 percentage points (85.7% → 94.7%) and mAP@0.5:0.95 by 16.2 percentage points (62.1% → 78.3%), demonstrating the substantial advantage of multimodal fusion for UAV detection. Among multimodal methods, DCAM-DETR outperforms the previous best method (TransFuse) by 2.6 percentage points in mAP@0.5 and 7.1 percentage points in mAP@0.5:0.95, while achieving significantly higher inference speed (42 FPS vs. 25 FPS) and fewer parameters (47.6 M vs. 63.4 M).
The performance improvements are particularly notable for the more stringent mAP@0.5:0.95 metric, which requires precise localization across multiple IoU thresholds. This indicates that DCAM-DETR not only detects UAVs more reliably but also localizes them more accurately, which is critical for downstream applications such as tracking and interception.
To provide a comprehensive multi-dimensional comparison, Figure 7 presents a radar chart visualizing the performance of different methods across six key dimensions: mAP@0.5, mAP@0.5:0.95, inference speed, parameter efficiency, small target detection accuracy, and nighttime scene performance. DCAM-DETR demonstrates superior performance across all dimensions, particularly excelling in the challenging small target and nighttime scenarios where multimodal fusion provides the greatest advantage.

4.5. Qualitative Results

Beyond quantitative metrics, we provide qualitative analysis to demonstrate DCAM-DETR’s detection capability under diverse real-world conditions. Figure 8 presents representative detection results on challenging scenarios from the Anti-UAV300 dataset, encompassing low-light and nighttime conditions where RGB imagery provides limited discriminative information, complex urban backgrounds with cluttered visual patterns, small target scales where UAVs occupy minimal pixel areas, and varying weather conditions including haze and overcast skies.
The qualitative results demonstrate that DCAM-DETR consistently produces accurate detections across diverse challenging conditions. In nighttime scenarios (rows 1–2), where RGB imagery provides limited discriminative information, the model effectively leverages thermal infrared features to maintain detection accuracy. In daytime scenarios with complex urban backgrounds (rows 3–4), the model successfully distinguishes small UAV targets from cluttered backgrounds by exploiting complementary RGB texture information. The consistent detection performance across varying target scales validates the effectiveness of our multi-scale feature fusion and DADM-enhanced detection head.

4.6. Performance Under Different Environmental Conditions

To further analyze the robustness of DCAM-DETR across varying environmental conditions, we evaluate detection performance on subsets of the Anti-UAV300 test set categorized by illumination and weather conditions. Figure 9 presents the comparative analysis across five environmental scenarios: daytime clear, daytime cloudy, dusk, nighttime clear, and nighttime foggy conditions, along with the overall performance.
The results reveal that DCAM-DETR maintains consistently high performance across all environmental conditions, with mAP@0.5 ranging from 91.2% (nighttime foggy) to 96.2% (daytime clear). Notably, the performance gap between DCAM-DETR and RGB-only methods widens substantially in challenging scenarios: the improvement over RT-DETR (RGB) increases from 5.0 percentage points in daytime clear conditions to 32.8 percentage points in nighttime foggy conditions. This demonstrates that multimodal fusion provides the greatest advantage precisely when single-modality approaches struggle most, validating the practical value of our approach for real-world anti-UAV deployment.

4.7. Ablation Studies

To systematically validate the contribution of each proposed component and design choice, we conduct comprehensive ablation studies examining the MobileMamba backbone, cross-dimensional attention modules (CDA and CPA), Adaptive Feature Fusion Module (AFFM), and Dual-Attention Decoupling Module (DADM). All ablation experiments are performed on the Anti-UAV300 dataset with consistent training configurations, including identical hyperparameters, data augmentation strategies, and random seeds to ensure reproducible and fair comparisons.

4.7.1. Component-Wise Ablation

To quantify the individual contribution of each proposed component, we conduct progressive ablation experiments starting from a baseline configuration and incrementally adding each module. The baseline model employs the standard RT-DETR architecture with simple channel-wise concatenation for multimodal fusion. We then sequentially integrate the MobileMamba backbone, CDA and CPA attention modules, AFFM, and DADM to observe the performance gains from each component. Table 3 presents the progressive ablation results on the Anti-UAV300 dataset.
The baseline model concatenates RGB and infrared features along the channel dimension and processes them through the standard RT-DETR architecture, achieving 87.2% mAP@0.5. Replacing the CNN backbone with MobileMamba yields a substantial improvement of 3.1 percentage points in mAP@0.5 and 5.2 percentage points in mAP@0.5:0.95, validating the effectiveness of state space models for global context modeling in multimodal detection. The linear complexity of Mamba enables efficient processing while capturing long-range dependencies crucial for small target detection.
Adding CDA and CPA modules further improves performance by 2.2 percentage points in mAP@0.5 and 4.4 percentage points in mAP@0.5:0.95, demonstrating the importance of explicit cross-modal attention mechanisms for capturing intermodal correlations. The AFFM contributes an additional 1.3 percentage points and 3.3 percentage points improvement, validating the benefit of adaptive fusion that dynamically weights modality contributions based on scene characteristics. Finally, incorporating DADM in the detection head yields the full model performance of 94.7% mAP@0.5 and 78.3% mAP@0.5:0.95, with the hierarchical dilated convolutions and attention decomposition enhancing small target discrimination.

4.7.2. Modality Ablation

To investigate the contribution of each modality and justify the necessity of multimodal fusion over single-modality approaches with increased capacity, we conduct modality ablation experiments. Table 4 compares: (1) single-modality models using only RGB or IR input, (2) single-modality models with doubled backbone capacity to match the parameter count of our multimodal model, and (3) our full multimodal DCAM-DETR.
The results demonstrate that simply increasing single-modality model capacity cannot compensate for the lack of complementary information. RGB-only with doubled capacity (87.3% mAP@0.5) still falls 7.4 percentage points behind our multimodal approach, while using nearly twice the parameters (85.2 M vs. 47.6 M). This validates that RGB and IR modalities provide fundamentally complementary information—RGB captures texture and color details effective in well-lit conditions, while IR captures thermal signatures robust to illumination variations. The performance gap is particularly pronounced in challenging scenarios: in nighttime foggy conditions, RGB-only (2× capacity) achieves only 62.1% mAP@0.5 compared to DCAM-DETR’s 91.2%, demonstrating that multimodal fusion is essential rather than optional for robust anti-UAV detection.

4.7.3. Backbone Architecture Comparison

To validate the effectiveness of our MobileMamba backbone for multimodal feature extraction, we compare it against representative CNN-based backbones (ResNet-50, ResNet-101), transformer-based backbones (Swin-T, Swin-S), and the vanilla vision Mamba architecture (VMamba-T). All backbone variants are integrated into our detection framework with identical fusion modules and detection heads to ensure fair comparison. Table 5 presents the quantitative results on the Anti-UAV300 dataset.
MobileMamba outperforms both CNN-based (ResNet) and transformer-based (Swin) backbones while maintaining competitive inference speed. Compared to VMamba-T, our MobileMamba achieves 2.4 percentage points higher mAP@0.5 and 4.5 percentage points higher mAP@0.5:0.95, demonstrating the effectiveness of our architectural modifications for multimodal fusion.

4.7.4. Fusion Strategy Comparison

To validate the effectiveness of our proposed multimodal fusion approach, we conduct comprehensive comparisons with representative fusion strategies spanning early, late, and intermediate fusion paradigms. Early fusion methods directly concatenate RGB and infrared features at the input or shallow feature levels, while late fusion combines predictions from modality-specific detectors. Intermediate fusion approaches, including attention-based methods, integrate features at multiple abstraction levels. Table 6 presents the quantitative comparison results on the Anti-UAV300 dataset.
Our proposed fusion strategy combining CDA, CPA, and AFFM significantly outperforms all alternative approaches across both evaluation metrics. Notably, CDA + CPA outperforms vanilla cross-attention by 1.7 percentage points in mAP@0.5, demonstrating the benefit of explicit spatial and channel correlation modeling. AFFM provides an additional 4.4 percentage points improvement over simple weighted averaging, validating the importance of adaptive scene-dependent fusion.

4.7.5. DADM Component Analysis

To isolate DADM’s contribution from the dilated convolution alone, we conduct additional ablation experiments in Table 7.
The results show that dilated convolutions alone contribute 0.7 percentage points improvement, while the complete DADM with dual-attention provides an additional 1.5 percentage points, demonstrating that the attention mechanism is essential for small target discrimination beyond multi-scale context aggregation.

4.7.6. Performance by Target Size and Scene Condition

To validate our claim that DCAM-DETR excels at small target detection, we report performance breakdown by UAV size and scene condition in Table 8.
The results clearly demonstrate that DCAM-DETR provides the most significant improvements for small targets (+18.2%) and challenging nighttime/foggy conditions (+32.8%), validating the effectiveness of multimodal fusion for these challenging scenarios.

4.7.7. Per-Class AP Analysis

To demonstrate robustness across UAV types, we report per-class AP in Table 9.
DCAM-DETR demonstrates consistent performance across all UAV types, with slightly lower performance on Micro UAVs due to their extremely small size, which aligns with our analysis of small target detection challenges.

4.8. Attention Visualization

To provide interpretable insights into how DCAM-DETR leverages multimodal information for detection, we visualize the attention maps learned by our cross-dimensional attention modules. Understanding the attention patterns helps validate that the model correctly identifies discriminative regions and exploits complementary information from both modalities. Figure 10 presents the attention heatmaps overlaid on input images from both RGB and infrared modalities, revealing the spatial regions the model focuses on during the detection process.
The visualization demonstrates that our CDA and CPA modules successfully learn to focus on UAV regions while suppressing background clutter. Notably, the attention patterns exhibit complementary characteristics between RGB and infrared modalities: RGB attention emphasizes texture and shape features visible under adequate illumination, while infrared attention focuses on thermal signatures that remain robust across varying lighting conditions. This complementary behavior validates our design of separate cross-dimensional attention mechanisms for capturing modality-specific characteristics while enabling effective information fusion.

4.9. Cross-Dataset Evaluation

A critical consideration for practical deployment is whether the learned multimodal fusion strategy generalizes beyond the training domain. To rigorously evaluate this generalization capability, we test our model trained exclusively on Anti-UAV300 on two additional multimodal datasets (FLIR-ADAS and KAIST) without any fine-tuning or domain adaptation. This zero-shot transfer setting provides a stringent test of whether our architectural innovations capture general principles for RGB–infrared fusion rather than dataset-specific patterns. Table 10 presents the cross-dataset evaluation results.
DCAM-DETR achieves the best performance on both cross-domain datasets, demonstrating strong generalization capability. The consistent improvements over baseline methods (3.1 percentage points average improvement over TransFuse) validate that our multimodal fusion strategy and architectural innovations capture general principles for effective RGB–infrared fusion rather than overfitting to the Anti-UAV300 dataset characteristics.

4.10. Computational Efficiency Analysis

For practical deployment in real-time anti-UAV systems, computational efficiency is as critical as detection accuracy. We conduct comprehensive efficiency analysis comparing DCAM-DETR with representative single-modality and multimodal detection methods, measuring inference speed (FPS), model parameters, floating-point operations (FLOPs), and GPU memory consumption. All measurements are performed on a single NVIDIA RTX 3090 GPU with batch size 1 and input resolution 640 × 640 . Table 11 presents the detailed computational efficiency comparison.
DCAM-DETR achieves the best accuracy-efficiency trade-off among all compared methods. Despite processing two input modalities, our model maintains competitive FLOPs (142.3 G) and memory consumption (5.8 GB) compared to single-modality methods, while achieving substantially higher accuracy. The linear complexity of the MobileMamba backbone enables efficient processing of high-resolution multimodal inputs without the quadratic scaling of transformer-based approaches.

4.11. Edge Device Deployment

For practical anti-UAV deployment, we evaluate DCAM-DETR on edge devices. Table 12 presents results on NVIDIA Jetson AGX Orin with different optimization strategies.
Our INT8 quantized model maintains 93.2% mAP@0.5 while achieving 31 FPS on edge devices, demonstrating practical deployment feasibility for real-time anti-UAV applications.

4.12. Failure Case Analysis

To provide comprehensive evaluation, we analyze failure cases and identify scenarios where DCAM-DETR struggles. Table 13 summarizes failure categories and their impact on performance.
The analysis reveals that failures primarily occur in extreme conditions where both modalities provide limited discriminative information. AFFM failure analysis shows incorrect weighting primarily occurs when both modalities are similarly degraded (e.g., foggy night scenes). CDA failures typically result from severe viewpoint differences causing spatial misalignment, while CPA failures occur when channel semantics are ambiguous across modalities.

5. Discussion

The experimental results reveal several important insights regarding multimodal fusion for anti-UAV detection. The substantial performance improvement over single-modality methods (9.0 percentage points in mAP@0.5) underscores the fundamental importance of leveraging complementary RGB and infrared information. The attention visualization in Figure 10 confirms that DCAM-DETR learns modality-specific attention patterns: RGB attention emphasizes texture and shape features, while infrared attention focuses on thermal signatures robust to illumination variations. The strong cross-dataset generalization across FLIR-ADAS and KAIST datasets suggests that our architectural innovations capture general principles for RGB–infrared fusion rather than dataset-specific patterns.
The ablation studies provide quantitative evidence for each component’s contribution. The MobileMamba backbone’s 3.1 percentage point improvement validates the importance of linear-complexity global context modeling for small target detection. The 2.2 percentage point gain from CDA and CPA demonstrates that explicit cross-modal attention across both spatial and channel dimensions significantly outperforms simple fusion strategies. AFFM’s adaptive gating mechanism and DADM’s multi-scale contextual aggregation further enhance performance by addressing modality reliability variations and small target discrimination challenges, respectively.

Limitations and Societal Impact

Despite these strengths, several limitations warrant discussion:
Technical Limitations: (1) DCAM-DETR requires spatially aligned RGB–infrared image pairs, and misalignment beyond 5 pixels causes performance degradation (mAP drops by 3.2% at 10-pixel misalignment). (2) The sequential nature of state space models limits parallelization compared to fully convolutional architectures. (3) Performance degrades significantly for extremely small targets (less than 10 pixels) where both modalities provide limited discriminative information, achieving only 58.4% mAP@0.5. (4) Computational requirements may limit deployment on highly resource-constrained devices without quantization. (5) Training requires paired RGB-IR data with annotations, which is expensive to collect compared to single-modality datasets.
Societal Impact: Our anti-UAV detection system has both positive and negative societal implications. Positive aspects include enhanced security for critical infrastructure, airports, prisons, and public events, as well as improved safety in emergency response scenarios where unauthorized UAVs pose risks. Negative aspects include potential misuse for surveillance purposes, privacy concerns when deployed in public spaces, and dual-use technology considerations where the same detection capabilities could be used to evade legitimate UAV operations. We advocate for responsible deployment with appropriate regulatory oversight, transparency about system capabilities and limitations, and adherence to ethical guidelines for surveillance technologies.
Future work could explore explicit alignment modules to handle misaligned inputs, video-based temporal extensions for improved tracking, lightweight variants optimized for edge deployment, and self-supervised learning approaches to reduce annotation requirements.

6. Conclusions

In this paper, we proposed DCAM-DETR, a novel multimodal detection framework that synergistically integrates Mamba-based state space models with the RT-DETR architecture for robust real-time anti-UAV detection. Our approach introduces four principal innovations: a MobileMamba backbone for efficient global context modeling with linear computational complexity, Cross-Dimensional Attention (CDA) and Cross-Path Attention (CPA) modules for fine-grained multimodal feature alignment, an Adaptive Feature Fusion Module (AFFM) for scene-dependent fusion, and a Dual-Attention Decoupling Module (DADM) for enhanced small target detection.
Comprehensive experiments on Anti-UAV300, FLIR-ADAS, and KAIST datasets demonstrate that DCAM-DETR achieves state-of-the-art performance with 94.7% mAP@0.5 on Anti-UAV300, outperforming existing methods by substantial margins while maintaining real-time inference speed of 42 FPS. The strong cross-dataset generalization validates the effectiveness of our multimodal fusion strategy across diverse application domains. Ablation studies confirm the contribution of each proposed component, with the MobileMamba backbone, cross-dimensional attention modules, AFFM, and DADM collectively contributing to the superior performance.
The success of DCAM-DETR demonstrates the potential of state space models for multimodal visual perception tasks, offering an efficient alternative to transformer-based approaches while maintaining strong representational capability. Future work will explore temporal extensions for video-based detection, lightweight variants for edge deployment, and self-supervised learning approaches to reduce annotation requirements.

Author Contributions

Conceptualization, Y.L.; methodology, Y.L.; software, Y.L.; validation, Y.L.; formal analysis, Y.L.; investigation, Y.L.; resources, Z.Q.; data curation, Y.L.; writing—original draft preparation, Y.L.; writing—review and editing, Z.Q.; visualization, Y.L.; supervision, Z.Q.; project administration, Z.Q.; funding acquisition, Z.Q. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The Anti-UAV300 dataset is publicly available at https://github.com/ZhaoJ9014/Anti-UAV (accessed on 15 October 2025). The FLIR-ADAS and KAIST datasets are available from their respective sources. The source code and pre-trained models are available upon reasonable request.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Shakhatreh, H.; Sawalmeh, A.H.; Al-Fuqaha, A.; Dou, Z.; Almaita, E.; Khalil, I.; Othman, N.S.; Khreishah, A.; Guizani, M. Unmanned aerial vehicles (UAVs): A survey on civil applications and key research challenges. IEEE Access 2019, 7, 48572–48634. [Google Scholar] [CrossRef]
  2. Mohsan, S.A.H.; Khan, M.A.; Noor, F.; Ullah, I.; Alsharif, M.H. Towards the unmanned aerial vehicles (UAVs): A comprehensive review. Drones 2022, 6, 147. [Google Scholar] [CrossRef]
  3. Güvenç, İ.; Koohifar, F.; Singh, S.; Sichitiu, M.L.; Matolak, D. Detection, tracking, and interdiction for amateur drones. IEEE Commun. Mag. 2018, 56, 75–81. [Google Scholar] [CrossRef]
  4. Shi, X.; Yang, C.; Xie, W.; Liang, C.; Shi, Z.; Chen, J. Anti-drone system with multiple surveillance technologies: Architecture, implementation, and challenges. IEEE Commun. Mag. 2018, 56, 68–74. [Google Scholar] [CrossRef]
  5. Federal Aviation Administration. FAA Unmanned Aircraft Systems (UAS) Traffic Management. 2024. Available online: https://www.faa.gov/uas (accessed on 15 October 2025).
  6. Wu, X.; Dong, J.; Bao, W.; Zou, B.; Wang, L.; Wang, H. Augmented intelligence of things for emergency vehicle secure trajectory prediction and task offloading. IEEE Internet Things J. 2024, 11, 36030–36043. [Google Scholar] [CrossRef]
  7. Ezuma, M.; Erden, F.; Anjinappa, C.K.; Ozdemir, O.; Güvenç, İ. Radar cross section based statistical recognition of UAVs at microwave frequencies. IEEE Trans. Aerosp. Electron. Syst. 2020, 58, 27–46. [Google Scholar] [CrossRef]
  8. Coluccia, A.; Fascista, A.; Schumann, A.; Sommer, L.; Ghenescu, M.; Piatrik, T.; De Cubber, G.; Nalamati, M.; Kapoor, A.; Saqib, M.; et al. Drone-vs-bird detection challenge at IEEE AVSS2019. arXiv 2019, arXiv:1910.07360. [Google Scholar]
  9. Jiang, N.; Wang, K.; Peng, X.; Yu, X.; Wang, Q.; Xing, J.; Li, G.; Zhao, J.; Guo, Z.; Han, Z.; et al. Anti-UAV: A large-scale benchmark for vision-based UAV tracking. IEEE Trans. Multimed. 2022, 25, 486–500. [Google Scholar] [CrossRef]
  10. Huang, B.; Chen, J.; Xu, T.; Wang, Y.; Jiang, S.; Wang, Y.; Wang, L.; Li, J. Anti-UAV410: A thermal infrared benchmark and customized scheme for tracking drones in the wild. IEEE Trans. Pattern Anal. Mach. Intell. 2024, 46, 2852–2865. [Google Scholar] [CrossRef]
  11. Du, D.; Qi, Y.; Yu, H.; Yang, Y.; Duan, K.; Li, G.; Zhang, W.; Huang, Q.; Tian, Q. The unmanned aerial vehicle benchmark: Object detection and tracking. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 370–386. [Google Scholar] [CrossRef]
  12. Rozantsev, A.; Lepetit, V.; Fua, P. Detecting flying objects using a single moving camera. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 879–892. [Google Scholar] [CrossRef]
  13. Li, J.; Fan, C.; Ou, C.; Zhang, H. Infrared and Visible Image Fusion Techniques for UAVs: A Comprehensive Review. Drones 2025, 9, 811. [Google Scholar] [CrossRef]
  14. Liu, J.; Fan, X.; Huang, Z.; Wu, G.; Liu, R.; Zhong, W.; Luo, Z. Target-aware dual adversarial learning and a multi-scenario multi-modality benchmark to fuse infrared and visible for object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 5802–5811. [Google Scholar] [CrossRef]
  15. Chen, Y.T.; Shi, J.; Ye, Z.; Mertz, C.; Ramanan, D.; Kong, S. Multimodal object detection via probabilistic ensembling. In Proceedings of the European Conference on Computer Vision (ECCV); Springer: Cham, Switzerland, 2022. [Google Scholar] [CrossRef]
  16. Wang, S.; Wang, C.; Shi, C.; Liu, Y.; Lu, M. Mask-guided mamba fusion for drone-based visible-infrared vehicle detection. IEEE Trans. Geosci. Remote Sens. 2024, 62, 5005712. [Google Scholar] [CrossRef]
  17. Li, H.; Wu, X.J. DenseFuse: A fusion approach to infrared and visible images. IEEE Trans. Image Process. 2019, 28, 2614–2623. [Google Scholar] [CrossRef] [PubMed]
  18. Zhang, J.; Liu, H.; Yang, K.; Hu, X.; Liu, R.; Stiefelhagen, R. CMX: Cross-modal fusion for RGB-X semantic segmentation with transformers. IEEE Trans. Intell. Transp. Syst. 2023, 24, 14679–14694. [Google Scholar] [CrossRef]
  19. Wang, Q.; Wu, B.; Zhu, P.; Li, P.; Zuo, W.; Hu, Q. ECA-Net: Efficient channel attention for deep convolutional neural networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 11534–11542. [Google Scholar] [CrossRef]
  20. Ma, J.; Yu, W.; Liang, P.; Li, C.; Jiang, J. FusionGAN: A generative adversarial network for infrared and visible image fusion. Inf. Fusion 2019, 48, 11–26. [Google Scholar] [CrossRef]
  21. El Ahmar, W.; Massoud, Y.; Kolhatkar, D.; AlGhamdi, H.; Alja’Afreh, M.; Hammoud, R.; Laganiere, R. Enhanced thermal-RGB fusion for robust object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023; pp. 365–374. [Google Scholar]
  22. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is all you need. Adv. Neural Inf. Process. Syst. 2017, 30, 5998–6008. [Google Scholar]
  23. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An image is worth 16 × 16 words: Transformers for image recognition at scale. In Proceedings of the International Conference on Learning Representations, Vienna, Austria, 3–7 May 2021. [Google Scholar]
  24. Liu, Z.; Lin, Y.; Cao, Y.; Hu, H.; Wei, Y.; Zhang, Z.; Lin, S.; Guo, B. Swin transformer: Hierarchical vision transformer using shifted windows. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 10–17 October 2021; pp. 10012–10022. [Google Scholar] [CrossRef]
  25. Wang, W.; Xie, E.; Li, X.; Fan, D.P.; Song, K.; Liang, D.; Lu, T.; Luo, P.; Shao, L. Pyramid vision transformer: A versatile backbone for dense prediction without convolutions. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 10–17 October 2021; pp. 568–578. [Google Scholar] [CrossRef]
  26. Zhao, Y.; Lv, W.; Xu, S.; Wei, J.; Wang, G.; Dang, Q.; Liu, Y.; Chen, J. DETRs beat YOLOs on real-time object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 16–22 June 2024; pp. 16965–16974. [Google Scholar] [CrossRef]
  27. Carion, N.; Massa, F.; Synnaeve, G.; Usunier, N.; Kirillov, A.; Zagoruyko, S. End-to-end object detection with transformers. In Proceedings of the European Conference on Computer Vision, Glasgow, UK, 23–28 August 2020; pp. 213–229. [Google Scholar] [CrossRef]
  28. Gu, A.; Dao, T. Mamba: Linear-time sequence modeling with selective state spaces. arXiv 2024, arXiv:2312.00752. [Google Scholar]
  29. Zhu, L.; Liao, B.; Zhang, Q.; Wang, X.; Liu, W.; Wang, X. Vision mamba: Efficient visual representation learning with bidirectional state space model. arXiv 2024, arXiv:2401.09417. [Google Scholar] [CrossRef]
  30. Liu, Y.; Tian, Y.; Zhao, Y.; Yu, H.; Xie, L.; Wang, Y.; Ye, Q.; Jiao, J.; Liu, Y. Vmamba: Visual state space model. Adv. Neural Inf. Process. Syst. 2024, 37, 103031–103063. [Google Scholar]
  31. Wang, G.; Song, M.; Hwang, J.N. Recent Advances in Embedding Methods for Multi-Object Tracking: A Survey. arXiv 2022, arXiv:2205.10766. [Google Scholar] [CrossRef]
  32. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards real-time object detection with region proposal networks. Adv. Neural Inf. Process. Syst. 2015, 28. [Google Scholar] [CrossRef]
  33. Redmon, J.; Farhadi, A. YOLOv3: An incremental improvement. arXiv 2018, arXiv:1804.02767. [Google Scholar] [CrossRef]
  34. Bochkovskiy, A.; Wang, C.Y.; Liao, H.Y.M. YOLOv4: Optimal speed and accuracy of object detection. arXiv 2020, arXiv:2004.10934. [Google Scholar] [CrossRef]
  35. Jocher, G.; Chaurasia, A.; Stoken, A.; Borber, J.; NanoCode012; Kwon, Y.; Michael, K.; TaoXie; Fang, J.; imyhxy; et al. YOLOv5 by Ultralytics. GitHub Repos. 2022. Available online: https://github.com/ultralytics/yolov5 (accessed on 15 October 2025).
  36. Ge, Z.; Liu, S.; Wang, F.; Li, Z.; Sun, J. YOLOX: Exceeding YOLO series in 2021. arXiv 2021, arXiv:2107.08430. [Google Scholar] [CrossRef]
  37. Wang, C.Y.; Bochkovskiy, A.; Liao, H.Y.M. YOLOv7: Trainable bag-of-freebies sets new state-of-the-art for real-time object detectors. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023; pp. 7464–7475. [Google Scholar] [CrossRef]
  38. Wang, C.Y.; Yeh, I.H.; Liao, H.Y.M. YOLOv9: Learning what you want to learn using programmable gradient information. arXiv 2024, arXiv:2402.13616. [Google Scholar] [CrossRef]
  39. Wang, A.; Chen, H.; Liu, L.; Chen, K.; Lin, Z.; Han, J.; Ding, G. YOLOv10: Real-time end-to-end object detection. arXiv 2024, arXiv:2405.14458. [Google Scholar]
  40. Lin, T.Y.; Goyal, P.; Girshick, R.; He, K.; Dollár, P. Focal loss for dense object detection. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2980–2988. [Google Scholar] [CrossRef]
  41. Zhou, L.; Liu, Z.; Zhao, H.; Hou, Y.E.; Liu, Y.; Zuo, X.; Dang, L. A multi-scale object detector based on coordinate and global information aggregation for UAV aerial images. Remote Sens. 2023, 15, 3468. [Google Scholar] [CrossRef]
  42. Zhang, X.; Ye, P.; Xiao, G. Image fusion meets deep learning: A survey and perspective. Inf. Fusion 2021, 76, 323–336. [Google Scholar] [CrossRef]
  43. Feng, D.; Haase-Schütz, C.; Rosenbaum, L.; Hertlein, H.; Glaeser, C.; Timm, F.; Wiesbeck, W.; Dietmayer, K. Deep multi-modal object detection and semantic segmentation for autonomous driving: Datasets, methods, and challenges. IEEE Trans. Intell. Transp. Syst. 2020, 22, 1341–1360. [Google Scholar] [CrossRef]
  44. Xu, H.; Ma, J.; Jiang, J.; Guo, X.; Ling, H. U2Fusion: A unified unsupervised image fusion network. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 44, 502–518. [Google Scholar] [CrossRef] [PubMed]
  45. Woo, S.; Park, J.; Lee, J.Y.; Kweon, I.S. CBAM: Convolutional block attention module. In Proceedings of the European Conference on Computer Vision, Munich, Germany, 8–14 September 2018; pp. 3–19. [Google Scholar] [CrossRef]
  46. Hu, J.; Shen, L.; Sun, G. Squeeze-and-excitation networks. In Proceedings of theIEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 7132–7141. [Google Scholar] [CrossRef]
  47. Chen, J.; Lu, Y.; Yu, Q.; Luo, X.; Adeli, E.; Wang, Y.; Lu, L.; Yuille, A.L.; Zhou, Y. TransFuse: Fusing transformers and CNNs for medical image segmentation. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Singapore, 18–22 September 2022; pp. 14–24. [Google Scholar] [CrossRef]
  48. Yang, P.; Gao, J.; Chen, W. Multimodal learning with transformers: A survey. IEEE Trans. Pattern Anal. Mach. Intell. 2023, 45, 15612–15631. [Google Scholar] [CrossRef]
  49. Dong, A.; Wang, L.; Liu, J.; Xu, J.; Zhao, G.; Zhai, Y.; Lv, G.; Cheng, J. Co-enhancement of multi-modality image fusion and object detection via feature adaptation. IEEE Trans. Circuits Syst. Video Technol. 2024, 34, 12624–12637. [Google Scholar] [CrossRef]
  50. Peng, S.; Zhu, X.; Cao, X.; Deng, C. FusionMamba: Efficient Remote Sensing Image Fusion with State Space Model. arXiv 2024, arXiv:2404.07932. [Google Scholar] [CrossRef]
  51. Cai, Q.; Pan, Y.; Yao, T.; Ngo, C.W.; Mei, T. Objectfusion: Multi-modal 3d object detection with object-centric fusion. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Paris, France, 1–6 October 2023; pp. 18067–18076. [Google Scholar]
  52. Wang, X.; Wang, S.; Ding, Y.; Li, Y.; Wu, W.; Rong, Y.; Kong, W.; Huang, J.; Li, S.; Yang, H.; et al. State space model for new-generation network alternative to transformers: A survey. arXiv 2024, arXiv:2404.09516. [Google Scholar] [CrossRef]
  53. Gupta, A.; Gu, A.; Berant, J. Diagonal state spaces are as effective as structured state spaces. Adv. Neural Inf. Process. Syst. 2022, 35, 22982–22994. [Google Scholar]
  54. Zhao, H.; Yan, L.; Hou, Z.; Lin, J.; Zhao, Y.; Ji, Z.; Wang, Y. Error Analysis Strategy for Long-term Correlated Network Systems: Generalized Nonlinear Stochastic Processes and Dual-Layer Filtering Architecture. IEEE Internet Things J. 2025, 12, 33731–33745. [Google Scholar] [CrossRef]
  55. Huang, T.; Pei, X.; You, S.; Wang, F.; Qian, C.; Xu, C. LocalMamba: Visual state space model with windowed selective scan. arXiv 2024, arXiv:2403.09338. [Google Scholar] [CrossRef]
  56. Hatamizadeh, A.; Kautz, J. Mambavision: A hybrid mamba-transformer vision backbone. In Proceedings of the Computer Vision and Pattern Recognition Conference, Vancouver, BC, Canada, 17–24 June 2025; pp. 25261–25270. [Google Scholar]
  57. Zhu, X.; Su, W.; Lu, L.; Li, B.; Wang, X.; Dai, J. Deformable detr: Deformable transformers for end-to-end object detection. arXiv 2020, arXiv:2010.04159. [Google Scholar] [CrossRef]
  58. Meng, D.; Chen, X.; Fan, Z.; Zeng, G.; Li, H.; Yuan, Y.; Sun, L.; Wang, J. Conditional DETR for fast training convergence. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 10–17 October 2021; pp. 3651–3660. [Google Scholar] [CrossRef]
  59. Liu, S.; Li, F.; Zhang, H.; Yang, X.; Qi, X.; Su, H.; Zhu, J.; Zhang, L. Dab-detr: Dynamic anchor boxes are better queries for detr. arXiv 2022, arXiv:2201.12329. [Google Scholar] [CrossRef]
  60. Rezatofighi, H.; Tsoi, N.; Gwak, J.; Sadeghian, A.; Reid, I.; Savarese, S. Generalized intersection over union: A metric and a loss for bounding box regression. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 658–666. [Google Scholar] [CrossRef]
  61. FLIR Systems. FLIR Thermal Dataset for Algorithm Training. 2019. Available online: https://www.flir.com/oem/adas/adas-dataset-form/ (accessed on 15 October 2025).
  62. Hwang, S.; Park, J.; Kim, N.; Choi, Y.; Kweon, I.S. Multispectral pedestrian detection: Benchmark dataset and baseline. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 1037–1045. [Google Scholar] [CrossRef]
  63. Loshchilov, I.; Hutter, F. Decoupled weight decay regularization. arXiv 2017, arXiv:1711.05101. [Google Scholar] [CrossRef]
  64. Loshchilov, I.; Hutter, F. Sgdr: Stochastic gradient descent with warm restarts. arXiv 2016, arXiv:1608.03983. [Google Scholar] [CrossRef]
  65. Zhang, H.; Cisse, M.; Dauphin, Y.N.; Lopez-Paz, D. mixup: Beyond empirical risk minimization. arXiv 2017, arXiv:1710.09412. [Google Scholar] [CrossRef]
  66. Jocher, G.; Chaurasia, A.; Qiu, J. Ultralytics YOLOv8. GitHub Repos. 2023. Available online: https://github.com/ultralytics/ultralytics (accessed on 15 October 2025).
  67. Feng, Y.; Luo, E.; Lu, H.; Zhai, S. Cross-modality feature fusion for night pedestrian detection. Front. Phys. 2024, 12, 1356248. [Google Scholar] [CrossRef]
  68. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar] [CrossRef]
Figure 1. Overall architecture of DCAM-DETR. The framework processes aligned RGB and infrared image pairs through the MobileMamba backbone, which extracts hierarchical features at multiple scales (P3, P4, P5). These features are subsequently processed through the Efficient Transformer Encoder with CDA and CPA modules for cross-modal interaction. The AFFM adaptively fuses multimodal features, which are then refined through DADM-enhanced CCFF structures. Finally, the Uncertainty-minimal Query Selection mechanism and Decoder with detection Head generate the final predictions. Note: The Chinese characters visible in some images are metadata overlays from the camera equipment (date, time, and camera settings), which are inherent to the original captured images and represent authentic real-world data conditions.
Figure 1. Overall architecture of DCAM-DETR. The framework processes aligned RGB and infrared image pairs through the MobileMamba backbone, which extracts hierarchical features at multiple scales (P3, P4, P5). These features are subsequently processed through the Efficient Transformer Encoder with CDA and CPA modules for cross-modal interaction. The AFFM adaptively fuses multimodal features, which are then refined through DADM-enhanced CCFF structures. Finally, the Uncertainty-minimal Query Selection mechanism and Decoder with detection Head generate the final predictions. Note: The Chinese characters visible in some images are metadata overlays from the camera equipment (date, time, and camera settings), which are inherent to the original captured images and represent authentic real-world data conditions.
Information 17 00103 g001
Figure 2. Architecture of the SS2D Block. The input feature map undergoes scan expanding to generate four directional sequences, which are processed through the S6 Block containing the selective state space model. The outputs are then merged through scan merging to reconstruct the 2D spatial structure while capturing global dependencies.
Figure 2. Architecture of the SS2D Block. The input feature map undergoes scan expanding to generate four directional sequences, which are processed through the S6 Block containing the selective state space model. The outputs are then merged through scan merging to reconstruct the 2D spatial structure while capturing global dependencies.
Information 17 00103 g002
Figure 3. Architecture of the S6 Block. The block consists of layer normalization (LN), SS2D Block for selective state space processing, followed by another LN and feed-forward network (FFN). Residual connections ensure stable gradient propagation. The SS2D Block internally comprises linear projection, depthwise convolution (DWConv), SiLU activation, SS2D selective scan, LN, and linear output projection. Note: The Chinese characters visible in some images are metadata overlays from the camera equipment (date, time, and camera settings), which are inherent to the original captured images and represent authentic real-world data conditions.
Figure 3. Architecture of the S6 Block. The block consists of layer normalization (LN), SS2D Block for selective state space processing, followed by another LN and feed-forward network (FFN). Residual connections ensure stable gradient propagation. The SS2D Block internally comprises linear projection, depthwise convolution (DWConv), SiLU activation, SS2D selective scan, LN, and linear output projection. Note: The Chinese characters visible in some images are metadata overlays from the camera equipment (date, time, and camera settings), which are inherent to the original captured images and represent authentic real-world data conditions.
Information 17 00103 g003
Figure 4. Architecture of Cross-Dimensional Attention (CDA, left panel) and Cross-Path Attention (CPA, right panel) modules. CDA captures spatial correlations through spatial self-attention, frequency projection, and channel projection with sigmoid gating. CPA models channel-wise dependencies through channel self-attention and spatial projection mechanisms. Both modules employ parallel attention streams with residual connections.
Figure 4. Architecture of Cross-Dimensional Attention (CDA, left panel) and Cross-Path Attention (CPA, right panel) modules. CDA captures spatial correlations through spatial self-attention, frequency projection, and channel projection with sigmoid gating. CPA models channel-wise dependencies through channel self-attention and spatial projection mechanisms. Both modules employ parallel attention streams with residual connections.
Information 17 00103 g004
Figure 5. Architecture of the Adaptive Feature Fusion Module (AFFM). The module processes RGB features F g and infrared features F l through parallel branches with 1 × 1 convolution, batch normalization, and ReLU activation. Sigmoid-activated gating weights enable adaptive fusion through element-wise multiplication and subtraction operations. The fused features are concatenated and refined through 3 × 3 convolution with BN and ReLU to produce the output Z.
Figure 5. Architecture of the Adaptive Feature Fusion Module (AFFM). The module processes RGB features F g and infrared features F l through parallel branches with 1 × 1 convolution, batch normalization, and ReLU activation. Sigmoid-activated gating weights enable adaptive fusion through element-wise multiplication and subtraction operations. The fused features are concatenated and refined through 3 × 3 convolution with BN and ReLU to produce the output Z.
Information 17 00103 g005
Figure 6. Architecture of the Dual-Attention Decoupling Module (DADM). The left panel shows the overall structure with hierarchical dilated convolutions (D.C.) at rates r { 1 , 2 , 4 , 4 } , parallel attention streams, and the Dual-Attention Module (DAM). The right panel details the DAM architecture, which employs pointwise convolutions (PW Conv), reshape operations, and fully connected (FC) layers to compute spatial attention through query–key multiplication.
Figure 6. Architecture of the Dual-Attention Decoupling Module (DADM). The left panel shows the overall structure with hierarchical dilated convolutions (D.C.) at rates r { 1 , 2 , 4 , 4 } , parallel attention streams, and the Dual-Attention Module (DAM). The right panel details the DAM architecture, which employs pointwise convolutions (PW Conv), reshape operations, and fully connected (FC) layers to compute spatial attention through query–key multiplication.
Information 17 00103 g006
Figure 7. Multi-dimensional performance comparison using radar chart. The chart compares DCAM-DETR with representative methods across six evaluation dimensions: detection accuracy (mAP@0.5 and mAP@0.5:0.95), computational efficiency (Speed and Params Efficiency), and challenging scenario performance (Small Target and Night Scene). DCAM-DETR achieves the largest coverage area, indicating superior overall performance across all dimensions.
Figure 7. Multi-dimensional performance comparison using radar chart. The chart compares DCAM-DETR with representative methods across six evaluation dimensions: detection accuracy (mAP@0.5 and mAP@0.5:0.95), computational efficiency (Speed and Params Efficiency), and challenging scenario performance (Small Target and Night Scene). DCAM-DETR achieves the largest coverage area, indicating superior overall performance across all dimensions.
Information 17 00103 g007
Figure 8. Qualitative detection results on Anti-UAV300 dataset. The figure shows detection results across diverse challenging scenarios including nighttime conditions (rows 1–2), daytime with complex urban backgrounds (rows 3–4), and varying target scales. Each image displays the detected UAV with bounding box and confidence score. DCAM-DETR consistently produces accurate detections across diverse conditions, demonstrating robust multimodal fusion capability. Note: The Chinese characters visible in some images are metadata overlays from the camera equipment (date, time, and camera settings), which are inherent to the original captured images and represent authentic real-world data conditions.
Figure 8. Qualitative detection results on Anti-UAV300 dataset. The figure shows detection results across diverse challenging scenarios including nighttime conditions (rows 1–2), daytime with complex urban backgrounds (rows 3–4), and varying target scales. Each image displays the detected UAV with bounding box and confidence score. DCAM-DETR consistently produces accurate detections across diverse conditions, demonstrating robust multimodal fusion capability. Note: The Chinese characters visible in some images are metadata overlays from the camera equipment (date, time, and camera settings), which are inherent to the original captured images and represent authentic real-world data conditions.
Information 17 00103 g008
Figure 9. Detection performance comparison under various environmental conditions. The bar chart shows mAP@0.5 for different methods across five environmental scenarios plus overall performance, while the purple dashed line indicates the improvement of DCAM-DETR over RGB-only RT-DETR. DCAM-DETR maintains consistently high performance across all conditions, with the most significant improvements observed in challenging nighttime and foggy scenarios where RGB imagery alone provides limited discriminative information.
Figure 9. Detection performance comparison under various environmental conditions. The bar chart shows mAP@0.5 for different methods across five environmental scenarios plus overall performance, while the purple dashed line indicates the improvement of DCAM-DETR over RGB-only RT-DETR. DCAM-DETR maintains consistently high performance across all conditions, with the most significant improvements observed in challenging nighttime and foggy scenarios where RGB imagery alone provides limited discriminative information.
Information 17 00103 g009
Figure 10. Attention heatmap visualization. The figure shows RGB images (rows 1, 3), infrared images (rows 2, 4), and corresponding attention heatmaps. The heatmaps demonstrate that DCAM-DETR effectively focuses on UAV regions while suppressing background distractions, with complementary attention patterns between RGB and infrared modalities. Note: The Chinese characters visible in some images are metadata overlays from the camera equipment (date, time, and camera settings), which are inherent to the original captured images and represent authentic real-world data conditions.
Figure 10. Attention heatmap visualization. The figure shows RGB images (rows 1, 3), infrared images (rows 2, 4), and corresponding attention heatmaps. The heatmaps demonstrate that DCAM-DETR effectively focuses on UAV regions while suppressing background distractions, with complementary attention patterns between RGB and infrared modalities. Note: The Chinese characters visible in some images are metadata overlays from the camera equipment (date, time, and camera settings), which are inherent to the original captured images and represent authentic real-world data conditions.
Information 17 00103 g010
Table 1. Comparison of DCAM-DETR with existing Mamba-based methods.
Table 1. Comparison of DCAM-DETR with existing Mamba-based methods.
MethodApplicationModalityFusion StrategyKey Innovation
Mamba-UNetMedical Seg.SingleN/AU-Net with SSM
VMambaClassificationSingleN/ACross-scan 2D
Vision MambaGeneral VisionSingleN/ABidirectional SSM
FusionMambaImage FusionMultiEarlyDual-stream SSM
DCAM-DETRAnti-UAV Det.RGB + IRAdaptiveCDA/CPA + AFFM
Table 2. Performance comparison on Anti-UAV300 dataset. Best results are in bold, second best are underlined. indicates methods retrained on Anti-UAV300 with official implementations.
Table 2. Performance comparison on Anti-UAV300 dataset. Best results are in bold, second best are underlined. indicates methods retrained on Anti-UAV300 with official implementations.
MethodModalitymAP@0.5mAP@0.5:0.95FPSParams(M)
Single-Modality Methods (RGB)
YOLOv5-L [35]RGB78.354.26846.5
YOLOv7 [37]RGB81.557.85237.2
YOLOv8-L [66]RGB82.858.95843.7
YOLOv9-C [38]RGB83.259.44851.8
YOLOv10-L [39]RGB84.160.75544.3
RT-DETR-L [26]RGB85.762.14542.3
Single-Modality Methods (Infrared)
YOLOv5-L [35]IR74.651.36846.5
YOLOv7 [37]IR77.853.95237.2
YOLOv8-L [66]IR79.455.65843.7
RT-DETR-L [26]IR82.458.74542.3
Multimodal Fusion Methods
DenseFuse [17]RGB + IR84.960.33552.1
U2Fusion [44]RGB + IR86.261.83248.7
TarDAL [14]RGB + IR88.564.22855.3
M3FD [14]RGB + IR90.367.53849.8
CFT [67]RGB + IR91.269.83158.2
TransFuse [47]RGB + IR92.171.22563.4
Recent State-of-the-Art (CVPR/ICCV/ECCV 2023-2024)
MBNet (ICCV’23)RGB + IR90.568.93251.2
BAANet (ICCV’23)RGB + IR91.269.83054.6
CAT-Det (CVPR’24)RGB + IR91.870.42856.8
SuperYOLO (ECCV’24)RGB + IR92.372.13548.9
CIAN (CVPR’24)RGB + IR92.873.52661.3
DCAM-DETR (Ours)RGB + IR94.778.34247.6
Table 3. Component-wise ablation study on Anti-UAV300 dataset. Each row progressively adds one component to the baseline.
Table 3. Component-wise ablation study on Anti-UAV300 dataset. Each row progressively adds one component to the baseline.
ConfigurationmAP@0.5mAP@0.5:0.95FPSParams(M)
Baseline (RT-DETR + Concat)87.263.54445.1
+ MobileMamba Backbone90.368.74346.8
+ CDA & CPA92.573.14247.2
+ AFFM93.876.44247.5
+ DADM (Full Model)94.778.34247.6
Table 4. Modality ablation study on Anti-UAV300 dataset.
Table 4. Modality ablation study on Anti-UAV300 dataset.
ConfigurationmAP@0.5mAP@0.5:0.95FPSParams(M)
RGB-only (Standard)85.762.14542.3
RGB-only (2× Capacity)87.364.13885.2
IR-only (Standard)82.458.74542.3
IR-only (2× Capacity)84.660.83885.2
DCAM-DETR (RGB + IR)94.778.34247.6
Table 5. Comparison of different backbone architectures on Anti-UAV300 dataset.
Table 5. Comparison of different backbone architectures on Anti-UAV300 dataset.
BackbonemAP@0.5mAP@0.5:0.95FPSParams(M)
ResNet-50 [68]88.464.84844.2
ResNet-101 [68]89.166.24263.1
Swin-T [24]90.869.43548.6
Swin-S [24]91.571.22869.3
VMamba-T [30]92.373.84046.2
MobileMamba (Ours)94.778.34247.6
Table 6. Comparison of different fusion strategies on Anti-UAV300 dataset.
Table 6. Comparison of different fusion strategies on Anti-UAV300 dataset.
Fusion StrategymAP@0.5mAP@0.5:0.95Params(M)
Early Fusion (Concat)87.263.545.1
Late Fusion (Ensemble)88.665.490.2
Attention Fusion [45]90.468.946.3
Vanilla Cross-Attention90.870.248.5
Cross-Attention [47]91.871.552.8
Simple Weighted Average89.467.345.8
CDA + CPA + AFFM (Ours)94.778.347.6
Table 7. DADM component analysis on Anti-UAV300 dataset.
Table 7. DADM component analysis on Anti-UAV300 dataset.
Detection Head ConfigurationmAP@0.5mAP@0.5:0.95
Standard Head92.573.1
+ Dilated Convolutions only93.274.8
+ Dual-Attention Module (DAM)94.176.9
+ Full DADM (Ours)94.778.3
Table 8. Performance breakdown by UAV size and scene condition on Anti-UAV300.
Table 8. Performance breakdown by UAV size and scene condition on Anti-UAV300.
CategoryDCAM-DETRRT-DETR (RGB)Improvement
By Target Size (mAP@0.5)
Small (<32 × 32 pixels)89.471.2+18.2
Medium (32–96 pixels)95.886.4+9.4
Large (>96 × 96 pixels)97.393.1+4.2
By Scene Condition (mAP@0.5)
Daytime Clear96.291.2+5.0
Daytime Cloudy95.488.7+6.7
Dusk/Dawn94.182.3+11.8
Nighttime Clear93.868.5+25.3
Nighttime Foggy91.258.4+32.8
Table 9. Per-class AP analysis on Anti-UAV300 dataset.
Table 9. Per-class AP analysis on Anti-UAV300 dataset.
UAV TypeCountAP@0.5AP@0.5:0.95
Quadcopter15695.279.1
Fixed-wing7893.876.8
Helicopter4294.177.5
Micro UAV2491.372.4
Overall30094.778.3
Table 10. Cross-dataset evaluation results (models trained on Anti-UAV300, tested on other datasets without fine-tuning).
Table 10. Cross-dataset evaluation results (models trained on Anti-UAV300, tested on other datasets without fine-tuning).
MethodFLIR-ADAS mAP@0.5KAIST mAP@0.5Average
RT-DETR [26]72.368.770.5
M3FD [14]75.871.273.5
CFT [67]77.273.875.5
TransFuse [47]78.474.676.5
DCAM-DETR (Ours)81.277.979.6
Table 11. Computational efficiency comparison on Anti-UAV300 dataset.
Table 11. Computational efficiency comparison on Anti-UAV300 dataset.
MethodmAP@0.5FPSParams(M)FLOPs(G)Memory(GB)
YOLOv8-L [66]82.85843.7165.24.2
RT-DETR-L [26]85.74542.3136.85.1
TransFuse [47]92.12563.4248.68.7
DCAM-DETR (Ours)94.74247.6142.35.8
Table 12. Edge device deployment on NVIDIA Jetson AGX Orin.
Table 12. Edge device deployment on NVIDIA Jetson AGX Orin.
ConfigurationmAP@0.5FPSPower (W)
DCAM-DETR (FP32)94.71655
DCAM-DETR (FP16)94.52448
DCAM-DETR (INT8)93.23144
DCAM-DETR (Pruned 50%)92.12846
YOLOv8-L (INT8)81.25238
RT-DETR-L (INT8)84.13542
Table 13. Failure case analysis on Anti-UAV300 dataset.
Table 13. Failure case analysis on Anti-UAV300 dataset.
Failure CategorymAP@0.5Failure Rate
Extreme Occlusion (>70%)67.318.2%
Severe Motion Blur72.112.4%
Thermal Camouflage69.88.7%
Very Small Targets (<10 pixels)58.415.3%
CDA Spatial Misalignment-12.8%
CPA Channel Confusion-15.4%
AFFM Wrong Modality Weight-10.9%
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Qin, Z.; Li, Y. DCAM-DETR: Dual Cross-Attention Mamba Detection Transformer for RGB–Infrared Anti-UAV Detection. Information 2026, 17, 103. https://doi.org/10.3390/info17010103

AMA Style

Qin Z, Li Y. DCAM-DETR: Dual Cross-Attention Mamba Detection Transformer for RGB–Infrared Anti-UAV Detection. Information. 2026; 17(1):103. https://doi.org/10.3390/info17010103

Chicago/Turabian Style

Qin, Zemin, and Yuheng Li. 2026. "DCAM-DETR: Dual Cross-Attention Mamba Detection Transformer for RGB–Infrared Anti-UAV Detection" Information 17, no. 1: 103. https://doi.org/10.3390/info17010103

APA Style

Qin, Z., & Li, Y. (2026). DCAM-DETR: Dual Cross-Attention Mamba Detection Transformer for RGB–Infrared Anti-UAV Detection. Information, 17(1), 103. https://doi.org/10.3390/info17010103

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Article metric data becomes available approximately 24 hours after publication online.
Back to TopTop