Next Article in Journal
Path Planning for Dragon-Fruit-Harvesting Robotic Arm Based on XN-RRT* Algorithm
Previous Article in Journal
A Mobile Wireless Sensor Coverage Optimization Method for Bridge Monitoring
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

REDETR-RISTD: Real-Time Long-Range Infrared Small Target Detection Network Based on the Reparameterized Efficient Detection Transformer

by
Ning Li
*,† and
Daozhi Wei
Air Defense and Antimissile School, Air Force Engineering University, Xi’an 710051, China
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Sensors 2025, 25(9), 2771; https://doi.org/10.3390/s25092771
Submission received: 31 March 2025 / Revised: 16 April 2025 / Accepted: 22 April 2025 / Published: 27 April 2025
(This article belongs to the Section Remote Sensors)

Abstract

:
The critical challenge of detecting infrared small targets at long ranges is that accuracy is compromised. This happens because the targets are small in size, have a weak signal-to-noise ratio (SNR), and are surrounded by complex backgrounds. A novel real-time long-range infrared small target detection network based on the Reparameterized Efficient Detection Transformer (REDETR-RISTD) is proposed. REDETR-RISTD maintains accuracy while significantly reducing computational complexity. First, we introduce a self-developed reparameterized multi-scale feature extraction module (RMSFE). This module helps to construct the lightweight RepEMSNet backbone. It substantially reduces model parameters while maintaining detection capabilities. Second, we design an Attention-Based Intra-Scale Contextual Features Interaction (AICFI) module within the hybrid encoder. This module enhances focus on infrared small targets. It also improves feature interaction across scales. Third, we implement a multi-scale pyramid feature fusion network (MSPFN) with bidirectional fusion mechanisms. This architecture helps to better capture and enhance small target features. Experimental results across three representative public datasets demonstrate the effectiveness of our approach. Compared to state-of-the-art (SOTA) models, REDETR-RISTD has only 13.814 M parameters. It achieves competitive performance with AP50 and recall rates of 96.5% and 92.7%, 84.3% and 83.3%, and 98.5% and 97.9%, respectively. REDETR-RISTD successfully balances the trade-off between detection accuracy and computational efficiency.

1. Introduction

Infrared small target detection technology has emerged as a critical tool across diverse operational domains, encompassing maritime reconnaissance, aerial surveillance, emergency response, environmental observation, and clinical diagnostics. This technology’s significance derives from its exceptional concealment capabilities, resistance to interference, and consistent functionality across all environmental conditions, establishing it as a fundamental component in contemporary monitoring systems [1,2,3,4].
The application landscape of infrared imaging extends well beyond these traditional domains. In the transportation sector, pioneering work by Bertozzi et al. [5] has yielded thermal night vision systems capable of identifying roadway hazards and pedestrians during limited visibility, substantially improving vehicular safety parameters. The behavioral sciences have similarly capitalized on this technology, with Koukiou and Anastassopoulos [6] establishing that thermal facial signatures provide reliable indicators of alcohol consumption, while subsequent research by Koukiou [7] has expanded these methodologies to broader cognitive state evaluation. Ecological applications have incorporated infrared systems for fauna monitoring and conservation initiatives, with particular emphasis on early thermal detection of forest conflagrations as elucidated by Pastor [8]. In healthcare contexts, Ring and Ammer [9] have documented the effectiveness of infrared imaging in non-invasive identification of inflammatory conditions and vascular irregularities. Currently, recent innovations documented by Koukiou [10] have advanced biometric security protocols through thermal facial vasculature mapping. This expansive functional spectrum underscores infrared imaging’s transformative capacity across scientific and practical domains.
The increasing requirement for precision in infrared small target detection has necessitated algorithmic advancement to overcome inherent limitations including diminutive target dimensions, suboptimal signal-to-noise ratios, and complex background interference [1,2]. As illustrated in Figure 1, infrared targets typically manifest as minimal, low-intensity signatures with substantial background similarity. Consequently, this investigation explores sophisticated infrared small target detection algorithms, with the objective of developing optimized methodologies that enhance detection precision, minimize false positive indicators, and improve infrared system efficacy across practical implementation scenarios.
Traditional infrared small target detection methods are mostly model-driven and include filtering-based, local contrast-based, and tensor-based approaches. For example, Lizhen Deng et al. used quantum genetic algorithms to enhance the top-hat filter for adaptive detection [11], and Yi Cui et al. combined heterogeneity filter convolution with hollow side window filtering and patch contrast measures for precise target identification [12]. These methods have been effective in certain specific applications but often require integration with additional techniques to optimize performance. X Zhang et al. and Zhonghua Wang combined various filtering and contrast mechanisms to differentiate targets from backgrounds and reduce false alarms [13,14], while Yongsong Li et al. and Xiangyue Zhang et al. introduced new filtering methods and multi-scale templates to improve detection in complex backgrounds [15,16].
However, these methods have limitations in complex scenes, particularly when there is significant noise and clutter, which reduces robustness. The complex background often complicates the extraction and utilization of local contrast information, and slight fluctuations in hyperparameters can significantly impact detection performance, limiting their generalization ability.
In recent years, deep learning, especially convolutional neural network (CNN)-based methods, has brought significant breakthroughs in infrared small target detection. For instance, Yu et al. proposed attention mechanisms that effectively guide shallow features to deep features, enhancing feature fusion [17], while Tong et al. introduced the EA-ATT module to improve the interaction between deep and shallow features [18]. Li et al. introduced the DNANet network, which uses densely nested attention networks to progressively fuse features [19]. These methods have improved detection accuracy but still face challenges, especially in long-range infrared small target detection, due to small target sizes, low contrast, and high background complexity [20].
Compared to traditional methods, deep learning-based techniques directly learn features from the raw image, offering stronger adaptability and generalization capabilities [20,21]. For example, Wang et al. and Zhao et al. employed generative adversarial networks (GANs) to balance missed detection and false alarm rates, focusing on noise suppression and target feature enhancement [21,22], while Zhou et al. proposed a competitive framework that achieves a Nash equilibrium between false negatives and false positives [23]. Additionally, Zhao et al.’s TBC-Net [24] and Hu et al.’s ST-Net [25] contributed significantly to small target detection by introducing residual connections and smoothing propagation.
Despite the significant improvement in accuracy, these deep learning-based methods still face the challenge of balancing real-time detection with high precision, especially for small targets. Therefore, we propose REDETR-RISTD, a single-stage infrared small target detection network based on the Reparameterized Efficient Detection Transformer. REDETR-RISTD combines the strengths of existing methods and optimizes them to enhance detection accuracy while maintaining high real-time performance.
The main contributions of this paper are summarized as follows:
(1) We propose a single-stage infrared small target detection method, REDETR-RISTD, which achieves real-time, high-precision detection in complex backgrounds and significantly reduces missed detections.
(2) We introduce a lightweight backbone network, RepEMSNet, built upon the RMSFE module. RepEMSNet enhances multi-scale feature extraction, improving the receptive field for small targets while reducing inference time through structural reparameterization.
(3) In the encoder, we use the AICFI module to capture salient features of infrared small targets and employ MSPFN for bottom-up feature fusion and top-down feature enhancement, thereby boosting the network’s detection ability.
(4) Experimental results demonstrate that REDETR-RISTD outperforms other SOTA detectors on three datasets, showing vast potential for applications in remote sensing and early warning detection.

2. Related Work

2.1. Vision Transformer (ViT)

Dosovitskiy et al. [26] first applied the Transformer [27] to computer vision by introducing the vanilla Vision Transformer (ViT). However, when applied to broader computer vision tasks such as object detection and semantic segmentation, vanilla ViT encountered several challenges. In particular, it struggled to capture local feature information and achieve precise target localization, required extensive data for training, and had a complex, computationally expensive structure.
To overcome these limitations, subsequent research has focused on enhancing ViT’s capabilities. Beal et al. [28] adopted a hybrid approach by integrating ViT with CNNs to create ViT-FRCNN, which leverages the spatial feature maps generated by ViT within the Faster R-CNN framework to improve detection efficiency. Fang et al. [29] further extended this idea by proposing YOLOS, a series of ViT-based object detection models, demonstrating the versatility of pre-trained Transformers in object detection through modifications in label and loss functions.
Li et al. [30] introduced complex masked image modeling (MIM) pre-training to boost ViT’s performance in object detection. Building on this, Fang et al. [31] proposed MIMDET, which utilizes MIM pre-trained ViT, replaces patchify stems with compact convolutional stems, and generates multi-scale feature maps from a single-scale ViT, leading to significantly faster convergence. In another line of work, Lou et al. [32] focused on multi-scale ViT to enhance defect target detection accuracy on printed circuit boards, achieving high detection accuracy with their approach.
Other contributions include Wang et al. [33], who tailored a visual transformer detector (ViTDet) for aerial image object detection and outperformed advanced CNNs on multiple datasets. In contrast, Motional et al. [34] argued that deeper models do not necessarily yield better results in 3D object detection and thus developed efficient network training strategies to reduce inference time while maintaining high performance. Zhang [35] proposed an efficient inductive visual Transformer framework and validated its effectiveness in optical remote sensing image detection tasks.
More recently, Li et al. [36] introduced the pyramid convolutional visual transformer (PCViT), which leverages a pyramid architecture and self-supervised pre-training to better capture multi-scale information in remote sensing images. Gong et al. [37] proposed a lightweight detection network combining Group Convolution, ShuffleNetV2, and Vision Transformer to optimize efficiency. Gao et al. [38] designed the spatio-temporal aggregation transformer (STAT) for neuromorphic datasets by incorporating density-based adaptive sampling, sparse event tensors, and lightweight triaxial vision transformers, achieving high accuracy.
While ViT-based models enhance feature representation and capture global context, they often struggle with local details and efficiency issues.

2.2. Detection Transformer (DETR)

Carion et al. [39] first introduced DETR, an end-to-end detector based on Transformer, which completely abandons traditional manual anchor points and the complex non-maximum suppression (NMS) component. Instead, DETR directly predicts one-to-one object sets through a binary matching mechanism. Although DETR offers several advantages, it suffers from slow convergence, high computational cost, and complicated query optimization.
To address the slow convergence, Zhu et al. [40] proposed a deformable attention module to improve the training convergence of multi-scale feature fusion. Liu et al. [41] introduced a dynamic anchor-based query formulation that uses box coordinates to enhance the similarity between queries and features, thereby alleviating the slow training issue in DETR. Li et al. [42] further proposed using real noisy bounding boxes based on DETR and Deformable DETR, which effectively reduced the difficulty of bipartite graph matching and accelerated convergence. Chen et al. [43] introduced a training method using one-to-many assignments within groups, coupled with separate decoder self-attention, to provide extra supervision and further speed up convergence. Additionally, Meng et al. [44] learned conditional space queries from decoder embeddings, allowing each cross-attention head to focus on different regions, which also contributed to faster convergence.
To tackle the high computational cost, Roh et al. [45] proposed a sparse encoder token and applied an auxiliary detection loss to it, thereby reducing computational burden while enhancing detection performance. Yao et al. [46] replaced the traditional six decoders with a single decoder, significantly boosting detection efficiency. Li et al. [47] improved the encoder’s efficiency by interleaving the update of high-level and low-level features, thereby increasing the reliability of predicted attention weights. Regarding query optimization, Chen et al. [48] represented object queries as box queries, reducing optimization difficulty while ensuring convergence speed. Wang et al. [49] introduced the concept of anchors for object queries, allowing for better query optimization. Finally, Zhao et al. [50] decoupled intra-scale interactions from cross-scale fusion to quickly process multi-scale features, improved encoding efficiency and convergence speed, and proposed queries with minimal uncertainty, thus significantly enhancing the quality of initial queries.
DETR-based approaches, though innovative in removing manual components like anchors and NMS, still face challenges in convergence speed and computational cost.

2.3. Loss Function for Infrared Small Target Detection

A key challenge in infrared small target detection is the severe class imbalance between the target and background pixels. To address this, researchers have developed specialized loss functions that balance localization accuracy and computational efficiency, while tackling challenges such as edge ambiguity and scale sensitivity.
In segmentation-based frameworks, precise edge delineation is critical. ISNet [51] introduces EdgeLoss, which independently computes binary cross-entropy and Dice losses for boundary regions. However, in low-contrast and noisy environments, this method may not robustly distinguish subtle edge features and can lead to misclassification. Similarly, Lv et al. [52] improved YOLOv3 for infrared targets by redesigning the edge loss function and network architecture. However, its high sensitivity to gradient changes may cause instability during training.
For bounding box-based detection, multi-component loss functions are widely adopted. Mao et al. [53] employed SoftIoU loss during training to handle ambiguous or uncertain object boundaries, leveraged Sparse SoftIoU Loss to simultaneously focus on coarse-grained and fine-grained information, and utilized Orthogonality Regularization Loss to enhance the model’s representational capacity and generalization ability. OSCAR [54] addressed class imbalance by employing Focal Loss, enhanced bounding box localization accuracy using IoU loss, and tackled the challenge of assessing small object localization quality with Quality Focal Loss. Mou et al. [55] further refined localization by applying the Complete IoU loss to multi-scale detection heads. The CIoU loss is defined as Equation (1).
L CIoU = 1 IoU + ρ 2 ( b , b g t ) c 2 + α v
where ρ is the Euclidean distance between bounding box centers, c is the diagonal length of the minimum enclosing box, and α v penalizes aspect ratio discrepancies, effectively addressing scale sensitivity in small target regression.
Recent advances leverage geometric distribution metrics to enhance similarity measurement. Zhu et al. [56] proposed the Normalized Wasserstein Distance (NWD) loss, which models predicted and ground-truth boxes as 2D Gaussian distributions N ( μ p , Σ p ) and N ( μ g t , Σ g t ) . The NWD loss computes their Wasserstein distance as Equation (2).
L NWD = exp 2 W 2 2 ( N p , N g t ) C
where W 2 2 is the squared Wasserstein distance and C is a normalization constant. This metric robustly evaluates similarity for non-overlapping or nested boxes, overcoming limitations of traditional IoU. Building on this, Zhao et al. [57] integrated NWD with Inner IoU, which focuses on the overlap between the inner regions of bounding boxes, as shown in Equation (3).
L Inner IoU = 1 Inner IoU IoU
where Inner-IoU is computed by shrinking boxes to avoid boundary noise. The hybrid loss L NWD + L Inner IoU accelerates convergence and improves precision in cluttered scenes.
But the NWD loss is sensitive to the selection of the normalization constant. It may require fine-tuning in different scenarios. Moreover, in tiny target detection, the local features of the targets are not distinct enough. Therefore, distribution-based modeling still has some limitations.
In summary, multi-component loss functions combine both coarse-grained and fine-grained information. They improve target localization accuracy to some extent. However, they also introduce multiple loss terms, which increases the complexity of parameter tuning. For extremely small targets with low contrast against the background, traditional loss functions still cannot fully capture the true shape of the target. This results in limited regression accuracy.

3. Materials and Methods

3.1. The REDETR-RISTD Model Architecture

The overall structure of the network is shown in Figure 2. The architecture of REDETR-RISTD consists of three key components: Backbone, Encoder, and Decoder.
In Backbone RepEMSNet, assuming that the input infrared image is I R C × H × W , downsampling is performed using convolution with a kernel of 3 × 3 and a step size of 2 to extract fine-grained localized features in a more extensive range of the image. Reparameterized Convolution (Repconv) in the Reparameterized Multi-scale Feature Extraction (RMSFE) module makes the entire backbone network structure reparameterizable, improving the inference speed of the network without loss of accuracy. Each REMSFE can effectively and progressively increase the receptive field of the neural network by stacking the convolution kernels layer-by-layer as a 3 × 3 convolution while efficiently utilizing the number of parameters [58]. Feature extraction is performed through multiple branches and scales to obtain feature extraction maps with image resolutions of (H/4, W/4), (H/8, W/8), (H/16, W/16), and (H/32, W/32), respectively.
In Encoder’s Attention-based Intra-scale Feature Interaction (AIFI), we improve it by utilizing the HiLo attention mechanism [59], and the resulting AICFI module captures local details and global dependencies in the deep backbone feature map in parallel, which improves computational efficiency and detection accuracy while enhancing the robustness of the model.
A new feature fusion framework, the Multi-scale Pyramid Fusion Network (MSPFN), is designed, in which the Multi-scale Fusion and Enhancement Feature (MSFEF) module can capture and fuse the multi-scale information of the small targets in the input feature map to increase feature diversity while maintaining feature information integrity, enabling the network to learn better and understand the small target features. The MSFEF module captures and fuses the multi-scale information of small targets in the input feature map to increase feature diversity while maintaining the integrity of the feature information, which enables the network to learn better and understand the small target features. The Multi-scale Feature Fusion Downsampling Module (MSF-FDM) retains the multi-scale information of infrared small target features while reducing the size of the feature map. It focuses on extracting local features of small targets and can effectively capture global contextual information. The Multi-scale Feature Fusion Upsampling Module (MSF-FUM) enhances the representation of feature maps by multipath feature extraction and fusion during upsampling, enabling efficient upsampling while maintaining detailed and global information. Finally, a new loss function, EnIoU loss, is designed to accelerate the convergence of network training.

3.2. Improved Feature Extraction Network RepEMSNet

Figure 3 clearly shows the architecture design of the proposed backbone network, RepEMSNet. The network input is defined as a 3D tensor X 0 R H 0 × W 0 × 3 , where H 0 , W 0 represents the spatial resolution of the infrared image. First, to rapidly compress redundant spatial data to effectively suppress interference from large homogeneous background regions in infrared images, while selectively extracting low-frequency background information encompassing scene geometric contours and temperature gradient distribution patterns, the input infrared image tensor X 0 R H 0 × W 0 × 3 undergoes initial feature extraction and downsampling via a convolutional module parameterized as ( 3 , 64 ; 3 , 2 ) , as shown in Equation (4).
P 1 = C ( 3 , 64 ; 3 , 2 ) ( X 0 )
where C ( c in , c out ; k , s ) denotes the standard convolutional operation in Equation (5).
C ( x ) = σ ( BN ( Conv 2 D ( c in , c out , k , s ) ( x ) ) )
where σ denotes the SiLU activation function, Conv 2 D represents the standard 2D convolution operation, and BN stands for batch normalization (BN). The symbols c in , c out , k , s represent the input channels, output channels, kernel size, and stride of the convolutional layer, respectively. In the function, x is the independent variable and does not represent any specific numerical value. All subsequent occurrences of x retain this definition. This operation reduces the resolution of X 0 to H 0 / 2 × W 0 / 2 with 64 output channels.
Next, to enhance feature diversity for addressing the ambiguity of infrared small targets caused by low SNR, while emphasizing critical characteristics such as target edge sharpness and local contrast enhancement, further downsampling is performed using a convolutional layer parameterized as ( 64 , 128 ; 3 , 2 ) , as shown in Equation (6). By implementing spatial compression, it reduces computational complexity in subsequent stages while preserving the positional information of targets.
P 2 = C ( 64 , 128 ; 3 , 2 ) ( P 1 )
where the output feature map P 2 R H 0 4 × W 0 4 has 128 channels.
To further enhance critical target characteristics while maintaining real-time processing capabilities and suppressing interference from complex backgrounds, the features P 2 are directed into the RMSFE module for enhancement through the following steps in Equations (7) to (12).
(1) Channel Expansion and Splitting. The input P 2 undergoes 1 × 1 convolution parameterized as ( 128 , 128 ; 1 , 1 ) in Equation (7).
F = C ( 128 , 128 ; 1 , 1 ) ( P 2 )
F is then split into two components along the channel dimension F ( a ) , F ( b ) in Equation (8).
{ F ( a ) , F ( b ) } = split ( F , 1 )
where F ( a ) , F ( b ) R H 0 4 × W 0 4 × 64 . F ( a ) preserves the integrity of original channel-wise data to prevent small targets from being diluted or lost during complex nonlinear transformations, while F ( b ) serves as an enhancement branch that employs targeted feature amplification mechanisms to significantly improve both the SNR and spatial activation intensity in target regions.
(2) Residual Feature Reconstruction. The branch F ( b ) undergoes processing in Equation (9) via the RepConv module defined in Equation (10). To ensure real-time processing capabilities, the RepConv leverages multi-scale convolutional kernel fusion to amplify edge sharpness and local contrast in target regions while suppressing background clutter.
G 0 = R ( 64 , 32 ; 3 , 1 ) ( F ( b ) )
where R ( · ) during training adopts a multi-branch structure in Equation (10).
R ( x ) = σ C ( 64 , 32 ; 3 , 1 ) ( x ) + C ( 64 , 32 ; 1 , 1 ) ( x ) + BN ( x )
(3) Recursive Convolutional Chain. The feature maps undergo progressive refinement through n 1 layers of recursive convolutional operations as defined in Equation (11), which iteratively enhances multi-scale feature to amplify sub-pixel details and strengthen the discriminative capability of microscopic targets.
G k = C ( 32 , 32 ; 3 , 1 ) ( G k 1 ) , k { 1 , , n 1 }
(4) Multi-Level Feature Fusion. All hierarchical representations are concatenated and dimensionally compressed through the 1 × 1 convolutional operation defined in Equation (12). This architecture synergistically enhances cross-scale feature complementarity, strengthens background clutter suppression capability, and preserves sub-pixel positional accuracy by maintaining native spatial resolution, thus effectively mitigating localization drift caused by conventional downsampling.
P 3 = C ( d total , 128 ; 1 , 1 ) concat ( F ( a ) , { G k } , G fin )
where d total = 64 + 32 ( n + 1 ) and G fin = C ( 32 , 32 ; 1 , 1 ) ( G n 1 ) . The output resolution remains H 0 4 × W 0 4 × 128 .
The preceding sections detail the computational processes of the first three core modules in the RepEMSNet backbone architecture. While subsequent network stages follow analogous feature transformation paradigms, we focus here on enumerating the critical parameters of RepEMSNet as summarized in Table 1.
Notably, the scaling factor s within the RMSFE module regulates the receptive fields and feature diversity through dynamic adjustment of intermediate channel dimensions. Specifically, s = 1 amplifies high-frequency feature extraction capability, enhancing sensitivity to infrared small targets by expanding intermediate channels to 384. s = 0.5 optimizes background suppression, reducing the false alarm rate (FAR) through channel compression. The multi-scale output feature maps generated by the backbone network RepEMSNet are hierarchically denoted as S1, S2, S3, S4, and S5, corresponding to spatial–semantic information at progressively increasing receptive fields.

3.3. Improved Feature Interaction AICFI

Existing studies [50] have demonstrated that applying self-attention mechanisms to high-level semantic-rich features can effectively model conceptual entity relationships, thereby enhancing object localization and recognition capabilities in downstream modules. However, conventional self-attention frameworks face critical limitations in infrared small target detection: (1) Standard attention mechanisms struggle to distinguish between high-frequency edge noise and low-frequency background radiation due to their indiscriminate spectral treatment. (2) Fixed-window local attention mechanisms lack adaptability to dynamic target scale variations. (3) The implementation of the global self-attention mechanism requires more computational resources. To overcome these limitations, we present the Attention-guided Intra-scale Contextual Feature Interaction (AICFI) module, a novel hierarchical architecture specifically designed for infrared small target characterization. As illustrated in Figure 4, the AICFI module employs a systematic frequency-aware decomposition-to-fusion paradigm. This innovative architecture employs a frequency-decoupling strategy combined with a dynamic fusion mechanism, implementing a hierarchical attention framework to enhance discriminative feature representation for infrared small targets.
As shown in Figure 4, N h denotes the total number of self-attentive heads in the layer and the ratio of the division between high-frequency and low-frequency heads. α serves as a balancing factor that dynamically adjusts the allocation ratio of attention heads between the high-frequency and low-frequency branches.
First, the input feature map X 1 R c 1 in × H 1 × W 1 is implicitly mapped to a dual-path interaction space through a parameterized feature decomposition strategy, where the high-frequency path (Hi-Fi) and low-frequency path (Lo-Fi) are dynamically allocated computational resources by the hyperparameters α [ 0 , 1 ] and the total number of attention heads N h . The high-frequency path is allocated N h ( 1 α ) heads to focus on fine-grained interactions within local windows, while the low-frequency path is allocated N h α heads to handle sparse global semantic modeling.
Second, the high-frequency path divides the input feature into H 2 s × W 2 s local windows (where s is the window size). Within each window, pixel-level dependency relationships are modeled via multi-head self-attention (MHSA), as shown in Equation (13).
X Hi Fi = i = 1 N h ( 1 α ) Softmax Q i K i T d V i
where ⨁ denotes the channel-wise concatenation of multi-head outputs, and d represents the dimension per attention head. This operation reinforces high-order statistical quantities such as target edges and textures within constrained receptive fields while suppressing high-frequency noise.
Simultaneously, the low-frequency path compresses the feature resolution to H 2 s × W 2 s through spatial downsampling and performs global multi-head attention in the compressed space, as shown in Equation (14).
X Lo Fi = UP s j = 1 N h α Softmax Q ˜ j K ˜ j T d V ˜ j
where UP s · denotes bilinear interpolation upsampling. The low-frequency path enhances semantic discriminability between targets and backgrounds by leveraging sparsified features to filter out local perturbations.
Next, the dual-path features undergo multi-granularity fusion through channel-wise projection and concatenation, as shown in Equation (15).
X 1 fused = P h ( X Hi Fi ) P l ( X Lo Fi )
Here, P h · and P l · are learnable channel projection matrices, and ⊕ represents the concatenation operation. The fused feature retains high-frequency details and low-frequency semantics, forming a complementary representation.
Finally, gradient propagation is stabilized via residual connections and a pre-normalization strategy, yielding the final output X out in Equation (16) of AICFI.
X out = LayerNorm X in + W 2 · GELU ( W 1 · X fused )
where W 1 R c 1 in × 4 c 1 in and W 2 R 4 c 1 in × c 1 in are the weight matrices of the multi-layer perceptron (MLP). This process ensures stable optimization in deep networks, producing the enhanced output feature X out R c 1 in × H 1 × W 1 .

3.4. Feature Fusion Framework MSPFN

In order to solve the problems of detail loss in multi-scale feature fusion, insufficient ability to model cross-scale context, and semantic differences between features at different levels of the feature pyramid in traditional feature fusion methods, we propose a novel feature fusion framework, MSPFN, which combines the pyramid feature fusion principle and incorporates the MSFEF module, MSF-FDM downsampling, and MSF-FUM upsampling. The overall structure of MSPFN is shown in Figure 5.

3.4.1. Multi-Scale Feature Fusion Upsampling Module

The structure of MSF-FUM is shown in Figure 5c. It achieves refined reconstruction of high-resolution features through a sequential process. First, the input feature map X iu R c iu × H iu × W iu undergoes Global Average Pooling (GAP) to compress spatial dimensions, generating channel-wise global statistics X pool R c iu × 1 × 1 . This step captures global semantic information for each channel to guide subsequent channel gating. Second, the global statistics pass through a channel gating module composed of a 1 × 1 convolution and a Hardsigmoid activation function, producing a dynamic weight matrix G iu [ 0 , 1 ] c iu × 1 × 1 as shown in Equation (17).
G iu = σ C ( c iu , c iu ; 1 ) ( GAP ( X iu ) )
where σ denotes the Hardsigmoid activation function. This matrix adaptively regulates channel contributions during feature fusion, suppressing noise and enhancing critical features. Next, the module processes the input through two parallel upsampling branches. Branch 1 uses a transposed convolution with learnable kernels to restore spatial resolution. The output F up 1 is derived by Equation (18).
F up 1 = T ( c iu , c iu / 2 ; 2 , 2 ) ( X iu )
where T ( c in , c out ; k , s ) is defined in Equation (19).
T ( c in , c out ; k , s ) = σ ( BN ( ConvTranspose 2 d ( k , s ) )
Here, σ represents the SiLU activation function. This operation restores the spatial resolution to 2 H iu × 2 W iu while enhancing nonlinearity via BN and SiLU.
Branch 2 combines bilinear interpolation and lightweight convolution. First, the feature map is upsampled to 2 H iu × 2 W iu via bilinear interpolation I ( · ) , followed by a 1 × 1 standard convolutional operation to adjust channels as shown in Equation (20).
F up 2 = C c iu , c iu / 2 ; 1 , 1 ( I ( X iu ) )
where I ( · ) denotes the bilinear interpolation operator. These two branches achieve multi-scale modeling of detailed information from the perspectives of local structure learning and global topology, respectively, forming complementary feature representations. The outputs of both branches are concatenated along the channel dimension to form an intermediate feature F ucat = F upl F up 2 R c in × 2 H iu × 2 W iu . The channel gating weights G iu are broadcast to match the spatial dimensions of F ucat and applied via channel-wise multiplication as shown in Equation (21).
F uweighted = G iu F ucat
This operation dynamically enhances target-related channels and suppresses background noise based on global semantic cues. Finally, the weighted features undergo a standard convolutional operation for channel fusion and nonlinear transformation, producing the high-resolution output F uout in Equation (22).
F uout = C ( c iu , c iu ; 1 ) F uweighted

3.4.2. Multi-Scale Feature Fusion Downsampling Module

The structure of MSF-FDM is shown in Figure 5d. It reduces the spatial resolution of feature maps while retaining key semantic information and suppressing background noise through multi-branch downsampling strategies and adaptive channel gating mechanisms. First, the input feature map X id R c id × H id × W id undergoes GAP to compress spatial dimensions, generating channel-wise global statistics X pool R c id × 1 × 1 . This step captures global semantic information for each channel, providing prior guidance for channel gating.
Second, the global statistics pass through a channel gating module composed of a 1 × 1 convolution and a Hardsigmoid activation function, producing a dynamic weight matrix G id [ 0 , 1 ] c id × 1 × 1 as shown in Equation (23).
G id = σ C ( c id , c id ; 1 ) ( GAP ( X id ) )
where σ denotes the Hardsigmoid function. This matrix G id adaptively regulates the importance of channels during downsampling, suppressing redundant features and enhancing target-related contextual information. Next, the module processes the input through two parallel downsampling branches. Branch 1 uses a standard convolution operation with a 3 × 3 kernel and stride 2 for spatial downsampling in Equation (24).
F down 1 = C ( c id , c id / 2 ; 3 , 2 ) ( X id )
The output feature map F down 1 has dimensions of H / 2 × W / 2 . This branch captures multi-scale context by expanding the local receptive field while reducing resolution. Branch 2 combines MaxPooling and lightweight convolution. First, a 2 × 2 MaxPool operation reduces the spatial size to H / 2 × W / 2 , followed by a 1 × 1 convolution for channel adjustment to obtain F down 2 in Equation (25).
F down 2 = C ( c id , c id / 2 ; 1 ) ( MaxPool ( X id ) )
where MaxPool ( · ) denotes the max pooling operation. This branch preserves salient local structures, achieving channel compression and nonlinear transformation via a standard convolution operation with a 1 × 1 kernel and stride 1. Then, the outputs of both branches are concatenated along the channel dimension to form an intermediate feature F dcat = F down 1 F down 2 , where F dcat R c id × H id / 2 × W id / 2 . The channel gating weights G id are broadcast to match the spatial dimensions of F dcat and applied via channel-wise multiplication as shown in Equation (26).
F dweighted = G id F dcat
This operation adaptively enhances target-related channel responses and suppresses noise based on global semantics. Finally, the weighted features undergo a standard convolution for channel fusion and nonlinear transformation, producing the low-resolution output feature map F dout in Equation (27).
F dout = C ( c dout , c dout ; 1 ) ( F dweighted )

3.4.3. Multi-Scale Feature Fusion Enhancement Module

The structure of MSFEF is illustrated in Figure 5a,b. MSFEF addresses the challenges of varying target scales and complex backgrounds in infrared small target detection through Cross Stage Partial connections and multi-scale parallel atrous convolutions, enabling multi-granularity semantic modeling and efficient fusion of input features.
First, the input feature map X 2 R c 2 in × H 2 × W 2 undergoes channel compression via two independent 1 × 1 standard convolution operations, generating the main branch feature X 21 and residual branch feature X 22 as shown in Equations (28) and (29).
X 21 = C ( c 2 in , c hid ; 1 ) ( X 2 )
X 22 = C ( c 2 in , c hid ; 1 ) ( X 2 )
where c hid = c 2 out · e denotes the hidden layer channels, with e = 0.5 as the default setting.
Second, the main branch feature X 21 is fed into the DPCConv module, where parallel atrous convolutions with different dilation rates d extract multi-scale features Y 1 , Y 2 , and Y 3 in Equations (30) to (32).
Y 1 = C ( c hid , c hid ; 3 , d = 1 ) ( X 21 )
Y 2 = C ( c hid , c hid / 2 ; 3 , d = 2 ) ( X 21 )
Y 3 = C ( c hid , c hid / 2 ; 3 , d = 3 ) ( X 21 )
where dilated convolution with d = 1 mainly focuses on the local details of infrared small targets. Dilated convolution with d = 2 primarily models the local context of the image. And dilated convolution with d = 3 is mainly used to suppress background noise. Next, Y 1 , Y 2 , and Y 3 are concatenated along the channel dimension to form Y cat = Y 1 Y 2 Y 3 R 2 c hid × H 2 × W 2 . A 1 × 1 standard convolution operation then compresses channels and fuses multi-scale information in Equation (33).
Y fused = C ( 2 c hid , c hid ; 1 ) ( Y cat )
This step enhances sensitivity to small target features while suppressing noise by jointly modeling local details and global context. Subsequently, the multi-scale fused feature Y fused is concatenated with the residual branch feature X 22 in Equation (34).
Z cat = Y fused X 22 R 2 c hid × H 2 × W 2
The residual branch preserves low-frequency components to mitigate the signal attenuation caused by multi-scale convolutions. Finally, a 1 × 1 standard convolution adjusts the channels of Z cat to generate the output feature F mout in Equation (35).
F mout = C ( 2 c hid , c mout ; 1 ) ( Z cat ) R c mout × H 2 × W 2
This output integrates multi-scale semantics while preserving critical target details, ensuring robust detection in complex infrared scenes.

3.5. Loss Function EnIoU Loss

The current IoU-based bounding box regression loss (BBRL) mainly accelerates convergence by adding new Loss terms while ignoring the shortcomings of the IoU Loss itself. Since the samples need to be differentiated in the regression process, using auxiliary bounding boxes of different scales to calculate the loss can effectively accelerate the regression process of the bounding box. Therefore, in the model training strategy, we use a smaller auxiliary bounding box to compute the loss for regressions with high-IoU samples and vice versa for low-IoU samples. To address the problem that the Generalized Intersection over Union (GIoU) [60] loss in the traditional bounding box regression loss function in RT-DETR is sensitive to the scale of small targets, we not only take into account the degree of overlap between the prediction bounding box and the ground-truth (GT) box but also the positional accuracy, which makes the prediction bounding box closer to the GT box. Equations (36) to (40) represent the construction process of EnIoU loss.
I = min b r g t , b r max b l g t , b l × min b b g t , b b max b t g t , b t
U = w g t × h g t r a t i o 2 + w × h r a t i o 2 I + ε = U 1 + U 2 + U 3
D 1 = P l t P l t g t 2
D 2 = P r b P r b g t 2
L EnIoU = 1 I U + D 1 2 + D 2 2
where b g t and b denote the GT box and anchor box, respectively, as shown in the red bounding box and green bounding box in Figure 6. min b r g t , b r denotes the right border of the area where the two bounding boxes overlap. max b l g t , b l denotes the left border of the area where the two bounding boxes overlap. min b b g t , b b indicates the lower boundary of the area where the two bounding boxes overlap. max b t g t , b t denotes the upper boundary of the area where the two bounding boxes overlap. I is the area of the portion of the intersection of the two bounding boxes, as shown in the dark filled area of Figure 6. w g t and h g t denote the width and height of the GT box, while the width and height of the prediction bounding box are denoted as w and h, respectively. The r a t i o is the ratio factor that controls the size of the auxiliary box. ε is a regular term. U is the total area occupied by the two bounding boxes, U 1 , U 2 , and U 3 denote the three components of U, and there is I = U 2 , as shown in Figure 6. P l t = ( x l t , y l t ) , P l t g t = ( x l t g t , y l t g t ) , P l b = ( x l b , y l b ) , P r b g t = ( x r b g t , y r b g t ) denotes the top-left and top-right vertices of the predicted and true boxes in Figure 6, respectively. · 2 denotes the Euclidean distance.

4. Results

4.1. Dataset

In this study, we use three open datasets to evaluate the performance of the proposed detector: SIRSTv2 [54], IRSTD-1k [51], and NUDT-SIRST [19]. To ensure a proper evaluation, we divide each dataset into training, validation, and test sets in a ratio of 7:1:2, and the final results are obtained from the test set. Next, we will provide a detailed introduction to the selected datasets.
The SIRSTv2 dataset, proposed by Dai Yimian from Nanjing University of Aeronautics and Astronautics, is a publicly available single-frame infrared small object detection dataset. This dataset contains 1024 typical infrared images, most of which have a resolution of 1280 × 1024, making it one of the highest-resolution datasets available for infrared small target detection. The images in SIRSTv2 are extracted from real-world video sequences, capturing various complex scenarios. These include urban environments where background interference, such as cranes and streetlights, resembles the appearance of targets, making detection more challenging. The dataset reflects the difficulties of infrared small object detection in realistic settings, where most targets are dim and difficult to distinguish from cluttered backgrounds due to low contrast and high noise. Unlike traditional approaches that rely on target saliency or low-rank sparse decomposition, SIRSTv2 requires a more advanced method that can leverage high-level semantic understanding of the entire image to effectively distinguish targets from non-target interference. The distribution of objects in the images is random, and many of the targets are not easily separable from the background, requiring robust contextual information for accurate detection.
The IRSTD-1k dataset, proposed by Zhang Mingjin and colleagues from Xidian University, is a publicly available benchmark dataset for infrared small object detection. The dataset covers a broader range of extreme conditions and challenging environments by integrating multi-source infrared imaging equipment and dynamic acquisitions from real-world scenarios. IRSTD-1k contains 1000 infrared images with a resolution of 512 × 512, annotated with 1520 infrared small objects. The data are sourced from both static and dynamic acquisitions across four major environmental categories: sky backgrounds (static cloud layers, dynamic flying birds), urban backgrounds (building thermal radiation, vehicle exhaust), natural backgrounds (forests, deserts), and extreme weather conditions (haze, rain, snow). The imaging spectrum spans short-wave, mid-wave, and long-wave infrared bands to simulate the imaging characteristics of different sensors. To enhance data diversity, synthetic noise (Gaussian noise, Poisson noise) is added to a subset of images, reducing the SNR. Approximately 40% of the images exhibit an SNR below 2 dB, where background noise nearly overwhelms object signals. The dataset highlights multi-scale object characteristics: approximately 30% of objects are smaller than 3 × 3 pixels (occupying less than 0.01% of the 512 × 512 image area), with some objects being single-pixel points (pixel coverage: 0.02%). The minimum object size is 1 × 1 pixels, while the maximum reaches 15 × 15 pixels. Regarding intensity distribution, only 20% of objects are the brightest regions globally, whereas 80% exhibit grayscale values similar to or lower than the background. The grayscale difference between objects and backgrounds is often below 10 in haze and desert scenarios.
The NUDT-SIRST dataset, developed by researchers from the National University of Defense Technology, is a publicly available synthesized single-frame infrared small target dataset.
Unlike traditional datasets built from object motion sequences, NUDT-SIRST is generated using simulation techniques that ensure highly accurate pixel-level annotations while offering extensive control over target characteristics and background complexity. The dataset comprises 1327 infrared images of 256 × 256 resolution, encompassing a diverse range of target categories and sizes embedded in richly cluttered and noisy backgrounds. Similar to real infrared scenarios, the images contain significant noise and low SNR, with the small targets often occupying less than 0.02% of the total image area—that is, in a standard 256 × 256 image, the target pixels typically cover an area smaller than 4 × 4 pixels. Furthermore, only about 35% of the objects appear as the brightest regions in the image, while the remaining 65% exhibit grayscale values similar to or even lower than their surroundings. This characteristic renders simple saliency-based approaches ineffective and necessitates the use of both global and local contextual information for reliable detection.

4.2. Evaluation Metrics

The three datasets contain only one infrared small target class. Therefore, we use AP50, AP50-95, and recall as precision evaluation metrics; computational and parametric quantities to measure the model’s time complexity and space complexity; and the precision–recall (PR) curve to verify the performance of the algorithm.
Wherein AP denotes average precision, AP50 denotes the average precision when the IoU of the detection model is set to the average precision of 0.5, and AP50-95 denotes the average value of the precision obtained when the IoU of the detection model is selected in the range of 0.5 to 0.95, and the value of AP can be obtained from the area under the PR curve. The unit of parametric quantity is M. Usually, the smaller the model computation and parametric quantity, the higher the FPS of the detection network, though this is not always the case.
Before describing the PR curve, it is necessary to clarify that the infrared small target detection results fall under two categories—target and non-target; therefore, this type of detection problem can be seen as a binary classification problem. Binary classification problems include four kinds of results: TP (true positive, both predicted and actual positive samples), FP (false positive, predicted positive samples, actual negative samples), FN (false positive, predicted negative samples, actual positive samples), and TN (true negative, both predicted and actual negative samples). The horizontal and vertical coordinates of the PR curves are recall (R) and precision (P), respectively. R denotes the proportion of all positive samples that are correctly predicted to be positive samples, and P denotes the proportion of samples predicted to be positive samples that are positive samples. R, P, and AP can be computed by Equations (41) to (43).
R = T P T P + F N
P = T P T P + F P
A P = 0 1 P ( R ) d ( R )
We also used target-level evaluation indicators, the probability of detection ( P d ) and false alarm rate ( F a ), to comprehensively evaluate the detector’s performance from multiple aspects.
Formally, the probability of detection is obtained by Equation (44).
P d = L t p L a l l
where L t p is the number of correctly predicted targets and L a l l is the number of all targets. The probability of a false alarm rate is obtained by Equation (45).
F a = P f P a l l
where P f is the number of targets for the error prediction, and P a l l is the target quantity of all predictions.

4.3. Training Details

The hardware configuration for the experiment is shown in Table 2.
The method proposed in this paper is implemented using Pytorch. The training parameters are set as follows: the image size is uniformly adjusted to 640 × 640. The YOLO series network selects the SGD optimizer with high computational efficiency and fast convergence speed for network training. The RTDETR series uses the AdamW optimizer. The momentum is set to 0.937. The initial learning rate is 0.01. The decay factor is 0.0001, and the training epoch is 500. A learning rate preheating strategy is used to improve the stability and performance of model training and reduce the risk of oscillations and gradient explosion at the beginning of training. The warmup epoch is set to 3, the warmup momentum is set to 0.8, and the learning rate of the warmup bias parameter is set to 0.1.
The boundary box-based object detection method TOOD uses the SGD optimizer with an initial learning rate of 0.0001, final learning rate of 0.0001, decay factor of 0.0001, and momentum of 0.9. Sparse R-CNN uses the AdamW optimizer with an initial learning rate of 0.000025, final learning rate of 0.0001, and decay factor of 0.0001. Mask R-CNN uses the SGD optimizer with an initial learning rate of 0.0002, final learning rate of 0.0001, decay factor of 0.0001, and momentum of 0.9. DINO uses the AdamW optimizer with an initial learning rate of 0.0001, final learning rate of 0.0001, and decay factor of 0.0001. All four methods use the MultiStepLR learning rate scheduler.
The segmentation-based detection methods ACM, ALCNet, DNANet, ISTDU-Net, RDIAN, and OSCAR use the default parameters from the original code. ACM, ALCNet, ISTDU-Net, and RDIAN use the Adam optimizer with a learning rate of 0.0005, and the learning rate scheduler is MultiStepLR for all of them. DNANet uses the Adagrad optimizer with a learning rate of 0.05 and a learning rate scheduler of CosineAnnealingLR for training. OSCAR uses the AdamW optimizer with a learning rate of 0.002 and a decay factor of 0.05.

4.4. Results and Analysis

4.4.1. Quantitative Analysis

The experimental results are shown in Table 3. REDETR-RISTD is compared with advanced target detectors. These include the YOLO series, DETR series, R-CNN series, and TOOD. On the SIRSTv2 dataset, REDETR-RISTD achieves a precision of 0.991, a recall of 0.927, an AP50 of 0.965, and an AP50-95 of 0.516. On the IRSTD-1k dataset, it obtains a precision of 0.991, a recall of 0.979, an AP50 of 0.985, and an AP50-95 of 0.745. REDETR-RISTD outperforms other models on these two datasets.
On the NUDT-SIRST dataset, REDETR-RISTD ranks second in detection metrics. Its recall is 0.979, its AP50 is 0.985, and its AP50-95 is 0.786. It is lower than the RT-DETR-HGNetv2-L model by 0.7% in recall, 1% in AP50, and 2.3% in AP50-95. The model size of REDETR-RISTD is reduced by 56.8% compared to RT-DETR-HGNetv2-L.
These results show that REDETR-RISTD performs very well on small targets. It is designed for detecting small objects. On the NUDT-SIRST dataset, the targets are relatively large. The algorithm does not fully realize its potential in this case, but it still shows strong detection capability. The high performance and lower model size make it a promising solution for object detection.
Figure 7 and Figure 8 show the PR and ROC curves of the three methods on the three datasets, respectively. The following conclusions can be drawn from the curve charts: (1) REDETR-RISTD obtains the best results relatively quickly on all three datasets. (2) REDETR-RISTD achieves the best balance between model parameters, precision, and recall on all three datasets. (3) On all three datasets, REDETR-RISTD can correctly detect more infrared small targets with less computational power.

4.4.2. Qualitative Analysis

To evaluate the performance of different detection methods for infrared small target detection, green and orange boxes are used in the visualization results to represent correct detections and abnormal detections, respectively.
Because the red boxes obtained in different scenes are small and the confidence scores are not clearly visible, green boxes and orange boxes are used instead to indicate correct detections and false or abnormal detections.
The number of green boxes reflects the number of correctly detected targets, while the number of orange boxes represents the number of false or missed detections.
In addition, the number inside the green box indicates the detection confidence, and the number inside the orange box indicates the false detection confidence. Red boxes are used to enclose the target areas.
The visualization results are shown in Figure 9, Figure 10 and Figure 11. The left side of each figure displays the ground-truth target boxes in eight typical scenes. These scenes include dark appearance, blurred background, complex architectural occlusion, cloudy scene, multi-target, bright clutter, high-contrast boundary, and high noise.
Figure 9 shows the ground-truth bounding boxes of targets in eight scenarios on the left and the detection results of YOLO series and REDETR-RISTD on the right.
In the dark-appearance scenario, the target brightness is extremely low. The YOLO series methods failed to detect the target and did not show any detection boxes. In contrast, REDETR-RISTD successfully detected the target with a confidence of 58%. This success is due to its reparameterized multi-scale feature extraction module, which effectively extracts target features in low-signal-to-noise and low-contrast conditions.
In the blurred-background scenario, all methods detected the single target. However, detection confidence varied. YOLOv7 achieved a confidence of 89%, while REDETR-RISTD maintained around 80%. This indicates that REDETR-RISTD has high stability in blurred backgrounds. The attention-guided intra-scale contextual feature interaction module effectively suppresses background interference.
In the complex architectural occlusion scenario, REDETR-RISTD achieved a detection confidence of 88%. This value is significantly higher than the YOLO series methods, which ranged from 70% to 82%. This improvement is mainly attributed to the AICFI module, which optimizes the feature differences between the target and occluded regions. This reduces the impact of building occlusion and enables accurate detection of partially occluded targets.
In the cloudy scenario, REDETR-RISTD reached a detection confidence of 85%, which is clearly higher than YOLOv9m’s 76%. Its advantage comes from the RMSFE module’s multi-scale feature fusion capability. Under cloud interference, this module enhances the expression of target edge features and improves the detection of small targets.
In the multi-target scenario, the YOLO series methods only detected the largest target. They failed to recognize the other two targets. In contrast, REDETR-RISTD successfully detected all three targets with detection confidences of 0.73, 0.27, and 0.32. This demonstrates that the multi-scale pyramid feature fusion architecture, through a bidirectional fusion mechanism, effectively integrates target features at different scales to achieve comprehensive detection.
In the bright-clutter scenario, REDETR-RISTD reached a detection confidence of 86%. This shows strong anti-interference capability. It maintains high detection accuracy in complex backgrounds and effectively avoids false detections.
In the high-contrast-boundary scenario, YOLOv6s failed to detect the target, and other YOLO methods showed low detection confidence. In contrast, REDETR-RISTD successfully detected the target with a confidence of 72%. This indicates that it has strong target localization ability in high-contrast environments and can overcome the influence of high-contrast noise.
Finally, in the high-noise scenario, the YOLO series methods only detected the largest target. REDETR-RISTD, however, successfully detected three targets with detection confidences of 0.68, 0.69, and 0.71. This further proves its strong robustness in noisy environments and its ability to maintain high detection stability under complex interference conditions.
Figure 10 shows the ground-truth bounding boxes of targets in eight scenarios on the left and the detection results of TOOD, Sparse R-CNN, Mask R-CNN, and REDETR-RISTD on the right.
In the dark-appearance scenario, TOOD failed to detect the target. Sparse R-CNN detected the target with a confidence of 0.799 and produced one false positive with a confidence of 0.511. Mask R-CNN did not detect the target and generated two false positives with confidences of 0.949 and 0.579. In contrast, REDETR-RISTD successfully detected the target with a confidence of 58%. This result demonstrates that REDETR-RISTD, with its reparameterized multi-scale feature extraction module, can better capture target features in low-signal-to-noise and low-contrast environments.
In the blurred-background scenario, all four methods successfully detected the single target, with detection confidences of 41.1%, 86.3%, 99.7%, and 80%, respectively. REDETR-RISTD uses an attention-guided intra-scale contextual feature interaction module to effectively reduce background interference and ensure stable detection performance.
In the complex architectural occlusion scenario, TOOD detected the target with a confidence of 40% and produced three false positives with confidences of 0.367, 0.303, and 0.368. Sparse R-CNN, Mask R-CNN, and REDETR-RISTD detected the target with confidences of 87.5%, 63.7%, and 88%, respectively. The multi-scale feature fusion module of REDETR-RISTD is particularly effective in this scenario. It reduces the negative impact of building occlusion on detection results.
In the cloudy scenario, all methods successfully detected the target with detection confidences of 42.7%, 88.8%, 98.4%, and 85%, respectively. Mask R-CNN produced two false positives with confidences of 0.925 and 0.326. REDETR-RISTD further enhances the expression of target edge features through its multi-scale fusion mechanism, thus improving detection accuracy.
In the multi-target scenario, TOOD detected only the largest target with a confidence of 0.646 and missed two targets. Sparse R-CNN detected three targets with confidences of 0.925, 0.385, and 0.456. Mask R-CNN detected one target with a confidence of 0.962 and missed two targets. In contrast, REDETR-RISTD successfully detected three targets with confidences of 0.73, 0.27, and 0.32. Its multi-scale pyramid feature fusion and bidirectional interaction mechanism effectively integrate target features at different scales, achieving comprehensive target detection.
In the high-contrast-boundary scenario, the detection confidences were 33.7%, 82.2%, 99.9%, and 86% for TOOD, Sparse R-CNN, Mask R-CNN, and REDETR-RISTD, respectively. These results indicate that REDETR-RISTD can maintain high detection accuracy under complex background conditions.
Finally, in the high-noise scenario, TOOD detected one target with a confidence of 0.327 and missed two targets. Sparse R-CNN detected one target with a confidence of 0.899 and missed two targets. Mask R-CNN failed to detect the target. In contrast, REDETR-RISTD successfully detected three targets with confidences of 0.68, 0.69, and 0.71. Its multi-scale fusion and self-attention mechanism effectively reduce the impact of noise on feature extraction, thereby maintaining high detection performance even in high-noise conditions.
Figure 11 shows the visualization results of infrared small target detection in eight typical scenarios. The leftmost part displays the ground-truth bounding boxes of the targets, while the detection outputs of DINO, RT-DETR-ResNet18, RT-DETR-HGNet-L, and REDETR-RISTD are shown from left to right.
In the dark-appearance scenario, DINO correctly detected the target with a confidence of 0.721 but also produced a false positive with a confidence of 0.827. This indicates that it is susceptible to background noise under low-light conditions. RT-DETR-ResNet18 failed to detect the target, while RT-DETR-HGNet-L only detected the target with a confidence of 0.33, showing significantly inadequate performance. In contrast, REDETR-RISTD successfully detected the target with a confidence of 58%, which can be attributed to its reparameterized multi-scale feature extraction module, which is highly sensitive to weak target features in low-signal-to-noise environments.
In the blurred-background scenario, all methods successfully detected the single target, with detection confidences of 88.9% for DINO and 80% for RT-DETR-ResNet18, RT-DETR-HGNet-L, and REDETR-RISTD. This result suggests that when the target is well distinguished from the background, all models can perform well, and the attention-guided contextual feature interaction module of REDETR-RISTD plays a crucial role in maintaining stable detection performance.
In the complex architectural occlusion scenario, DINO detected the target with a confidence of 90.1%, while RT-DETR-ResNet18 and RT-DETR-HGNet-L achieved confidences of 85.1% and 80.7%, respectively. REDETR-RISTD reached 88%. The higher confidence of DINO indicates that it can extract stronger features when dealing with partial occlusion. However, REDETR-RISTD effectively mitigated the interference caused by occlusion through multi-scale feature fusion, ensuring accurate detection.
In the cloudy scenario, DINO detected the target with a confidence of 89.6%, but also produced a false positive with a confidence of 0.745. RT-DETR-ResNet18 and REDETR-RISTD achieved confidences of 83% and 85%, respectively, while RT-DETR-HGNet-L only reached 77%. Although DINO had a high detection confidence, its false positives show that it tends to capture erroneous features under cloud interference. REDETR-RISTD, on the other hand, uses its reparameterized structure to suppress interference information, resulting in more robust detection.
In the multi-target scenario, DINO detected all three targets with confidences of 0.783, 0.673, and 0.656, demonstrating an advantage when there are more targets. RT-DETR-ResNet18 detected only one target and missed two, while RT-DETR-HGNet-L detected two targets but produced one false positive and one missed target. In contrast, REDETR-RISTD successfully detected all three targets with confidences of 0.73, 0.27, and 0.32. This result confirms the effectiveness of REDETR-RISTD’s multi-scale pyramid feature fusion and bidirectional interaction mechanism in integrating features from different scales to handle multi-target detection tasks comprehensively.
In the bright-clutter scenario, all methods exhibited high detection confidences, with DINO reaching 92.6%, RT-DETR-ResNet18 at 87%, RT-DETR-HGNet-L at 85%, and REDETR-RISTD at 86%. This shows that all methods can recognize the target in a bright and cluttered background, with REDETR-RISTD maintaining a relatively stable performance.
For the high-contrast-boundary scenario, DINO achieved a detection confidence of 89%, showing strong target localization capability. In comparison, RT-DETR-ResNet18, RT-DETR-HGNet-L, and REDETR-RISTD achieved confidences of 74%, 69%, and 72%, respectively. Although DINO performed relatively well under these conditions, REDETR-RISTD effectively reduced the negative impact of high-contrast noise through its adaptive noise suppression mechanism, ensuring stable detection results.
In the high-noise scenario, DINO detected two targets with confidences of 0.875 and 0.871, but missed one target. RT-DETR-ResNet18 and RT-DETR-HGNet-L also detected two targets, but both missed one target, with confidences ranging from 0.74 to 0.73 and 0.67 to 0.61, respectively. In contrast, REDETR-RISTD detected three targets with confidences of 0.68, 0.69, and 0.71, demonstrating stronger noise suppression ability. This indicates that in noisy environments, traditional methods are prone to interference in feature extraction, while REDETR-RISTD’s multi-scale feature fusion and self-attention mechanism effectively alleviated this issue, ensuring comprehensive target detection.

4.5. Ablation Experiment

REDETR-RISTD contains three key components: the backbone RepEMSNet, the intra-scale contextual feature interaction AICFI, and the feature fusion framework MSPFN, which compose a model capable of accurately detecting infrared small targets. Ablation experiments were conducted to compare with the baseline model RT-DETR-ResNet18 to validate the effectiveness of the proposed modules, and the ablation experiments included six different combinations.
Group 1: RT-DETR-ResNet18 (Baseline).
Group 2: RT-DETR+RepEMSNet.
Group 3: RT-DETR+RepEMSNet+AICFI.
Group 4: RT-DETR+RepEMSNet+AICFI +MSPFN.
Group 5: REDETR-RISTD (Ours).
The performance of each module is qualitatively analyzed by plotting the PR curves of the five combination methods. As shown in Figure 12, compared with the baseline model, the PR curves of the proposed modules’ combinations Group 2, Group 3, Group 4, and Group 5 have a larger area enclosed with the horizontal and vertical axes than that of Group 1, which fully proves the effectiveness of the three proposed modules.
Table 4 demonstrates the detection results, where Group 1, which is absent from all three modules, represents the baseline model RT-DETR-ResNet18, and Group 1–Group 4 all use GIoU.
As shown in Table 4, the proposed method exhibits an increase in recall, AP50, and AP50-95, and a decrease in the number of parameters compared to the baseline model, achieving an overall improvement in the detector’s performance.
Specifically, compared to the baseline combination Group 1, after replacing the RepEMSNet backbone, the model parameters in Group 2 decreased by 30.3% and 21.97%, respectively, and the detection accuracy metrics, AP50 and AP50-95, increased to 95% and 47.8%, respectively.
In Group 3, after adding the AICFI module, AP50-95 is improved to 49.6%, and the number of model parameters is further reduced to 13.811 M.
After combining the first three modules, Group 4 is obtained. Currently, AP50 reaches 96.4%, recall is improved to 92.6%, and the number of parameters is reduced by 30.3%, respectively.
Finally, the original GIoU is replaced by EnIoU, which accelerates the network’s convergence while increasing AP50-95 to 51.6%. At this time, the number of parameters and computations are further reduced by 30.5% and 19.5%, respectively, compared to the baseline model.
The significant change in AP50-95 is caused by the small number of pixels occupied by the small targets, resulting in a small area covered by the genuine bounding box. Slight deviations occurring in the prediction bounding box can also change significantly with the IoU of the GT box. Using AICFI to interact with the deep feature map output from RepEMSNet with in-scale features makes the network more sensitive to the location information of small targets. The EnIoU is also utilized to focus on the degree and location of overlap between the prediction frame and the GT box to accelerate the network’s convergence and generate a prediction bounding box consistent with the GT box, thus improving AP50-95.
To validate the performance of the proposed backbone model in this thesis for detecting infrared small targets, we conducted a comprehensive comparison with nine current advanced backbone models such as ResNet, Transformer, etc., and the comparison results are shown in Table 5.
In Table 5, combining the four metrics of recall, AP50, AP50-95 and Parameters, only the proposed RepEMSNet backbone performs in the top four, realizing an effective balance between detection accuracy, parameter count, and computational cost. Compared to the latest StarNet backbone, RepEMSNet’s recall, AP50, and AP50-95 are improved by 5.9%, 4.7%, and 2.6%, respectively. The number of parameters is reduced by 1.944. Compared to ResNet-50 and CSwinTransformer, recall, AP50, and AP50-95 are improved by 5.9%, 4.7%, and 2.6%, respectively. AP50 and AP50-95 metrics decreased by about 1%, but Parameters decreased by 28.111 M and 16.64 M.

4.6. Impacts of Reparameterized Structures

The proposed RepEMSNet backbone network employs a structural reparameterization operation, which converts a multi-branch structure into a single-branch structure and integrates the batch normalization operation into the convolution operation. These operations can effectively reduce memory occupancy and improve computational efficiency without affecting detection accuracy.
Table 6 demonstrates the change in REDETR-RISTD inference time before and after structural reparameterization. The batch size and input image resolution are 16 and 640 × 640 pixels, respectively. The results show that the overall inference time of the proposed model is reduced by 21% on CPU and 15.1% on GPU by structural reparameterization.

4.7. Loss Function Comparison Experiment

To verify the effectiveness of the proposed EnIoU loss function, we analyzed it in comparison with existing loss functions (e.g., GIoU [60], DIoU, CIoU [61], EIoU [62], SIoU [63], Inner-IoU [64], and MPDIoU [65]). The auxiliary bounding box was resized and tested using various scale settings. The data in Table 7 show that the model equipped with the EnIO loss function at a scale of ratio = 0.7 exhibits higher detection accuracy. Compared to the baseline model with the GIoU loss function, the recall metric improved by 0.1%, AP50 by 0.1%, and AP50-95 by 3.1%. The results suggest that using the EnIoU loss function may lead to more consistent bounding box regression and improved prediction accuracy.
To clearly illustrate the detection performance of models trained with different IoU loss functions, we use radar charts to display the four metrics P, R, AP50, and AP50-95. To ensure the metrics are comparable, we normalize them. This allows us to present the performance of the models in a unified coordinate system.
From Figure 13, it can be observed that when EnIoU is used as the bounding box loss function with the hyperparameter ratio = 0.7, the radar chart exhibits the largest coverage area. This result indicates superior performance in P, R, AP50, and AP50-95, demonstrating the strongest overall detection capability. The larger polygon area reflects not only higher values in individual metrics but also a balanced improvement in multi-metric performance. The superior performance is attributed to two key factors. First, EnIoU achieves coordinated optimization by addressing sample differentiation, positional accuracy, and multi-scale feature integration through its adaptive weighting mechanism. It effectively balances the trade-off between precision and recall. Second, EnIoU leverages geometric constraints and multi-scale feature fusion to enhance adaptability to complex scenarios, ensuring robust detection performance.
Furthermore, Figure 14 demonstrates the training results of EnIoU under different hyperparameter ratio settings. It is evident that variations in ratio lead to significant fluctuations in model performance across metrics. Among these configurations, the radar chart corresponding to ratio = 0.7 exhibits the largest coverage area. This observation further validates the advantage of this parameter configuration in balancing loss weighting and gradient feedback, ensuring optimal training dynamics.
To analyze the speed of convergence of model training under different IoU, we show the convergence curves of the eight loss functions in Figure 15. All loss functions eventually converge, yet EnIoU achieves a stable state in fewer epochs. The red circle in Figure 15 marks the convergence epoch. This rapid convergence is attributed to the design of EnIoU. It adopts an adaptive loss weighting mechanism that adjusts the size of the auxiliary bounding box based on the sample’s IoU, using a smaller box for high-IoU samples and a larger box for low-IoU samples. In addition, EnIoU improves positional accuracy by calculating the Euclidean distance between the vertices of the predicted and GT boxes, which ensures closer alignment. The design further incorporates multi-scale feature fusion to enhance the model’s adaptability to targets of different sizes. Experimental results confirm that this design not only speeds up convergence but also improves overall detection performance.

4.8. Comparative Experiment of Bounding Box-Based Detection Method and Segmentation-Based Method

In this section, we present a comprehensive comparison between the bounding box-based detection method and segmentation-based approaches for infrared small target detection. We first offer a quantitative evaluation of performance metrics across multiple challenging datasets, and then provide qualitative visualizations that further illustrate detection outcomes under diverse conditions. The following results, including the performance table and detection visualizations, demonstrate the robustness and superiority of our method over traditional segmentation techniques.
From Table 8, it is clear that the infrared small target detection method based on bounding box detection (REDETR-RISTD) significantly outperforms the six segmentation-based methods on multiple metrics. First, REDETR-RISTD achieves the highest performance in three core metrics: precision (P), recall (R), and detection probability (Pd) across the SIRSTv2, IRSTD-1k, and NUDT-SIRST datasets. For example, on the SIRSTv2 dataset, the method achieves a precision of 0.991, a recall of 0.927, and a detection probability of 0.976. The values of other methods are much lower. This consistent advantage shows that the bounding box method is more accurate and robust in locating small targets and capturing their features.
On the other hand, segmentation-based methods depend on pixel-level segmentation. They are easily affected by background noise and blurred target details. As a result, their precision and recall are lower than those of the bounding box method. Although REDETR-RISTD shows a slightly higher false alarm rate (Fa) on some datasets, its false alarm rate is only 1.25 on the NUDT-SIRST dataset. This is a clear advantage compared to other methods. In infrared small target detection, high precision and high recall are critical. A slight increase in the false alarm rate can be remedied by further optimization and threshold adjustment.
Furthermore, REDETR-RISTD demonstrates strong generalization ability across different datasets, which further proves its robustness. Whether it is the complex background of the SIRSTv2 dataset or other scenarios, the method consistently delivers high detection performance. This advantage shows that the bounding box detection method can effectively overcome the shortcomings of traditional segmentation methods, such as loss of detail and blurred target localization, and thus achieves more accurate detection results.
Based on the visual results in Figure 16, we discuss the performance of each method in different scenarios in sequence. First, the ACM method fails to detect the target in low-brightness (dim target) images. In the Noise image with three targets and significant background noise, it only correctly detects two targets, showing false positives and false negatives. This indicates that the ACM method is sensitive to low contrast and noise interference. Next, ALCNet also has false positives and false negatives in low-brightness images. Although it detects all targets in other scenarios, its robustness is limited under challenging conditions. DNANet shows a false positive in low-brightness images and only detects one target correctly. It also misses one target in the multi-target scenario but performs reasonably well in other situations. In contrast, ISTDU-Net detects all targets across all scenarios, demonstrating good adaptability to complex backgrounds and low-contrast conditions. Furthermore, both RDIAN and U-Net fail to detect the target in low-brightness images and miss one target in the multi-target scenario. While their performance is generally good in other conditions, their overall stability is weaker. Finally, our proposed bounding box-based detection method, REDETR-RISTD, accurately detects targets in all scenarios, whether in low brightness, complex backgrounds, or multi-target conditions, with no false positives or false negatives.
Overall, traditional segmentation methods are prone to be affected by background noise and blurred target details in low-brightness and complex scenarios, leading to unstable detection results. In contrast, the REDETR-RISTD bounding box-based method shows higher robustness and accuracy. These results confirm the advantages of using bounding box strategies in infrared small target detection tasks, providing a more reliable solution for practical applications.

5. Conclusions

We propose REDETR-RISTD, an infrared small target detection network with a reparameterizable backbone and multi-scale feature fusion. We introduce a reparameterizable backbone called RepEMSNet that speeds up inference, improves accuracy, and gradually enlarges the sensory field using stacked 3×3 convolutions. In the encoder, we use AICFI to capture local details and global dependencies in the multi-scale feature maps, enabling the network to focus on salient features for infrared small target detection. We also developed a new framework named MSPFN to efficiently fuse multi-scale features using bottom-up fusion and top-down path enhancement strategies. Three modules were designed to extract both local and global contextual information and to significantly enhance feature diversity and expressive capability. Experimental results show that our network outperforms other advanced detectors, and ablation experiments confirm the effectiveness of RepEMSNet, AICFI, and MSPFN. Given its enhanced accuracy and efficiency in detecting small, dim targets against complex backgrounds, REDETR-RISTD demonstrates significant potential for real-world applications, particularly in demanding fields such as infrared search-and-rescue missions and various aerial or ground-based monitoring scenarios where timely and reliable detection is critical.
Infrared small target detection still faces challenges such as background occlusion, interference, and camouflage spoofing, and it is evolving toward infrared tracking missions. In the future, we will further optimize the REDETR-RISTD architecture by exploring improved backbone and feature fusion methodologies to reduce parameters and computation while enhancing accuracy and speed. We will also investigate interpretable learning combined with spatio-temporal information and apply these advances to infrared small target tracking.

Author Contributions

Conceptualization, N.L. and D.W.; methodology, N.L. and D.W.; software, N.L.; validation, N.L. and D.W.; formal analysis, D.W.; investigation, N.L. and D.W.; data curation, D.W.; writing—original draft preparation, N.L.; visualization, N.L.; supervision, D.W.; project administration, D.W. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data underlying the findings of this study are available upon reasonable request from the corresponding author.

Acknowledgments

The authors declare that no additional support, beyond the contributions outlined in the author contribution, was received for this study.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Han, J.; Ma, Y.; Zhou, B.; Fan, F.; Liang, K.; Fang, Y. A robust infrared small target detection algorithm based on human visual system. IEEE Geosci. Remote Sens. Lett. 2014, 11, 2168–2172. [Google Scholar]
  2. Han, J.; Liang, K.; Zhou, B.; Zhu, X.; Zhao, J.; Zhao, L. Infrared small target detection utilizing the multiscale relative local contrast measure. IEEE Geosci. Remote Sens. Lett. 2018, 15, 612–616. [Google Scholar] [CrossRef]
  3. Dai, Y.; Wu, Y.; Zhou, F.; Barnard, K. Asymmetric contextual modulation for infrared small target detection. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Virtual, 5–9 January 2021; pp. 950–959. [Google Scholar]
  4. Wu, Y.; Tan, S.; Zhao, Y.; Liang, L.; Zhou, M.; Ji, G. Broadband multispectral compatible absorbers for radar, infrared and visible stealth application. Prog. Mater. Sci. 2023, 135, 101088. [Google Scholar] [CrossRef]
  5. Bertozzi, M.; Broggi, A.; Fascioli, A.; Graf, T.; Meinecke, M.M. Pedestrian detection for driver assistance using multiresolution infrared vision. IEEE Trans. Veh. Technol. 2004, 53, 1666–1678. [Google Scholar] [CrossRef]
  6. Koukiou, G.; Anastassopoulos, V. Drunk person identification using thermal infrared images. Int. J. Electron. Secur. Digit. Forensics 2012, 4, 229–243. [Google Scholar] [CrossRef]
  7. Koukiou, G. Intoxication identification using thermal imaging. Hum.-Robot Interact.-Theory Appl. 2017, 8, 145–173. [Google Scholar]
  8. Pastor, E.; Planas, E. Infrared imagery on wildfire research. Some examples of sound capabilities and applications. In Proceedings of the 2012 3rd International Conference on Image Processing Theory, Tools and Applications (IPTA), Istanbul, Turkey, 15–18 October 2012; pp. 31–36. [Google Scholar] [CrossRef]
  9. Ring, E.; Ammer, K. Infrared thermal imaging in medicine. Physiol. Meas. 2012, 33, R33. [Google Scholar] [CrossRef]
  10. Koukiou, G. Thermal Biometric Features for Drunk Person Identification Using Multi-Frame Imagery. Electronics 2022, 11, 3924. [Google Scholar] [CrossRef]
  11. Deng, L.; Zhu, H.; Zhou, Q.; Li, Y. Adaptive top-hat filter based on quantum genetic algorithm for infrared small target detection. Multimed. Tools Appl. 2018, 77, 10539–10551. [Google Scholar] [CrossRef]
  12. Cui, Y.; Lei, T.; Chen, G.; Zhang, Y.; Peng, L.; Hao, X.; Zhang, G. Hollow side window filter with saliency prior for infrared small target detection. IEEE Geosci. Remote Sens. Lett. 2023, 21, 6001505. [Google Scholar] [CrossRef]
  13. Zhang, X.; Ru, J.; Wu, C. Infrared small target detection based on gradient correlation filtering and contrast measurement. IEEE Trans. Geosci. Remote Sens. 2023, 61, 5603012. [Google Scholar] [CrossRef]
  14. Wang, Z.; Duan, S.; Sun, C. Infrared small target detection method combined with bilateral filter and local entropy. Secur. Commun. Netw. 2021, 2021, 6661852. [Google Scholar] [CrossRef]
  15. Li, Y.; Li, Z.; Shen, Y.; Li, J. Infrared small target detection based on 1-D difference of guided filtering. IEEE Geosci. Remote Sens. Lett. 2023, 20, 1–5. [Google Scholar] [CrossRef]
  16. Zhang, X.; Ru, J.; Wu, C. A nonparametric regression-based multi-scale gradient correlation filtering method for infrared small target detection. Electronics 2023, 12, 1562. [Google Scholar] [CrossRef]
  17. Yu, C.; Liu, Y.; Wu, S.; Xia, X.; Hu, Z.; Lan, D.; Liu, X. Pay attention to local contrast learning networks for infrared small target detection. IEEE Geosci. Remote Sens. Lett. 2022, 19, 3512705. [Google Scholar] [CrossRef]
  18. Tong, X.; Sun, B.; Wei, J.; Zuo, Z.; Su, S. EAAU-Net: Enhanced asymmetric attention U-Net for infrared small target detection. Remote Sens. 2021, 13, 3200. [Google Scholar] [CrossRef]
  19. Li, B.; Xiao, C.; Wang, L.; Wang, Y.; Lin, Z.; Li, M.; An, W.; Guo, Y. Dense nested attention network for infrared small target detection. IEEE Trans. Image Process. 2022, 32, 1745–1758. [Google Scholar] [CrossRef]
  20. He, X.; Ling, Q.; Zhang, Y.; Lin, Z.; Zhou, S. Detecting dim small target in infrared images via subpixel sampling cuneate network. IEEE Geosci. Remote Sens. Lett. 2022, 19, 6513005. [Google Scholar] [CrossRef]
  21. Zhao, B.; Wang, C.; Fu, Q.; Han, Z. A novel pattern for infrared small target detection with generative adversarial network. IEEE Trans. Geosci. Remote Sens. 2020, 59, 4481–4492. [Google Scholar] [CrossRef]
  22. Wang, A.; Li, W.; Wu, X.; Huang, Z.; Tao, R. Mpanet: Multi-patch attention for infrared small target object detection. In Proceedings of the IGARSS 2022—2022 IEEE International Geoscience and Remote Sensing Symposium, Kuala Lumpur, Malaysia, 17–22 July 2022; pp. 3095–3098. [Google Scholar]
  23. Zhou, H.; Tian, C.; Zhang, Z.; Li, C.; Xie, Y.; Li, Z. PixelGame: Infrared small target segmentation as a Nash equilibrium. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2022, 15, 8010–8024. [Google Scholar] [CrossRef]
  24. Zhao, M.; Cheng, L.; Yang, X.; Feng, P.; Liu, L.; Wu, N. TBC-Net: A real-time detector for infrared small target detection using semantic constraint. arXiv 2019, arXiv:2001.05852. [Google Scholar]
  25. Hu, K.; Sun, W.; Nie, Z.; Cheng, R.; Chen, S.; Kang, Y. Real-time infrared small target detection network and accelerator design. Integration 2022, 87, 241–252. [Google Scholar] [CrossRef]
  26. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An image is worth 16×16 words: Transformers for image recognition at scale. arXiv 2020, arXiv:2010.11929. [Google Scholar]
  27. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, L.; Polosukhin, I. Attention is all you need. In Proceedings of the 31st International Conference on Neural Information Processing Systems, NIPS’17, Long Beach, CA, USA, 4–9 December 2017; pp. 6000–6010. [Google Scholar]
  28. Beal, J.; Kim, E.; Tzeng, E.; Park, D.H.; Zhai, A.; Kislyuk, D. Toward transformer-based object detection. arXiv 2020, arXiv:2012.09958. [Google Scholar]
  29. Fang, Y.; Liao, B.; Wang, X.; Fang, J.; Qi, J.; Wu, R.; Niu, J.; Liu, W. You only look at one sequence: Rethinking transformer in vision through object detection. Adv. Neural Inf. Process. Syst. 2021, 34, 26183–26197. [Google Scholar]
  30. Li, Y.; Xie, S.; Chen, X.; Dollar, P.; He, K.; Girshick, R. Benchmarking detection transfer learning with vision transformers. arXiv 2021, arXiv:2111.11429. [Google Scholar]
  31. Fang, Y.; Yang, S.; Wang, S.; Ge, Y.; Shan, Y.; Wang, X. Unleashing vanilla vision transformer with masked image modeling for object detection. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Paris, France, 2–3 October 2023; pp. 6244–6253. [Google Scholar]
  32. Lou, L.; Lu, K.; Xue, J. Multi-Scale Vision Transformer for Defect Object Detection. Procedia Comput. Sci. 2023, 222, 397–406. [Google Scholar] [CrossRef]
  33. Wang, L.; Tien, A. Aerial image object detection with vision transformer detector (ViTDet). In Proceedings of the IGARSS 2023—2023 IEEE International Geoscience and Remote Sensing Symposium, Pasadena, CA, USA, 16–21 July 2023; pp. 6450–6453. [Google Scholar]
  34. Singh, A. Training Strategies for Vision Transformers for Object Detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023; pp. 110–118. [Google Scholar]
  35. Zhang, C.; Su, J.; Ju, Y.; Lam, K.M.; Wang, Q. Efficient inductive vision transformer for oriented object detection in remote sensing imagery. IEEE Trans. Geosci. Remote Sens. 2023, 61, 5616320. [Google Scholar] [CrossRef]
  36. Li, J.; Tian, P.; Song, R.; Xu, H.; Li, Y.; Du, Q. PCViT: A Pyramid Convolutional Vision Transformer Detector for Object Detection in Remote Sensing Imagery. IEEE Trans. Geosci. Remote Sens. 2024, 62, 5608115. [Google Scholar] [CrossRef]
  37. Gong, W. Lightweight Object Detection: A Study Based on YOLOv7 Integrated with ShuffleNetv2 and Vision Transformer. arXiv 2024, arXiv:2403.01736. [Google Scholar]
  38. Guo, Z.; Gao, J.; Ma, G.; Xu, J. Spatio-Temporal Aggregation Transformer for Object Detection with Neuromorphic Vision Sensors. IEEE Sens. J. 2024, 24, 19397–19406. [Google Scholar] [CrossRef]
  39. Carion, N.; Massa, F.; Synnaeve, G.; Usunier, N.; Kirillov, A.; Zagoruyko, S. End-to-end object detection with transformers. In Proceedings of the European Conference on Computer Vision, Glasgow, UK, 23–28 August 2020; pp. 213–229. [Google Scholar]
  40. Zhu, X.; Su, W.; Lu, L.; Li, B.; Wang, X.; Dai, J. Deformable detr: Deformable transformers for end-to-end object detection. arXiv 2020, arXiv:2010.04159. [Google Scholar]
  41. Liu, S.; Li, F.; Zhang, H.; Yang, X.; Qi, X.; Su, H.; Zhu, J.; Zhang, L. Dab-detr: Dynamic anchor boxes are better queries for detr. arXiv 2022, arXiv:2201.12329. [Google Scholar]
  42. Li, F.; Zhang, H.; Liu, S.; Guo, J.; Ni, L.M.; Zhang, L. Dn-detr: Accelerate detr training by introducing query denoising. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 13619–13627. [Google Scholar]
  43. Chen, Q.; Chen, X.; Zeng, G.; Wang, J. Group detr: Fast training convergence with decoupled one-to-many label assignment. arXiv 2022, arXiv:2207.13085. [Google Scholar]
  44. Meng, D.; Chen, X.; Fan, Z.; Zeng, G.; Li, H.; Yuan, Y.; Sun, L.; Wang, J. Conditional detr for fast training convergence. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 3651–3660. [Google Scholar]
  45. Roh, B.; Shin, J.; Shin, W.; Kim, S. Sparse detr: Efficient end-to-end object detection with learnable sparsity. arXiv 2021, arXiv:2111.14330. [Google Scholar]
  46. Yao, Z.; Ai, J.; Li, B.; Zhang, C. Efficient detr: Improving end-to-end object detector with dense prior. arXiv 2021, arXiv:2104.01318. [Google Scholar]
  47. Li, F.; Zeng, A.; Liu, S.; Zhang, H.; Li, H.; Zhang, L.; Ni, L.M. Lite detr: An interleaved multi-scale encoder for efficient detr. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023; pp. 18558–18567. [Google Scholar]
  48. Chen, X.; Wei, F.; Zeng, G.; Wang, J. Conditional detr v2: Efficient detection transformer with box queries. arXiv 2022, arXiv:2207.08914. [Google Scholar]
  49. Wang, Y.; Zhang, X.; Yang, T.; Sun, J. Anchor DETR: Query design for transformer-based object detection. arXiv 2021, arXiv:2109.07107. [Google Scholar]
  50. Zhao, Y.; Lv, W.; Xu, S.; Wei, J.; Wang, G.; Dang, Q.; Liu, Y.; Chen, J. Detrs beat yolos on real-time object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 16–22 June 2024; pp. 16965–16974. [Google Scholar]
  51. Zhang, M.; Zhang, R.; Yang, Y.; Bai, H.; Zhang, J.; Guo, J. ISNet: Shape Matters for Infrared Small Target Detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18–24 June 2022; pp. 877–886. [Google Scholar]
  52. Lv, H.; Jing, D.; Zhan, W. Research on Infrared Image Target Detection Technology Based on YOLOv3 and Computer Vision. J. Phys. Conf. Ser. 2021, 2033, 012142. [Google Scholar] [CrossRef]
  53. Mao, Q.; Li, Q.; Wang, B.; Zhang, Y.; Dai, T.; Philip Chen, C.L. SpirDet: Toward Efficient, Accurate, and Lightweight Infrared Small-Target Detector. IEEE Trans. Geosci. Remote Sens. 2024, 62, 1–12. [Google Scholar] [CrossRef]
  54. Dai, Y.; Li, X.; Zhou, F.; Qian, Y.; Chen, Y.; Yang, J. One-Stage Cascade Refinement Networks for Infrared Small Target Detection. IEEE Trans. Geosci. Remote Sens. 2023, 61, 1–17. [Google Scholar] [CrossRef]
  55. Mou, X.; Lei, S.; Zhou, X. YOLO-FR: A YOLOv5 Infrared Small Target Detection Algorithm Based on Feature Reassembly Sampling Method. Sensors 2023, 23, 2710. [Google Scholar] [CrossRef]
  56. Zhu, J.; Qin, C.; Choi, D. YOLO-SDLUWD: YOLOv7-based small target detection network for infrared images in complex backgrounds. Digit. Commun. Netw. 2023. [Google Scholar] [CrossRef]
  57. Zhao, L.; Wang, J.; Chen, Y.; Yin, Q.; Rong, G.; Zhou, S.; Tang, J. IST-DETR: Improved DETR for Infrared Small Target Detection. IEEE Access 2024, 12, 164303–164314. [Google Scholar] [CrossRef]
  58. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  59. Pan, Z.; Cai, J.; Zhuang, B. Fast vision transformers with hilo attention. Adv. Neural Inf. Process. Syst. 2022, 35, 14541–14554. [Google Scholar]
  60. Rezatofighi, H.; Tsoi, N.; Gwak, J.; Sadeghian, A.; Reid, I.; Savarese, S. Generalized intersection over union: A metric and a loss for bounding box regression. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 658–666. [Google Scholar]
  61. Zheng, Z.; Wang, P.; Liu, W.; Li, J.; Ye, R.; Ren, D. Distance-IoU loss: Faster and better learning for bounding box regression. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; Volume 34, pp. 12993–13000. [Google Scholar]
  62. Zhang, Y.F.; Ren, W.; Zhang, Z.; Jia, Z.; Wang, L.; Tan, T. Focal and efficient IoU loss for accurate bounding box regression. Neurocomputing 2022, 506, 146–157. [Google Scholar] [CrossRef]
  63. Gevorgyan, Z. SIoU loss: More powerful learning for bounding box regression. arXiv 2022, arXiv:2205.12740. [Google Scholar]
  64. Zhang, H.; Xu, C.; Zhang, S. Inner-IoU: More effective intersection over union loss with auxiliary bounding box. arXiv 2023, arXiv:2311.02877. [Google Scholar]
  65. Ma, S.; Xu, Y. MPDIoU: A loss for efficient and accurate bounding box regression. arXiv 2023, arXiv:2307.07662. [Google Scholar]
Figure 1. Typical infrared small targets. The red circles indicate the locations of the small targets.
Figure 1. Typical infrared small targets. The red circles indicate the locations of the small targets.
Sensors 25 02771 g001
Figure 2. Overall structure of REDETR-RISTD. The decoder and auxiliary prediction heads iteratively refine object queries to generate category and bounding box predictions for three distinct objects, which are displayed on the right side as bounding boxes in three different colors.
Figure 2. Overall structure of REDETR-RISTD. The decoder and auxiliary prediction heads iteratively refine object queries to generate category and bounding box predictions for three distinct objects, which are displayed on the right side as bounding boxes in three different colors.
Sensors 25 02771 g002
Figure 3. Structure of RepEMSNet. The overall architecture of the network is illustrated, with the right side detailing the structural composition of Conv1 through Conv5 layers and the RepConv module.
Figure 3. Structure of RepEMSNet. The overall architecture of the network is illustrated, with the right side detailing the structural composition of Conv1 through Conv5 layers and the RepConv module.
Sensors 25 02771 g003
Figure 4. Framework of AICFI. The main architecture of AICFI is presented in the upper portion of the figure, while the lower portion illustrates the internal structure of the HiLo attention mechanism.
Figure 4. Framework of AICFI. The main architecture of AICFI is presented in the upper portion of the figure, while the lower portion illustrates the internal structure of the HiLo attention mechanism.
Sensors 25 02771 g004
Figure 5. Overall structure of MSPFN. The figure illustrates the main components (a) MSFEF, (b) DPCConv, (c) MSF-FUM, and (d) MSF-FDM, along with the internal structures of the ConvTranspose, Conv1, and Conv3 modules shown on the right.
Figure 5. Overall structure of MSPFN. The figure illustrates the main components (a) MSFEF, (b) DPCConv, (c) MSF-FUM, and (d) MSF-FDM, along with the internal structures of the ConvTranspose, Conv1, and Conv3 modules shown on the right.
Sensors 25 02771 g005
Figure 6. Diagram illustrating the loss function. The red box represents the ground-truth box, while the green box represents the predicted bounding box.
Figure 6. Diagram illustrating the loss function. The red box represents the ground-truth box, while the green box represents the predicted bounding box.
Sensors 25 02771 g006
Figure 7. PR curves comparison between 11 detectors and the REDETR-RISTD model on three infrared small target detection datasets (SIRSTv2, IRSTD-1k, and NUDT-SIRST).
Figure 7. PR curves comparison between 11 detectors and the REDETR-RISTD model on three infrared small target detection datasets (SIRSTv2, IRSTD-1k, and NUDT-SIRST).
Sensors 25 02771 g007
Figure 8. ROC curves comparison between 11 detectors and the REDETR-RISTD model on three infrared small target detection datasets (SIRSTv2, IRSTD-1k, and NUDT-SIRST).
Figure 8. ROC curves comparison between 11 detectors and the REDETR-RISTD model on three infrared small target detection datasets (SIRSTv2, IRSTD-1k, and NUDT-SIRST).
Sensors 25 02771 g008
Figure 9. Comparison results of the detection performance of the YOLO series methods. On the far left, the ground truth provides locally zoomed-in annotations of infrared small targets. The first to the last rows show the results for detecting infrared small targets with dark appearances, blurred backgrounds, complex architectural masks, cloudy scenes, multiple targets, bright clutter, high-contrast boundaries, and high noise levels, respectively. Green bounding boxes denote true positive (TP) detection results, with numerical annotations indicating confidence scores; orange bounding boxes characterize anomalous detection events: those containing numerical values represent false positive (FP) predictions, where values reflect the confidence level of false detection, while unnumbered orange boxes indicate missing detection instances.
Figure 9. Comparison results of the detection performance of the YOLO series methods. On the far left, the ground truth provides locally zoomed-in annotations of infrared small targets. The first to the last rows show the results for detecting infrared small targets with dark appearances, blurred backgrounds, complex architectural masks, cloudy scenes, multiple targets, bright clutter, high-contrast boundaries, and high noise levels, respectively. Green bounding boxes denote true positive (TP) detection results, with numerical annotations indicating confidence scores; orange bounding boxes characterize anomalous detection events: those containing numerical values represent false positive (FP) predictions, where values reflect the confidence level of false detection, while unnumbered orange boxes indicate missing detection instances.
Sensors 25 02771 g009
Figure 10. Comparison results of the detection performance of the typical one-stage and two-stage methods. On the far left, the ground truth provides locally zoomed-in annotations of infrared small targets. The first to the last rows show the results for detecting infrared small targets with dark appearances, blurred backgrounds, complex architectural masks, cloudy scenes, multiple targets, bright clutter, high-contrast boundaries, and high noise levels, respectively. Green bounding boxes denote true positive (TP) detection results, with numerical annotations indicating confidence scores; orange bounding boxes characterize anomalous detection events: those containing numerical values represent false positive (FP) predictions, where values reflect the confidence level of false detection, while unnumbered orange boxes indicate missing detection instances.
Figure 10. Comparison results of the detection performance of the typical one-stage and two-stage methods. On the far left, the ground truth provides locally zoomed-in annotations of infrared small targets. The first to the last rows show the results for detecting infrared small targets with dark appearances, blurred backgrounds, complex architectural masks, cloudy scenes, multiple targets, bright clutter, high-contrast boundaries, and high noise levels, respectively. Green bounding boxes denote true positive (TP) detection results, with numerical annotations indicating confidence scores; orange bounding boxes characterize anomalous detection events: those containing numerical values represent false positive (FP) predictions, where values reflect the confidence level of false detection, while unnumbered orange boxes indicate missing detection instances.
Sensors 25 02771 g010
Figure 11. Comparison results of the detection performance of the transformer-based methods. On the far left, the ground truth provides locally zoomed-in annotations of infrared small targets. The first to the last rows show the results for detecting infrared small targets with dark appearances, blurred backgrounds, complex architectural masks, cloudy scenes, multiple targets, bright clutter, high-contrast boundaries, and high noise levels, respectively. Green bounding boxes denote true positive (TP) detection results, with numerical annotations indicating confidence scores; orange bounding boxes characterize anomalous detection events: those containing numerical values represent false positive (FP) predictions, where values reflect the confidence level of false detection, while unnumbered orange boxes indicate missing detection instances.
Figure 11. Comparison results of the detection performance of the transformer-based methods. On the far left, the ground truth provides locally zoomed-in annotations of infrared small targets. The first to the last rows show the results for detecting infrared small targets with dark appearances, blurred backgrounds, complex architectural masks, cloudy scenes, multiple targets, bright clutter, high-contrast boundaries, and high noise levels, respectively. Green bounding boxes denote true positive (TP) detection results, with numerical annotations indicating confidence scores; orange bounding boxes characterize anomalous detection events: those containing numerical values represent false positive (FP) predictions, where values reflect the confidence level of false detection, while unnumbered orange boxes indicate missing detection instances.
Sensors 25 02771 g011
Figure 12. PR curve of the different groups on three infrared small target detection datasets (SIRSTv2, IRSTD-1k, and NUDT-SIRST).
Figure 12. PR curve of the different groups on three infrared small target detection datasets (SIRSTv2, IRSTD-1k, and NUDT-SIRST).
Sensors 25 02771 g012
Figure 13. Performance comparison of eight IoU loss functions under the metrics P, R, AP50, and AP50-95.
Figure 13. Performance comparison of eight IoU loss functions under the metrics P, R, AP50, and AP50-95.
Sensors 25 02771 g013
Figure 14. Performance comparison of the EnIoU loss function with eight different hyperparameter ratio settings under the metrics P, R, AP50, and AP50-95.
Figure 14. Performance comparison of the EnIoU loss function with eight different hyperparameter ratio settings under the metrics P, R, AP50, and AP50-95.
Sensors 25 02771 g014
Figure 15. Convergence curves of seven typical IoU and EnIoU loss functions. The red circle indicates the convergence point of the proposed EnIoU training.
Figure 15. Convergence curves of seven typical IoU and EnIoU loss functions. The red circle indicates the convergence point of the proposed EnIoU training.
Sensors 25 02771 g015
Figure 16. The detection results of segmentation-based methods and REDETR-RISTD across eight different scenarios are presented. Dashed circles indicate missed detections, while solid ellipses mark false detections. The green box indicates that the target has been detected correctly.
Figure 16. The detection results of segmentation-based methods and REDETR-RISTD across eight different scenarios are presented. Dashed circles indicate missed detections, while solid ellipses mark false detections. The green box indicates that the target has been detected correctly.
Sensors 25 02771 g016
Table 1. Configurations of RepEMSNet. The original input infrared image resolution is 640 × 640 pixels.
Table 1. Configurations of RepEMSNet. The original input infrared image resolution is 640 × 640 pixels.
ModuleNumberScaling FactorOutput Resolution (Pixels)Output ChannelOutput Feature Map
Conv11-320 × 32064S1
Conv11-160 × 160128-
RMSFE1 s = 0.5 160 × 160128S2
Conv11-80 × 80256-
RMSFE1 s = 0.5 80 × 80256S3
Conv11-40 × 40384-
RMSFE1 s = 1 40 × 40384S4
Conv11-20 × 20384-
RMSFE3 s = 1 20 × 20384S5
Table 2. Experimental hardware parameters.
Table 2. Experimental hardware parameters.
NameConfiguration
Operating systemWin11
Computing platformCUDA 11.7
CPUIntel Core i5
GPUNVIDIA GeForce RTX 3060 (GALAX Technology, Hong Kong, China)
GPU memory size12 G
Table 3. Quantitative analysis results. The best-performing method is presented in bold, and the second-best method is shown in bold italics.
Table 3. Quantitative analysis results. The best-performing method is presented in bold, and the second-best method is shown in bold italics.
MethodsDataset
SIRSTv2IRSTD-1kNUDT-SIRSTParameters/M
PRAP50AP50-95PRAP50AP50-95PRAP50AP50-95
YOLOv6s0.9060.7760.8640.4460.8310.7210.790.3590.9040.9410.9650.75416.298
YOLOv70.890.8470.8760.4070.8270.7180.7590.3290.9590.9150.940.6236.195
YOLOv8m0.9260.7580.8750.4470.8830.7210.8090.3730.9770.9040.9680.76625.84
TOOD0.6890.6610.7040.2680.8390.7450.8090.3630.9520.9250.9580.69632.018
Sparse R-CNN0.8970.8630.8880.4770.8260.7430.810.3840.9860.910.9440.64977.8
Mask R-CNN0.9230.790.8880.5110.8070.5610.6910.3570.8110.8140.8770.56343.991
DINO0.9270.9230.9480.4850.8360.8160.8250.3480.9830.9640.9780.77147.54
YOLOv9m0.9180.8120.9170.4720.870.7350.8080.370.9220.8910.9630.76920.156
YOLOv10b0.9540.7740.8860.4870.7760.7280.8070.3810.970.9150.9780.66819.005
RT-DETR-ResNet180.9580.9110.940.4680.8410.7990.8270.370.9890.9680.990.72419.873
RT-DETR-HGNetv2-L0.9740.9180.9430.4770.8570.7840.8410.3810.9950.9860.9950.80931.986
REDETR-RISTD (Ours)0.9910.9270.9650.5160.8780.8330.8430.3880.9920.9790.9850.78613.814
Table 4. Detection results for different modules. ✓ and × indicate the presence or absence of the corresponding module, respectively.
Table 4. Detection results for different modules. ✓ and × indicate the presence or absence of the corresponding module, respectively.
GroupRepEMSNetAICFIMSPFNEnIoURAP50AP50-95Parameters/M
1××××0.9110.940.46819.873
2×××0.9290.950.47813.845
3××0.9180.9470.49613.811
4×0.9260.9640.48513.851
50.9270.9650.51613.814
Table 5. Experimental results of different backbones. The best and the fourth-best values are denoted in bold, respectively.
Table 5. Experimental results of different backbones. The best and the fourth-best values are denoted in bold, respectively.
BackbonesRAP50AP50-95Parameters/M
ResNet-180.9110.940.46819.873
ResNet-340.8870.9470.48031.106
ResNet-500.9350.9510.49341.956
HGNet-V20.9180.9430.47731.986
StarNet0.8700.9030.45215.789
EfficientViT0.9030.9220.44210.703
RepViT0.8870.9070.4244.471
CSwinTransformer0.9320.9480.48330.485
SwinTransformer0.8760.9130.44036.313
RepEMSNet0.9290.9500.47813.845
Table 6. Inference time on GPUs and CPUs. ✓ means adopting reparameterization, and × means not adopting reparameterization.
Table 6. Inference time on GPUs and CPUs. ✓ means adopting reparameterization, and × means not adopting reparameterization.
ReparameterDevicePreprocessInferencePostprocess
×CPU1.0 ms390.4 ms0.1 ms
CPU1.2 ms307.8 ms0.1 ms
×RTX30600.7 ms13.1 ms0.1 ms
RTX30600.8 ms10.8 ms0.2 ms
Table 7. Results of the effect of different loss functions. The best values are denoted in bold.
Table 7. Results of the effect of different loss functions. The best values are denoted in bold.
Loss FunctionPRAP50AP50-95
GIoU0.9830.9260.9640.485
DIoU0.9570.9030.9450.479
CIoU0.9530.9110.9540.512
EIoU0.9250.8920.9270.452
SIoU0.920.930.9410.481
Inner-IoU0.9560.880.9480.49
MPDIoU0.9490.90.9250.487
EnIoU (ratio = 0.65)0.9720.9030.9440.474
EnIoU (ratio = 0.7)0.9910.9270.9650.516
EnIoU (ratio = 0.75)0.9680.9190.9410.467
EnIoU (ratio = 0.8)0.9020.8890.9020.457
EnIoU (ratio = 1.1)0.920.8350.8960.433
EnIoU (ratio = 1.15)0.9470.8950.9180.451
EnIoU (ratio = 1.2)0.9660.9140.9440.486
EnIoU (ratio = 1.25)0.9060.8630.9260.44
Table 8. Our model vs. SOTA models: comparison of P, R, Pd, and Fa ( × 10 5 ) values on the SIRSTv2, IRSTD-1k, and NUDT-SIRST datasets. The best performers are in bold.
Table 8. Our model vs. SOTA models: comparison of P, R, Pd, and Fa ( × 10 5 ) values on the SIRSTv2, IRSTD-1k, and NUDT-SIRST datasets. The best performers are in bold.
MethodsDataset
SIRSTv2IRSTD-1kNUDT-SIRST
PRPdFaPRPdFaPRPdFa
ACM0.7210.7770.8795.850.6790.7570.9235.930.7060.8690.9712.2
ALCNet0.8380.6650.9021.840.70.820.91360.8090.7970.9631.29
DNANet0.8760.8630.971.80.820.7260.8791.310.9540.9590.991.22
ISTDU-Net0.8520.7960.9622.150.780.770.8993.510.9470.9410.9871.28
RDIAN0.8990.720.9396.260.8280.670.8660.670.9170.8820.9781.29
OSCAR0.8730.7420.9094.20.7690.760.8962.830.90.9270.9741.35
REDETR-RISTD (Ours)0.9910.9270.9766.90.9220.8330.9224.590.9920.9790.9921.25
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Li, N.; Wei, D. REDETR-RISTD: Real-Time Long-Range Infrared Small Target Detection Network Based on the Reparameterized Efficient Detection Transformer. Sensors 2025, 25, 2771. https://doi.org/10.3390/s25092771

AMA Style

Li N, Wei D. REDETR-RISTD: Real-Time Long-Range Infrared Small Target Detection Network Based on the Reparameterized Efficient Detection Transformer. Sensors. 2025; 25(9):2771. https://doi.org/10.3390/s25092771

Chicago/Turabian Style

Li, Ning, and Daozhi Wei. 2025. "REDETR-RISTD: Real-Time Long-Range Infrared Small Target Detection Network Based on the Reparameterized Efficient Detection Transformer" Sensors 25, no. 9: 2771. https://doi.org/10.3390/s25092771

APA Style

Li, N., & Wei, D. (2025). REDETR-RISTD: Real-Time Long-Range Infrared Small Target Detection Network Based on the Reparameterized Efficient Detection Transformer. Sensors, 25(9), 2771. https://doi.org/10.3390/s25092771

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop