Next Article in Journal
Investigating LiDAR Metrics for Old-Growth Beech- and Spruce-Dominated Forest Identification in Central Europe
Previous Article in Journal
Accelerating CO2 Outgassing in the Equatorial Pacific from Satellite Remote Sensing
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

STIDNet: Spatiotemporally Integrated Detection Network for Infrared Dim and Small Targets

1
Research Center for Space Optical Engineering, Harbin Institute of Technology, Harbin 150001, China
2
Shanghai Institute of Satellite Engineering, Shanghai 201109, China
3
Zhengzhou Research Institute, Harbin Institute of Technology, Zhengzhou 450000, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2025, 17(2), 250; https://doi.org/10.3390/rs17020250
Submission received: 2 December 2024 / Revised: 5 January 2025 / Accepted: 10 January 2025 / Published: 12 January 2025
(This article belongs to the Section Remote Sensing Image Processing)

Abstract

:
Infrared dim and small target detection (IRDSTD) aims to obtain target position information from the background, clutter, and noise. However, for infrared dim and small targets with low signal-to-clutter ratios (SCRs), the detection difficulty lies in the fact that their poor local spatial saliency will lead to missed detections and false alarms. In this work, a spatiotemporally integrated detection network (STIDNet) is proposed for IRDSTD. In the network, a spatial saliency feature generation module (SSFGM) employs a U-shaped network to extract deep features from the spatial dimension of the input image in a frame-by-frame manner and splices them based on the temporal dimension to obtain an airtime feature tensor. IRDSTs with direction-of-motion consistency and strong interframe correlation are reinforced, and randomly generated spurious waves, noise, and other false alarms are inhibited via a fixed-weight multiscale motion feature-based 3D convolution kernel (FWMFCK-3D). A mapping from the features to the target probability likelihood map is constructed in a spatiotemporal feature fusion module (STFFM) by performing 3D convolutional fusion on the spatially localized saliency and time-domain motion features. Finally, several ablation and comparison experiments indicate the excellent performance of the proposed network. For infrared dim and small targets with SCRs < 3, the average AUC value still reached 0.99786.

1. Introduction

Infrared imaging methods with good concealment, strong anti-interference capabilities, long imaging distances, and low costs have been extensively explored in relation to infrared search and tracking (IRST), civil maritime surveillance, air surveillance, and vehicle detection fields [1,2]. Among them, infrared dim and small target detection (IRDSTD), which accurately locates targets of interest in infrared images, has received much attention as a key technology. Nevertheless, their small sizes, limited shape features, indefinite textures, and low signal-to-clutter ratios (SCRs) make infrared dim and small targets (IRDSTs) prone to being submerged in background clutter and noise, which makes detection difficult.

1.1. Related Works

1.1.1. Single-Frame Infrared Dim and Small Target Detection

Single-frame infrared small target detection methods can be divided into traditional methods and deep learning-based methods.
Filter-based methods, which are traditional single-frame methods for IRDSTD, aim to predict the background of an image as accurately as possible from local spatial information. In these methods, initially, filtering is performed on a local region centered on each pixel, the original local region is replaced with the computed result, and each pixel in the image is traversed to obtain a background prediction image. Afterward, saliency images are obtained by calculating the differences between the original image and the background prediction image. Finally, the detection result is obtained via threshold segmentation. Filter-based methods typically include median filter [3], two-dimensional least-mean-square filter (LMS) [4], bilateral filter (BF) [5], mathematical morphology [6], and Robinson guard filter [7] methods. These methods are fast and effective but exhibit a limited ability to detect IRDSTs with low SCRs.
The design inspiration of the IRDSTD algorithm is based on the local contrast characteristics of the human visual system; small high-intensity targets in complex infrared images usually attract the human eye very quickly, so the background region can be suppressed by the high contrast level of the target. Chen et al. [8] reported that IRDSTs could be detected via local contrast methods (LCMs), in which a contrast map is obtained by traversing the input image to compute the differences between each pixel and its neighboring pixels; then, those pixels in the contrast map whose differences are larger than the threshold are marked as targets. Han et al. [9] proposed an improved local contrast measurement (ILCM) approach to solve the problem that LCM algorithms are prone to pixel-sized noise with high brightness (PNHB) and low efficiency. PNHB is suppressed by smoothing the block mean during preprocessing, and the efficiency of the algorithm is improved by increasing the slide step size and performing block processing. A multiscale patch-based contrast measurement (MPCM) [10] scheme based on image blocks was reported and could effectively enhance real targets by benefiting from patch-based contrast measurement (PCM) and a multiscale strategy adapted to the target scale. Moreover, LCMs, such as relative local contrast measurement (RLCM) [11], multiscale improved LCM (MLCM) [12], and tri-layer window-based local contrast measurement (TLLCM) [13], have been widely explored. Unfortunately, these methods have poor IRDST detection capabilities because they primarily exploit the airspace saliency characteristics of the target.
After fully considering the distributional characteristics of the IRDST of interest and the background, based on the nonlocal autocorrelation of the infrared background and the sparsity of the target, low-rank decomposition methods can separate the target from the background, and related strategies have been widely researched. The infrared patch image (IPI) technique [14] separates the IRDST from the background through sparse representation and low-rank matrix restoration, which can transform the detection problem into a convex optimization problem with a low rank and sparse matrix separation. A non-negative infrared image block (NIPPS) [15] and a weighted infrared image block tensor (RIPT) [16] were proposed to improve the accuracy of IRDSTD by improving the accuracy of the low-rank matrix recovery process to reduce the error resulting from IRDST shrinkage and noise residuals.
Unlike traditional algorithms based on artificial features, deep learning methods use large amounts of data to autonomously learn the characteristics of the IRDST. They are beneficial for mining the unique and potential information contained in image targets, thus achieving satisfactory IRDST detection effects.
Compared with RISTDNet [17], EAAU-Net [18], and DNANet [19], ALCNet [20] highlights infrared small target features by embedding the traditional local contrast measurement method into the end-to-end network and using a bottom-up attention mechanism for information fusion. An attention fusion-based feature pyramid network for detecting IRDSTs was designed by combining a feature extraction module with a feature fusion module [21]. A coarse-to-fine internal attention perception network was proposed for infrared small target detection by assuming that the pixels of the target or background are related [22]. UIU-Net [23] was reported to achieve multilevel and multiscale learning by embedding a tiny U-Net model into a larger U-Net backbone. In IRDSTD tasks, compared with traditional algorithms, deep learning methods have higher detection rates and lower false alarm rates.
However, under severe clutter and complex background interference, the performance of single-frame detection methods degrades when the target is poorly salient in a single image. It is important to utilize multiframe images to fuse spatiotemporal information when conducting IRDSTD in scenarios with low SCRs.

1.1.2. Multiframe Infrared Dim and Small Target Detection

Multiframe detection methods aim to achieve IRDSTD and reduce false alarms by integrating the spatial local salient information of targets with time-domain motion and change information. Traditional multiframe methods are generally based on spatial–temporal contrast methods and low-rank decomposition models. The spatial–temporal local contrast filter (STLCF) [24] and spatial–temporal local difference measurement [25] algorithms expand the two-dimensional contrast operator to three-dimensional space to compute the spatial–temporal contrast between the current frame and the historical frames and fuse the output with the spatial-domain saliency results to extract the target location. The multiple-subspace learning and spatial–temporal patch-tensor model (MSLSTIPT) [26] can detect IRDSTs in multiframe infrared images by extending the low-rank decomposition process to three-dimensional space.
In the field of general-purpose image processing, two main deep learning-based multiframe processing methods are available. The method represented by DeepSORT [27] is based on single-frame detection, Hungarian correlation, and Kalman filtering for multiframe target detection [28]. The method represented by SiamFC [29] is based on deep correlation filtering for conducting multiframe processing after completing single-frame detection [30]. These factors are unfavorable for the detection of IRDSTs with low SCRs because of the threshold segmentation issues encountered during the single-frame detection process.
Deep learning tasks dedicated to multiframe IRDSTD are still in their infancy, but what is clear is that the fusion of spatiotemporal information facilitates the detection of IRDSTs with low SCRs. Liu et al. [31] proposed the use of a bidirectional convolutional long short-term memory structure and 3D convolution to learn the differences between IRDST and background spatiotemporal features by inputting multiframe infrared images. This approach can monitor the temporal fluctuations of each pixel, and it assumes that pixels exhibiting sudden temporal energy changes are likely to belong to the target. The motion differences between the target and clutter were utilized in DTUM [32] to design a direction-coded convolutional block (DCCB), which combines spatial saliency with linear motion features to improve the detectability of IRDSTs.

1.2. Motivation

However, the existing methods mentioned above still have several shortcomings and deficiencies.
The use of airspace alone is not sufficient for detecting IRDSTs with low SCRs since they are spatially nonsignificant, as shown in Figure 1a. The detectability of IRDSTs can be effectively improved by extracting the spatiotemporal features of targets from multiframe images.
As illustrated in Figure 1b, a common strategy involves obtaining detection results for each frame through single-frame detection and then using the correlations between consecutive frames to associate targets, thus enabling target detection across a sequence of images [27]. However, owing to the threshold segmentation operation implemented during single-frame detection, the useful information possessed by IRDSTs with low spatial significance levels and low SCRs tends to be lost, leading to missed detections of IRDSTs in sequential images. To solve this problem, a one-step multiframe detection method, as shown in Figure 1c, has been explored. It adopts the idea of accumulation. A spatial and temporal feature tensor is formed by the superposition of spatial saliency features without a loss, and this spatial and temporal feature tensor is mapped to the target probability with an end-to-end network. The integrated design method effectively combines the spatial and temporal features of IRDSTs, which can greatly improve the detectability of IRDSTs with low SCRs.
Herein, a spatiotemporally integrated detection network was proposed for IRDSTs, spatial saliency characteristics were extracted in a frame-by-frame manner via U-Net, and the characteristic tensor was obtained via reorganization according to the temporal dimension. Afterward, the motion characteristics of the target were extracted through the proposed convolution kernel for multiscale 3D motion features with fixed weights. Finally, a mapping from the features to the target probability likelihood map was established, benefiting from the spatiotemporal feature fusion module based on 3D convolution. In summary, this strategy successfully achieves multiframe IRDST detection.
Overall, the main contributions of this work are as follows.
(1)
The spatial and temporal dimensions were integrated into a multiframe IRDST detection network. The salient airspace features and time-domain motion characteristics of the target were incorporated into a unified network architecture based on 3D convolution, which achieved multiframe IRDST detection via an end-to-end network.
(2)
According to the consistency of the short-term moving direction of the IRDST, its multiscale three-dimensional motion features with fixed weights were convolved to synchronously enhance the temporal and spatial significance of the target and inhibit random clutter and noise, which improved the detectability of IRDSTs with low SCRs.
(3)
The spatiotemporal characteristics of five parallel and independent optimization strategies were incorporated into the designed spatiotemporal feature fusion-based processing module, which successfully mapped spatiotemporal tensors to target multiframe probability maps.
(4)
Compared with the current methods, the proposed strategy exhibited better detection performance, especially for IRDSTs with low SCRs.

2. Methods

In this section, a spatiotemporally integrated detection network (STIDNet) is realized and applied. First, the main components of the integrated detection network are detailed, the design of the feature extraction layer of the motion saliency and spatiotemporal saliency extraction network is displayed, and the specific implementation of the proposed algorithm under input sequence images is described.

2.1. Overall Architecture

We focus on the design of STIDNet; the network can independently learn the spatiotemporal characteristics of IRDSTs to strengthen its ability to detect IRDSTs with extremely low SCRs. As shown in Figure 2, the overall structure of the network consists of three parts—a spatial saliency feature generation module (SSFGM), a motion feature extraction module (MFEM), and a spatiotemporal feature fusion module (STFFM).
Section 2.2. presents the design of a U-shaped network for extracting the spatial saliency for IRDSTs. Section 2.3. describes the proposed multiscale motion feature-based 3D convolution kernel (FWMFCK-3D) for accumulating the spatial saliency of IRDSTs via a prior motion. Section 2.4. details the mapping from the spatiotemporal motion feature tensor to an IRDST probability. Finally, Section 2.5 and Section 2.6 introduce the loss functions and applications of the network, respectively.

2.2. Spatial Saliency Feature Generation Module

The network input is a sequence of infrared images T × W × H , where T denotes the length of the input sequence (5 frames are used in the paper), and W and H denote the width and height of the images, respectively. For a single-frame image I 5 × W × H in a sequence, a U-shaped network [33] is used as the backbone network to extract the significance of the spatial domain, and the weight parameters of the network are shared by all the images contained in the sequence.
To ensure that the diversity of the spatial features is conducive to the subsequent temporal information mining process, the head structure of the single-frame detection network is modified to expand the number of output channels from 1 to 32. This modification allows for the extraction of spatially significant feature tensors S t 32 × W × H , which are then concatenated in a frame-by-frame manner.
S 5 × 32 × W × H = c a t [ S 1 32 × W × H , S 2 32 × W × H , S 3 32 × W × H , S 4 32 × W × H , S 5 32 × W × H ]
where c a t [ ] denotes the concatenation operator.
The feature tensor S 5 × 32 × W × H is reconstructed as spatiotemporal tensor S T 32 × 5 × W × H according to the temporal dimension for the subsequent extraction of temporal motion features.

2.3. Motion Feature Extraction Module

In the infrared search and discovery system, because of the long target distance and the high frame rate of the input image, the speed of the target on the image plane is limited and the direction of the target motion has strong consistency within a short period. According to the above characteristics, we designed FWMFCK-3D to convolve the spatiotemporal tensor S T 32 × 5 × W × H and accumulate the spatial significance of IRDSTs according to their prior motion to improve the detectability of IRDSTs with low SCRs.
The eight 3D motion directions are shown in Figure 3; these are left-right, left-up, and right-down; up-down, right-up, and left-down; right-left, right-down, and left-up; and down-up, left-down, and right-up.
To adapt to targets with different motion speeds and scales, convolution kernels with four sizes—5 × 5 × 5, 5 × 7 × 7, 5 × 9 × 9, and 5 × 11 × 11—were designed, and their depth direction was 5. Eight 5 × 5 × 5 convolution kernels are shown in Figure 4. The value of the red pixel is taken to be 1, and the values of the non-red pixels are taken to be 0. The other sizes used for the fixed-weight convolution kernels were set up with reference to the 5 × 5 × 5 dimensions.
Thirty-two 3D convolution kernels with fixed weights are obtained, and they are denoted as follows:
3 D F W C K D i = c a t ( F W C K L R D i , F W C K L U R D D i , F W C K U D D i , F W C K R U L D D i , F W C K R L D i , F W C K R D L U D i , F W C K D U D i , F W C K L D R U D i )       D i 5 , 7 , 9 , 11
The padding was set so that the input and output W × H dimensions were consistent, and the spatiotemporal tensor was convolved with 3 D F W C K D i to obtain a spatiotemporal motion feature tensor S T M , as follows:
S T M = S T 32 × 5 × W × H 3 D F W C K D i
As show in Figure 5, S T 32 × 5 × W × H was convolved with the convolution kernel at each scale to generate a tensor possessing a size of 8 × 5 × W × H. The size of spatiotemporal motion feature tensor S T M obtained by splicing the tensors at four scales was 32 × 5 × W × H, which is consistent with that of S T 32 × 5 × W × H .
Ablation experiments were designed to demonstrate the effectiveness of this structure in Section 3.4.

2.4. Spatiotemporal Feature Fusion Module

The STFFM, which is based on 3D convolution, was adopted to map the spatiotemporal motion feature tensor to the probability map of the IRDST. As shown in Figure 6, a U-shaped network was employed in a spatiotemporal significance module (STSM) to better extract and fuse the shallow high-resolution and deep low-resolution feature maps. It used three rounds of up-and-down-sampling to extract features at different scales, and fuses the features with different depths via a cascading approach.
As shown in Figure 7, five STSMs were adopted to map STM to the target probability maps of the five frames. The modules operate in parallel with each other and without sharing weights.
P = { P 1 , , P 5 } = S T S M 1 ( S T M ) , , S T S M 5 ( S T M )       t 1 , 2 , 3 , 4 , 5
where P is the target probability likelihood map generated by STSMs.

2.5. Loss Function

IRDSTD is actually a binary classification problem involving each pixel point of the input image, but since the number of IRDST samples is extremely low in proportion to the background, the numbers of training samples are extremely imbalanced, which leads to slow training convergence. Therefore, the focus loss [34] was chosen as the loss function for the probability map of each image frame to increase the weights of the small number of positive samples and improve the convergence speed of the training process. The sum of the focus losses calculated for 5 image frames is used as the total loss of the network.
FL t ( P t , y t ) = y t 1 P t γ log ( P t ) 1 y t P t γ log ( 1 P t ) L o s s = FL t     t [ 1 , 2 , 3 , 4 , 5 ]
where y t is the label, and the value of γ is set as 2.

2.6. Result-Level Fusion in the Implementation

The application process for sequence images I N W × H = I 1 W × H , I 2 W × H , , I T W × H is shown in Figure 8a.
Step 1: Time slide window
A time slider was used to group the input sequences along the time direction to obtain the following network input:
G N 5 × W × H = G 1 5 × W × H , G 2 5 × W × H , , G T 4 5 × W × H
where
G 1 5 × W × H = S 1 , S 2 , S 3 , S 4 , S 5 , , G T 4 5 × W × H = S T 4 , S T 3 , S T 2 , S T 1 , S T
Step 2: Neural network processing
G N 5 × W × H was entered into the network to obtain the result I R N ;
I R N = F S T I D N e t ( G N 5 × W × H )       N [ 1 , 2 , , T 4 ]
where F S T I D N e t ( ) denotes inference using the STIDNet network via the trained weights.
Step 3: Result level fusion
For a frame such as I 10 W × H from the input sequence image I N W × H , as shown in Figure 8b,   I 10 W × H has 5 detection results in I R 6 ~ I R 10 , whose inputs are G 6 5 × W × H ~ G 10 5 × W × H   I 6 W × H , I 7 W × H , , I 10 W × H , , I 14 W × H , so the process of obtaining R 10 (the final result of I 10 W × H ) via I R 6 ~ I R 10 utilized up to 9 image frames, thus using more contextual information.
Specifically, we proposed the result set fusion strategy of the maximum method to obtain the final result, as shown in Figure 8b. The fusion formula is as follows:
R K ( i , j ) = M a x [ I R K 4 _ 5 ( i , j ) , I R K 3 _ 4 ( i , j ) , I R K 2 _ 3 ( i , j ) , I R K 1 _ 2 ( i , j ) , I R K _ 1 ( i , j ) ]       1 i W , 1 j H , 1 K T
where ( i , j ) is the pixel position, R K ( i , j ) is the probability of an IRDST pixel being located at ( i , j ) in the Kth frame, and M a x [ ] is the maximum operation. I R K 4 _ 5 ( i , j ) is the processing result of the network, and ( K 4 ) _ 5 denotes the 5th frame of the network result obtained for the network input G K 4 5 × W × H .

3. Experiments and Results

3.1. Dataset

The open-source dataset NUDT-MIRSDT [32] was used as the main training and validation data for the algorithm proposed in this paper.
Training dataset: This dataset included a total of 100 sequences, each containing 100 images, and each sequence constitutes 96 training groups, totaling 9600 training samples.
Test dataset: This dataset included a total of 20 sequences, totaling 1920 test samples.
In the test dataset, the mean SCR of the target across eight sequences is less than 3, making this a typical dim target. As the attention focus related to the data, the SCR change curve is shown in Figure 9.

3.2. Performance Evaluation Indices

Regarding the IRDST detection performance of the proposed approach, the receiver operating characteristic (ROC) curve, which can characterize global performance, and the signal clutter ratio gain (SCRG) and background suppression factor (BSF), which can measure local performance, were used as evaluation indices.
The vertical coordinate of the ROC curve is the true-positive rate (TPR), and the horizontal coordinate is the false-positive rate (FPR). The TPR and FPR measure the classification ability of a model from the perspectives of positive samples and negative samples, respectively. The larger the AUC value is, the stronger the classification ability.
The calculation formulas for the TPR and FPR are as follows:
F P R = F P N T P R = T P P
where TP, FP, P, and N represent the number of true positives (correct detections), the number of false positives (false alarms), the total number of positive samples, and the total number of negative samples, respectively. The area under the curve (AUC) is the quantitative index of the ROC curve.
BSF and SCRG are calculated as follows:
B S F = C i n C o u t
where C i n represents the standard deviation of the original infrared image and C o u t represents the standard deviation of the infrared image processed by the detection algorithm. A larger BSF indicates better background suppression.
S C R G = S C R o u t S C R i n
S C R = μ T μ B σ B
where S C R o u t and S C R i n denote the SCR of the original target and the SCR of the final target, respectively. μ T , μ B , and σ B represent the average pixel intensity of the target region, the average pixel intensity of the neighboring region around the target, and the standard deviation of the neighboring region, respectively. S C R G demonstrates the enhancement effect the algorithm attained for the target, and a larger S C R G indicates an algorithmic gain for the IRDST.

3.3. Network Training

The computer was equipped with an Intel(R) Core (TM) i9-14900K CPU @ 3.20 GHz (Intel Corporation, Santa Clara, United States of America) and an NVIDIA GeForce RTX 4090 GPU @ 24 GB (NVIDIA Corporation, Santa Clara, United States of America). The main software versions were as follows: Python 3.8.16, Pytorch 1.13.1, and Cuda 11.6.1.
Adaptive moment estimation (Adam) was used as the optimizer for training, the initial learning rate was set to 0.001, and the weight attenuation coefficient was set to 0.5. We adopted the ROC optimization method to save the best weights.
B e s t _ w e i g h t = W e i g h t e p o c h ,               i f   A U C e p o c h b e s t _ A U C B e s t _ w e i g h t ,       o t h e r w i s e
where A U C e p o c h is the AUC for the dataset in the current epoch. b e s t _ A U C is the historically best AUC. W e i g h t e p o c h is the weight for the current epoch. B e s t _ w e i g h t is the historically best weight. If A U C e p o c h b e s t _ A U C , we updated the optimal weights.

3.4. Ablation Study

To demonstrate the necessity of the MFEM and the result-level maximum fusion strategy, we designed the following groups for ablation experiments.
The three experimental groups were designated as STIDNet-NMFE-Sum, STIDNet-NMFE-Max, and STIDNet-NMFE-None, with None MFEM. However, they adopted different fusion strategies—Sum, Max, and with None fusion. A comparison of these three experiments may reveal which fusion strategy has advantages.
For the purpose of comparison, the remaining three experimental groups, namely STIDNet-MFE-Sum, STID-Net-MFE-Max, and STIDNet-MFE-None, all utilize the MFEM. This experiment aims to demonstrate the importance of the MFEM in extracting the motion features of IRDST.
The setups of the six groups of ablation experiments are shown in Table 1.
The same training data and strategy were used to train the network, and the optimal weights obtained from each group were used to statistically analyze the data with low SCRs (SCR < 3) in the test set. The obtained ROC curves are shown in Figure 10, and the AUC statistics are shown in Table 2.
(1) Evaluation of MFEM’s effect on IRDSTD performance
Under the same fusion method, the ROC curves and AUCs produced with and without MFEM were compared. From a qualitative perspective, as shown in Figure 10, the ROC curves of the experimental groups that employed MFEM (red) were, in general, superior to those that did not (green), as evidenced by their location in the upper left corner of the axis. From a quantitative perspective, the AUC enhancement of 0.0087/0.0068/0.0142 for a IRDST (SCR < 3) was achieved with the employed of MFEM by comparing STID-Net-MFE-Sum with STID-Net-NMFE-Sum, STID-Net-MFE-Max with STID-Net-NMFE-Max, and STID-Net-MFE-None with STID-Net-NMFE-None. This is attributable to the utilization of a priori information gathering energy from the motion of a IRDST with a low SCR using MFEM, thereby enhancing the distinction between the targets and clutter, and achieving a higher AUC.
(2) Evaluation of maximal fusion strategy effect on IRDSTD performance
The influence of the fusion strategy on the resulting performance was analyzed on the premise of ensuring the consistency of the motion feature extraction settings. The ROC curves and average AUCs of STIDNet-MFE-Max and STIDNet-MFE-Sum were better than those of STIDNet-MFE-None, which indicated that the performance of fusion processing was superior to that of non-fusion processing. This was due to the fact that the fusion strategy employs a greater quantity of contextual information for IRDSTD.
In particular, by comparing STIDNet-MFE-Max and STIDNet-MFE-Sum, as well as STIDNet-MFE-Max and STIDNet-MFE-None, the AUC values increased by 0.0023/0.0080, respectively. In the experimental group utilizing the MFEM, the AUC was also enhanced by 0.00043/0.00068 through the implementation of the maximum fusion strategy.
The above analysis results can be described as follows.
(1)
The AUC achieved on the test set by the network with motion feature extraction was better than that of the network without motion feature extraction. Motion feature extraction can significantly improve the ability to detect IRDSTs.
(2)
Fusion processing was superior to non-fusion processing. After conducting a comparison, the AUC of the maximum value fusion method was the highest.
Hence, in the subsequent comparative experiments with other methods, STIDNet-MFE-Max was adopted as the optimal version of the algorithm proposed in this work.

3.5. Visual Analysis of Feature Maps

(1) Visualization of MFEM’s usefulness
The following discussion attempted to provide a visual representation of the benefits of MFEM. A visual analysis was performed via the feature maps produced before and after the application of the MFEM, as shown in Figure 11.
As shown in Figure 11a, strong responses were observed for the target and lower-right building areas when utilizing the SSFGM without the MFEM. However, strong clutter was significantly attenuated at the edge of the building after the MFEM was applied. This was due to the fact that MFEM possessed the effect of enhancing the saliency of the moving IRDST, while the edge of the building was stationary and was thus suppressed after MFEM was applied.
Furthermore, as shown in Figure 11b, a weak target response was found in the results of the SSFGM, with incomplete extracted target edge features. After applying the MFEM, the target was completely and accurately extracted, which benefited the accumulation of information from multiple frames.
These results indicate that the MFEM can both effectively suppress false alarms and enhance target features.
(2) Visualization of STFFM ’s usefulness
Furthermore, as demonstrated in Figure 12, the efficacy of the STFFM was evidently substantiated.
Feature_O3, Feature_O2, and Feature_O1 represented the deepest, middle, and shallowest features of the upsampled outputs derived from the decoder in the STFFM, respectively.
As the feature map was decoded in a layer-by-layer manner within the network, the response of the target area became increasingly focused, whereas that of the background and clutter weakened. This was because the encoding and decoding of spatiotemporal feature maps was carried out in the STFFM, which accumulated the spatial energy of the IRDSTs in the time dimension, thereby significantly enhancing the saliency of the IRDST.

3.6. Comparison Experiments

To explore the performance of the method developed in this work, we compared it with the state-of-the-art IRDSTD methods, including traditional methods (MPCM, HBMLCM, WSLCM, RLCM, TLLCM, NIPPS, RIPT, WLDM, FKRW, MGRG, and STLCF) and methods based on deep learning (ISTDUNet, RISTDNet, DNANet, and DTUMNet). All the traditional methods used their default parameters, and the methods based on deep learning were retrained on the dataset to obtain the optimal weights. The specific parameter settings of the algorithms, the FPS results, and the parameters of the deep learning algorithms are shown in Table 3.
As shown in Figure 13., the performance of the proposed approach was compared with that of other algorithms, and the ROC curves are described as follows.
As evidenced by the ROC curves, deep learning consistently exhibited superior performance in comparison to traditional methods, particularly in scenarios where targets had low signal-to-noise ratios. The ROC curves of the other algorithms were obviously biased toward the lower-right corner of the coordinate axis, whereas those of the proposed algorithm were all in the upper-left corner of the coordinate system, particularly for low-SCR targets, indicating its excellent detection ability.
The AUCs of the ROC curves are illustrated as the quantitative analysis results in Table 4.
Compared with the other algorithms, the proposed method achieved the best performance when the SCR > 3. Moreover, when SCR < 3, the mean AUC value attained for eight sequences still reached 0.99786, indicating a significant advantage. The ROC curve and AUC results indicate that the developed algorithm has an excellent performance, particularly for IRDSTs with low SCRs.
Moreover, the performance differences observed between this algorithm and the other state-of-the-art methods in the local areas of small and dim targets are displayed in Table 5 and Table 6. The BSF and SCRG results indicate that the proposed method is at the leading level among the compared algorithms.
Typical processing results obtained for eight input images are displayed in Figure 14 to observe the detection capabilities of the tested algorithms.
Both the targets from Image 1 and Image 2, which had high SCRs, were successfully detected by all six listed algorithms. However, NIPPS, STLCF, and RISTDNet failed to detect the targets in Image 3–Image 8 because of their low SCRs. The targets in Image 3, Image 4, and Image 7 could be detected by ISTDUNet, whereas the targets in Image 5, Image 6, and Image 8 were not detected. Although DTUMNet successfully detected all the targets, false alarms occurred in Image 4 and Image 8. In the Image 2, Image 3, Image 5, and Image 7 results of DTUMNet, the targets were split into multiple regions and had no integrity, so this approach had difficulty in terms of accurately localizing the target.
Compared with these methods, the proposed method successfully detected all the targets without false alarms, and the detection results were centralized rather than dispersed, implying its satisfactory detection ability.
Furthermore, Figure 15 visualizes the results produced by the algorithm for the sequence images, where the targets were all accurately detected with low SCRs and no false alarms.

4. Conclusions

In this work, spatial saliency and temporal motion feature extraction were incorporated into a unified network architecture via an end-to-end design to construct a spatiotemporally integrated network for the detection of IRDSTs. According to the transcendental motion paradigm, the spatial saliency of the IRDST of interest was accumulated through an FWMFCK-3D model. The FWMFCK-3D model employed a fixed-weight convolutional kernel with eight directions and four scales to aggregate energy from IRDSTs. During spatiotemporal feature fusion processing, the feature tensor was mapped to the probability of the IRDST to detect the target via 3D convolution. In the ablation experiments, the AUC enhancement of 0.0087/0.0068/0.0142 for the IRDST (SCR < 3) was achieved with the FWMFCK-3D module in comparison with the results when the FWMFCK-3D module was not used. The implementation of a maximal fusion strategy resulted in the augmentation of the AUC by 0.0023/0.0080 and 0.00043/0.00068, in comparison to the summation and no fusion methodologies, respectively. Compared with other algorithms, the proposed algorithm presented the best ROC curves, AUCs, and BSFs on the NUDT-MIRSDT dataset, especially for IRDSTs with a SCR < 3, the average AUC value still reached 0.99786.

Author Contributions

Conceptualization: L.Z. and Q.H.; investigation: Z.Z. and Y.X.; methodology: L.Z., Z.Z., and Q.H.; software: Z.Z., F.T., and Y.X.; formal analysis: L.Z., Q.H., and F.T.; validation: Z.Z. and Y.X.; writing—original draft preparation: L.Z.; writing—review and editing: L.Z. and Q.H. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

The original data presented in the study are openly available in Li, R.J.; An, W.; Xiao, C.; Li, B.Y.; Wang, Y.Q.; Li, M.; Guo, Y.L. Direction-Coded Temporal U-Shape Module for Multiframe Infrared Small Target Detection. IEEE Trans. Neural Networks Learn. Syst. 2023, 36, 555–568. https://doi.org/10.1109/TNNLS.2023.3331004.

Acknowledgments

The authors would like to thank the National University of Defense Technology for providing the NUDT-MIRSDT dataset.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Kim, S.; Lee, J. Scale invariant small target detection by optimizing signal-to-clutter ratio in heterogeneous background for infrared search and track. Pattern Recognit. 2012, 45, 393–406. [Google Scholar] [CrossRef]
  2. Zhao, M.J.; Li, W.; Li, L.; Hu, J.; Ma, P.G.; Tao, R. Single-Frame Infrared Small-Target Detection: A Survey. IEEE Geosci. Remote Sens. Mag. 2022, 10, 87–119. [Google Scholar] [CrossRef]
  3. Deshpande, S.D.; Er, M.H.; Ronda, V.; Chan, P. Max-Mean and Max-Median filters for detection of small-targets. In Proceedings of the Signal and Data Processing of Small Targets, SPIE’s International Symposium on Optical Science Engineering, and Instrumentation, Denver, CO, USA, 4 October 1999; pp. 74–83. [Google Scholar]
  4. Han, J.H.; Liu, S.B.; Qin, G.; Zhao, Q.; Zhang, H.H.; Li, N.N. A Local Contrast Method Combined With Adaptive Background Estimation for Infrared Small Target Detection. IEEE Geosci. Remote Sens. Lett. 2019, 16, 1442–1446. [Google Scholar] [CrossRef]
  5. Bae, T.W.; Lee, S.H.; Sohng, K.I. Small target detection using the Bilateral Filter based on Target Similarity Index. IEICE Electron. Expr. 2010, 7, 589–595. [Google Scholar] [CrossRef]
  6. Bai, X.Z.; Zhou, F.G. Analysis of new top-hat transformation and the application for infrared dim small target detection. Pattern Recognit. 2010, 43, 2145–2156. [Google Scholar] [CrossRef]
  7. Zhang, S.F.; Huang, X.H.; Wang, M. Background Suppression Algorithm for Infrared Images Based on Robinson Guard Filter. In Proceedings of the 2nd International Conference on Multimedia and Image Processing (ICMIP), Wuhan, China, 17–19 March 2017; IEEE: New York, NY, USA, 2017; pp. 250–254. [Google Scholar]
  8. Chen, C.L.P.; Li, H.; Wei, Y.T.; Xia, T.; Tang, Y.Y. A Local Contrast Method for Small Infrared Target Detection. IEEE Trans. Geosci. Remote Sens. 2014, 52, 574–581. [Google Scholar] [CrossRef]
  9. Han, J.H.; Ma, Y.; Zhou, B.; Fan, F.; Liang, K.; Fang, Y. A Robust Infrared Small Target Detection Algorithm Based on Human Visual System. IEEE Geosci. Remote Sens. Lett. 2014, 11, 2168–2172. [Google Scholar] [CrossRef]
  10. Wei, Y.T.; You, X.G.; Li, H. Multiscale patch-based contrast measure for small infrared target detection. Pattern Recognit. 2016, 58, 216–226. [Google Scholar] [CrossRef]
  11. Han, J.H.; Liang, K.; Zhou, B.; Zhu, X.Y.; Zhao, J.; Zhao, L.L. Infrared Small Target Detection Utilizing the Multiscale Relative Local Contrast Measure. IEEE Geosci. Remote Sens. Lett. 2018, 15, 612–616. [Google Scholar] [CrossRef]
  12. Yao, S.K.; Chang, Y.; Qin, X.J. A Coarse-to-Fine Method for Infrared Small Target Detection. IEEE Geosci. Remote Sens. Lett. 2019, 16, 256–260. [Google Scholar] [CrossRef]
  13. Han, J.H.; Moradi, S.; Faramarzi, I.; Liu, C.Y.; Zhang, H.H.; Zhao, Q. A Local Contrast Method for Infrared Small-Target Detection Utilizing a Tri-Layer Window. IEEE Geosci. Remote Sens. Lett. 2020, 17, 1822–1826. [Google Scholar] [CrossRef]
  14. Gao, C.; Meng, D.; Yang, Y.; Wang, Y.; Zhou, X.; Hauptmann, A.G. Infrared patch-image model for small target detection in a single image. IEEE Trans Image Process. 2013, 22, 4996–5009. [Google Scholar] [CrossRef] [PubMed]
  15. Dai, Y.M.; Wu, Y.Q.; Song, Y.; Guo, J. Non-negative infrared patch-image model: Robust target-background separation via partial sum minimization of singular values. Infrared Phys. Technol. 2017, 81, 182–194. [Google Scholar] [CrossRef]
  16. Dai, Y.M.; Wu, Y.Q. Reweighted Infrared Patch-Tensor Model With Both Nonlocal and Local Priors for Single-Frame Small Target Detection. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2017, 10, 3752–3767. [Google Scholar] [CrossRef]
  17. Hou, Q.Y.; Wang, Z.P.; Tan, F.J.; Zhao, Y.; Zheng, H.L.; Zhang, W. RISTDnet: Robust Infrared Small Target Detection Network. IEEE Geosci. Remote Sens. Lett. 2022, 19. [Google Scholar] [CrossRef]
  18. Tong, X.Z.; Sun, B.; Wei, J.Y.; Zuo, Z.; Su, S.J. EAAU-Net: Enhanced Asymmetric Attention U-Net for Infrared Small Target Detection. Remote Sens. 2021, 13, 3200. [Google Scholar] [CrossRef]
  19. Li, B.Y.; Xiao, C.; Wang, L.G.; Wang, Y.Q.; Lin, Z.P.; Li, M.; An, W.; Guo, Y.L. Dense Nested Attention Network for Infrared Small Target Detection. IEEE Trans. Image Process. 2023, 32, 1745–1758. [Google Scholar] [CrossRef]
  20. Dai, Y.M.; Wu, Y.Q.; Zhou, F.; Barnard, K. Attentional Local Contrast Networks for Infrared Small Target Detection. IEEE Trans. Geosci. Remote Sens. 2021, 59, 9813–9824. [Google Scholar] [CrossRef]
  21. Zuo, Z.; Tong, X.Z.; Wei, J.Y.; Su, S.J.; Wu, P.; Guo, R.Z.; Sun, B. AFFPN: Attention Fusion Feature Pyramid Network for Small Infrared Target Detection. Remote Sens. 2022, 14, 3412. [Google Scholar] [CrossRef]
  22. Wang, K.W.; Du, S.Y.; Liu, C.X.; Cao, Z.G. Interior Attention-Aware Network for Infrared Small Target Detection. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5002013. [Google Scholar] [CrossRef]
  23. Wu, X.; Hong, D.F.; Chanussot, J. UIU-Net: U-Net in U-Net for Infrared Small Object Detection. IEEE Trans. Image Process. 2023, 32, 364–376. [Google Scholar] [CrossRef] [PubMed]
  24. Deng, L.Z.; Zhu, H.; Tao, C.; Wei, Y.T. Infrared moving point target detection based on spatial-temporal local contrast filter. Infrared Phys. Technol. 2016, 76, 168–173. [Google Scholar] [CrossRef]
  25. Zhu, H.; Guan, Y.S.; Deng, L.Z.; Li, Y.S.; Li, Y.J. Infrared moving point target detection based on an anisotropic spatial-temporal fourth-order diffusion filter. Comput. Electr. Eng. 2018, 68, 550–556. [Google Scholar] [CrossRef]
  26. Sun, Y.; Yang, J.G.; An, W. Infrared Dim and Small Target Detection via Multiple Subspace Learning and Spatial-Temporal Patch-Tensor Model. IEEE Trans. Geosci. Remote Sens. 2021, 59, 3737–3752. [Google Scholar] [CrossRef]
  27. Wojke, N.; Bewley, A.; Paulus, D. Simple Online and Realtime Tracking with a Deep Association Metric. In Proceedings of the 24th IEEE International Conference on Image Processing (ICIP), Beijing, China, 17–20 September 2017; IEEE: New York, NY, USA, 2017; pp. 3645–3649. [Google Scholar]
  28. Zhang, T.; Zhao, D.F.; Chen, Y.S.; Zhang, H.L.; Liu, S.L. DeepSORT with siamese convolution autoencoder embedded for honey peach young fruit multiple object tracking. Comput. Electron. Agric. 2024, 217, 108583. [Google Scholar] [CrossRef]
  29. Bertinetto, L.; Valmadre, J.; Henriques, J.F.; Vedaldi, A.; Torr, P.H.S. Fully-Convolutional Siamese Networks for Object Tracking. In Proceedings of the Computer Vision ECCV 2016 Workshops, PT II, Amsterdam, The Netherlands, 8–10 and 15–16 October 2016; Springer: Berlin, Germany, 2016; pp. 850–865. [Google Scholar]
  30. Mei, Y.P.; Yan, N.; Qin, H.X.; Yang, T.; Chen, Y.Y. SiamFCA: A new fish single object tracking method based on siamese network with coordinate attention in aquaculture. Comput. Electron. Agric. 2024, 216, 108542. [Google Scholar] [CrossRef]
  31. Liu, X.; Li, X.Y.; Li, L.Y.; Su, X.F.; Chen, F.S. Dim and Small Target Detection in Multi-Frame Sequence Using Bi-Conv-LSTM and 3D-Conv Structure. IEEE Access 2021, 9, 135845–135855. [Google Scholar] [CrossRef]
  32. Li, R.J.; An, W.; Xiao, C.; Li, B.Y.; Wang, Y.Q.; Li, M.; Guo, Y.L. Direction-Coded Temporal U-Shape Module for Multiframe Infrared Small Target Detection. IEEE Trans. Neural Networks Learn. Syst. 2023, 36, 555–568. [Google Scholar] [CrossRef]
  33. Hou, Q.Y.; Zhang, L.W.; Tan, F.J.; Xi, Y.Y.; Zheng, H.L.; Li, N. ISTDU-Net: Infrared Small-Target Detection U-Net. IEEE Geosci. Remote Sens. Lett. 2022, 19, 7506205. [Google Scholar] [CrossRef]
  34. Lin, T.-Y.; Goyal, P.; Girshick, R.; He, K.; Dollar, P. Focal Loss for Dense Object Detection. IEEE Trans. Neural Networks Learn. Syst. 2020, 42, 318–327. [Google Scholar] [CrossRef] [PubMed]
  35. Shi, Y.F.; Wei, Y.T.; Yao, H.; Pan, D.H.; Xiao, G.R. High-Boost-Based Multiscale Local Contrast Measure for Infrared Small Target Detection. IEEE Geosci. Remote Sens. Lett. 2018, 15, 33–37. [Google Scholar] [CrossRef]
  36. Han, J.H.; Moradi, S.; Faramarzi, I.; Zhang, H.H.; Zhao, Q.; Zhang, X.J.; Li, N. Infrared Small Target Detection Based on the Weighted Strengthened Local Contrast Measure. IEEE Geosci. Remote Sens. Lett. 2021, 18, 1670–1674. [Google Scholar] [CrossRef]
  37. Deng, H.; Sun, X.P.; Liu, M.L.; Ye, C.H.; Zhou, X. Small Infrared Target Detection Based on Weighted Local Difference Measure. IEEE Trans. Geosci. Remote Sens. 2016, 54, 4204–4214. [Google Scholar] [CrossRef]
  38. Qin, Y.; Bruzzone, L.; Gao, C.Q.; Li, B. Infrared Small Target Detection Based on Facet Kernel and Random Walker. IEEE Trans. Geosci. Remote Sens. 2019, 57, 7104–7118. [Google Scholar] [CrossRef]
  39. Huang, S.Q.; Peng, Z.M.; Wang, Z.R.; Wang, X.Y.; Li, M.H. Infrared Small Target Detection by Density Peaks Searching and Maximum-Gray Region Growing. IEEE Geosci. Remote Sens. Lett. 2019, 16, 1919–1923. [Google Scholar] [CrossRef]
Figure 1. Schematic diagram of current detection methods. The red squares represent the location of the IRDSTs.
Figure 1. Schematic diagram of current detection methods. The red squares represent the location of the IRDSTs.
Remotesensing 17 00250 g001
Figure 2. Overall network structure.
Figure 2. Overall network structure.
Remotesensing 17 00250 g002
Figure 3. Eight 3D motion directions.
Figure 3. Eight 3D motion directions.
Remotesensing 17 00250 g003
Figure 4. Three-dimensional fixed-weight convolution kernel ( F W C K D 5 ). The value of the red pixel is taken to be 1, and the values of the non-red pixels are taken to be 0.
Figure 4. Three-dimensional fixed-weight convolution kernel ( F W C K D 5 ). The value of the red pixel is taken to be 1, and the values of the non-red pixels are taken to be 0.
Remotesensing 17 00250 g004
Figure 5. Detailed graphical representation of Equation (3).
Figure 5. Detailed graphical representation of Equation (3).
Remotesensing 17 00250 g005
Figure 6. Spatiotemporal significance module.
Figure 6. Spatiotemporal significance module.
Remotesensing 17 00250 g006
Figure 7. Spatiotemporal feature fusion module.
Figure 7. Spatiotemporal feature fusion module.
Remotesensing 17 00250 g007
Figure 8. Fusion of result level in implementation. (a) Process flow, (b) The details of Step 3.
Figure 8. Fusion of result level in implementation. (a) Process flow, (b) The details of Step 3.
Remotesensing 17 00250 g008
Figure 9. The SCR curves of the sequences.
Figure 9. The SCR curves of the sequences.
Remotesensing 17 00250 g009
Figure 10. ROC curves for ablation experiments.
Figure 10. ROC curves for ablation experiments.
Remotesensing 17 00250 g010aRemotesensing 17 00250 g010b
Figure 11. The feature map generated by SSFGM and MFEM. For better visualization, the target area is highlighted by a red circle. The false alarm area is marked by a yellow tag.
Figure 11. The feature map generated by SSFGM and MFEM. For better visualization, the target area is highlighted by a red circle. The false alarm area is marked by a yellow tag.
Remotesensing 17 00250 g011
Figure 12. The feature maps from the different layers of the STFFM. The target area is highlighted by a red tag.
Figure 12. The feature maps from the different layers of the STFFM. The target area is highlighted by a red tag.
Remotesensing 17 00250 g012
Figure 13. ROC curves for comparative experiments. All results will be made public on https://github.com/zhanglw882/STIDNet (accessed on 1 August 2024).
Figure 13. ROC curves for comparative experiments. All results will be made public on https://github.com/zhanglw882/STIDNet (accessed on 1 August 2024).
Remotesensing 17 00250 g013aRemotesensing 17 00250 g013b
Figure 14. Visualized results of different methods. Image 1 and Image 2 are from Seq. 1 and Seq. 2 in the high SCR group and Image 3–Image 8 are from Seq. 1–Seq. 6 in the low SCR group. For better visualization, the target area is enlarged in the top-right corner and highlighted by a yellow circle. Missed targets are marked with dotted boxes. The false alarm area is marked by a green tag.
Figure 14. Visualized results of different methods. Image 1 and Image 2 are from Seq. 1 and Seq. 2 in the high SCR group and Image 3–Image 8 are from Seq. 1–Seq. 6 in the low SCR group. For better visualization, the target area is enlarged in the top-right corner and highlighted by a yellow circle. Missed targets are marked with dotted boxes. The false alarm area is marked by a green tag.
Remotesensing 17 00250 g014
Figure 15. Visualized results of our methods in sequence. Specifically, they are Seq. 7 and Seq. 8 from the low SCR group, showing a total of 9 frames, e.g., Frame 10, Frame 20, Frame 30, etc. More test results will be made public on https://github.com/zhanglw882/STIDNet (accessed on 1 August 2024).
Figure 15. Visualized results of our methods in sequence. Specifically, they are Seq. 7 and Seq. 8 from the low SCR group, showing a total of 9 frames, e.g., Frame 10, Frame 20, Frame 30, etc. More test results will be made public on https://github.com/zhanglw882/STIDNet (accessed on 1 August 2024).
Remotesensing 17 00250 g015
Table 1. The setup of the ablation experiments.
Table 1. The setup of the ablation experiments.
No.MethodsMotion Feature ExtractionFusion Method
1STIDNet-NMFE-SumSUM
2STIDNet-NMFE-MaxMAX
3STIDNet-NMFE-None
4STIDNet-MFE-SumSUM
5STIDNet-MFE-MaxMAX
6STIDNet-MFE-None
Table 2. AUC results in the ablation experimental group.
Table 2. AUC results in the ablation experimental group.
No.STIDNet-NMFE-SumSTIDNet-NMFE-MaxSTIDNet-NMFE-NoneSTIDNet-MFE-SumSTIDNet-MFE-MaxSTIDNet-MFE-None
Seq. 10.999060.999160.999070.999210.999210.99916
Seq. 20.950760.965220.924760.998320.999040.9945
Seq. 30.999090.999060.998890.999120.999120.99913
Seq. 40.998780.998760.998660.998770.998760.99878
Seq. 50.998850.998850.998850.998850.998850.99885
Seq. 60.998490.998810.998710.998890.998880.99889
Seq. 70.999310.99930.999310.999310.99930.99931
Seq. 80.96530.96950.945870.986990.989740.9888
Mean0.9887050.99108250.9830150.99743250.99786250.9971775
Table 3. Parameterization of the contrast method.
Table 3. Parameterization of the contrast method.
MethodsPublicationParameter SettingsParams (M)FPS
MPCM [10]Pattern Recognit. 2016L = 3; Window size = [3, 5, 7, 9]-36.331
HBMLCM [35]IEEE Geosci. Remote Sens. Lett. 2018External window size = 15 × 15;
target size = [3, 5, 7, 9]
-126.098
WSLCM [36]IEEE Geosci. Remote Sens. Lett. 2021Gauss_krl = [1, 2, 1; 2, 4, 2; 1, 2, 1]./16; Scs = [5, 7, 9, 11]-0.801
RLCM [11]IEEE Geosci. Remote. Sens. Lett. 2018Scale = 3; k1 = [2, 5, 9]; k2 = [4, 9, 16]-1.024
TLLCM [13]IEEE Geosci. Remote. Sens. Lett. 2019GS = [1/16, 1/8, 1/16; 1/8, 1/4, 1/8; 1/16, 1/8, 1/16]-1.533
NIPPS [15]Infr. Phys. Technol. 2017PatchSize = 50; SlideStep = 10; LambdaL = 2; RatioN = 0.005-0.249
RIPT [16]IEEE J. Sel. Topics Appl. Earth Observ. Remote Sens. 2017PatchSize = 30; SlideStep = 10; LambdaL = 0.7; MuCoef = 5; h = 1-1.238
WLDM [37]IEEE Trans. Geosci. Remote Sens. 2016L = 9-0.389
FKRW [38]IEEE Trans. Geosci. Remote Sens. 2019L = [−4, −1, 0, −1, −4; −1, 2, 3, 2, −1; 0, 3, 4, 3, 0;−1, 2, 3, 2, −1; −4, −1, 0, −1, −4]-9.034
MGRG [39]IEEE Geosci. Remote. Sens. Lett. 2019numSeeds = 20; tarRate = 0.01 × 0.15-1.422
STLCF [25]Computers and Electrical Engineering. 2018tspan_rng = 2; swind_rng = 7-2.749
ISTDUNet [33]IEEE Trans. Geosci. Remote Sens. Lett. 2022hyperparameter of channel = 22.76123.027
RISTDNet [17]IEEE Trans. Geosci. Remote Sens. Lett. 2022-0.7634.723
DNANet [19]IEEE Trans. Image Process. 2023-1.13420.747
DTUMNet [32]IEEE Trans. on Neural Networks and Learning Systems. 2023Res-Unet0.2989.494
STIDNet--4.8745.945
Table 4. AUC results of comparative experiments.
Table 4. AUC results of comparative experiments.
Seq.WSLCMRLCMTLLCMNIPPSRIPTSTLCFISTDUNetRISTDNetDNANetDTUMNetSTIDNet
S
C
R
>
3
10.99870.99720.99790.99870.90870.98330.99870.99870.97270.99870.99865
20.96890.97130.99900.86900.82410.99260.99900.99720.98900.99900.99901
30.99890.99890.99890.99390.88400.99890.99890.99890.97750.99890.9989
40.99890.99890.99890.99890.99890.99890.99890.99890.99840.99890.99885
50.99810.99330.99650.99820.68370.98440.99820.99820.99820.99820.9982
60.99870.99300.99850.97870.72900.99870.99870.99870.50540.99870.99866
70.99900.99360.99900.99900.99900.99900.99900.99900.99900.99900.99903
80.95010.95770.99570.56910.49930.98230.99540.96410.76760.99870.99873
90.98810.99590.99810.94260.49940.98710.99870.99830.93620.99890.99889
100.99820.99590.99830.99790.59400.99520.99840.99850.48480.99850.99845
110.97800.92490.99830.88850.64900.99650.99850.99780.94380.99850.99853
120.99900.99890.99900.99890.99890.99830.99900.99900.99890.99890.99895
Mean0.989550.984960.998180.936120.772330.992930.998450.995610.880960.998740.99874
S
C
R
<
3
10.517610.509630.877210.725350.518330.975730.953660.79330.783890.988420.99921
20.52130.78610.98320.49960.49930.73480.99330.58190.45790.77370.99904
30.55050.53430.85370.55890.55090.89410.99620.93190.64910.98340.99912
40.54360.46580.88970.49940.49890.63920.99000.83210.57950.90530.99876
50.52740.59130.97780.69230.49810.78930.99790.97720.89190.99890.99885
60.50420.48590.92440.50910.49930.85810.98330.90370.66590.99750.99888
70.49920.47800.57880.73510.56560.97890.99930.85760.74100.81560.9993
80.49890.76420.92020.76580.62770.92870.92170.72480.81540.92850.98974
Mean0.520340.576900.875630.623190.532270.849850.979420.825310.698070.923920.99786
Table 5. BSF results of comparative experiments. The best results are in red; the second-best results are in blue; and the third-best results are in green.
Table 5. BSF results of comparative experiments. The best results are in red; the second-best results are in blue; and the third-best results are in green.
Seq.WSLCMRLCMTLLCMNIPPSRIPTSTLCFISTDUNetRISTDNetDNANetDTUMNetSTIDNet
S
C
R
>
3
138.65.69.14278.332.52.11358.4278.35.71365.32251.7
2988.57.29.59524.61860.22.9240.347.61.3588.31711.0
32109.615.464.0257.91046.15.9746.3204.17.1795.71848.9
420.33.06.411.28.12.352.48.42.71300.91557.6
514.21.42.615.219.51.313.711.50.9529.6935.2
618.53.46.811.024.32.243.539.41.6534.51484.5
723.11.33.46.49.81.18.114.90.3196.0410.6
813.71.32.37.3619.71.452.421.20.6509.5598.9
9688.25.314.555.128.62.1589.5114.63.0652.31131.9
1028.710.321.917.828.44.344.214.38.31392.92205.1
1121.06.210.535.155.01.8765.7340.94.2630.01814.0
122.82.32.337.182.61.2126.529.98.1467.41445.0
S
C
R
<
3
125.54.926.451.046.42.633.512.54.9708.61163.0
222.58.09.342.220.51.4147.929.25.21910.9887.6
322.24.05.617.57.21.656.631.01.4501.01045.2
415.12.76.924.56.41.513.016.11.8965.81164.4
5137.52.13.339.08.71.496.711.11.6352.3545.8
684.78.716.532.126.52.1204.627.34.4487.71768.7
7104.94.76.774.7622.72.3414.874.73.6857.61687.2
81054.46.012.785.838.92.2624.2502.94.11341.51629.2
Table 6. SCRG results of comparative experiments. The best results are in red; the second-best results are in blue; and the third-best results are in green.
Table 6. SCRG results of comparative experiments. The best results are in red; the second-best results are in blue; and the third-best results are in green.
Seq.WSLCMRLCMTLLCMNIPPSRIPTSTLCFISTDUNetRISTDNetDNANetDTUMNetSTIDNet
S
C
R
>
3
11313.71.0277.21518.3614.13.4578.424.91.11022.41241.5
22194.11.51492.11746.6951.82.468.95.30.1659.41127.2
32806.61.61718.91239.013914.0919.31053.91.81133.91817.0
40.00.3182.714.80.05.8256.888.02.39490.66081.9
545.91.43.516.40.01.233.98.60.31430.91620.5
6753.01.4819.3190.10.01.34.45.50.2438.7221.8
7535.61.2882.461.626.11.08.34.50.1422.4203.1
8793.90.71352.1918.6421.31.627.914.10.21141.4651.4
91664.21.2149.61107.812352.2417.156.01.01004.71381.4
100.40.17.15.20.02.06.028.20.75321.52736.5
110.00.1145.3666.2633.24.3683.0215.73.02859.42466.1
120.00.023221458.015711412.725336.618766.92130.77416719398266
S
C
R
<
3
120.80.612.318.85.02.9131.939.90.83263.63281.2
20.80.82.90.00.01.1251.71.11.16947.34837.8
36.60.618.1231.9141.43.1360.273.31.64968.610788
41.50.0036.00.00.02.116.56.60.64408.64652.2
51094.30.827.71085.5549.21.420.615.30.7593.1847.8
6832.51.6112.5485.3789.31.265.0105.00.9454.41340.9
71275.71.123.8890.2961.41.364.889.00.6715.6566.5
81440.60.968.41053.61117.32.172.4487.71.01235.31101.9
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhang, L.; Zhou, Z.; Xi, Y.; Tan, F.; Hou, Q. STIDNet: Spatiotemporally Integrated Detection Network for Infrared Dim and Small Targets. Remote Sens. 2025, 17, 250. https://doi.org/10.3390/rs17020250

AMA Style

Zhang L, Zhou Z, Xi Y, Tan F, Hou Q. STIDNet: Spatiotemporally Integrated Detection Network for Infrared Dim and Small Targets. Remote Sensing. 2025; 17(2):250. https://doi.org/10.3390/rs17020250

Chicago/Turabian Style

Zhang, Liuwei, Zhitao Zhou, Yuyang Xi, Fanjiao Tan, and Qingyu Hou. 2025. "STIDNet: Spatiotemporally Integrated Detection Network for Infrared Dim and Small Targets" Remote Sensing 17, no. 2: 250. https://doi.org/10.3390/rs17020250

APA Style

Zhang, L., Zhou, Z., Xi, Y., Tan, F., & Hou, Q. (2025). STIDNet: Spatiotemporally Integrated Detection Network for Infrared Dim and Small Targets. Remote Sensing, 17(2), 250. https://doi.org/10.3390/rs17020250

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop