Next Article in Journal
Transformer-Based Decomposition of Electrodermal Activity for Real-World Mental Health Applications
Previous Article in Journal
Termite Detection Techniques in Embankment Maintenance: Methods and Trends
Previous Article in Special Issue
YOLOv8-LCNET: An Improved YOLOv8 Automatic Crater Detection Algorithm and Application in the Chang’e-6 Landing Area
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Infrared Small Target Detection via Modified Fast Saliency and Weighted Guided Image Filtering

1
Institute of Optics and Electronics, Chinese Academy of Sciences, Chengdu 610209, China
2
National Laboratory on Adaptive Optics, Chengdu 610209, China
*
Author to whom correspondence should be addressed.
Sensors 2025, 25(14), 4405; https://doi.org/10.3390/s25144405
Submission received: 22 April 2025 / Revised: 25 June 2025 / Accepted: 5 July 2025 / Published: 15 July 2025

Abstract

The robust detection of small targets is crucial in infrared (IR) search and tracking applications. Considering that many state-of-the-art (SOTA) methods are still unable to suppress various edges satisfactorily, especially under complex backgrounds, an effective infrared small target detection algorithm inspired by modified fast saliency and the weighted guided image filter (WGIF) is presented in this paper. Initially, the fast saliency map modulated by the steering kernel (SK) is calculated. Then, a set of edge-preserving smoothed images are produced by WGIF using different filter radii and regularization parameters. After that, utilizing the fuzzy sets technique, the background image is predicted reasonably according to the results of the saliency map and smoothed or non-smoothed images. Finally, the differential image is calculated by subtracting the predicted image from the original one, and IR small targets are detected through a simple thresholding. Experimental results on four sequences demonstrate that the proposed method can not only suppress background clutter effectively under strong edge interference but also detect targets accurately with a low false alarm rate.

1. Introduction

Infrared small target detection is critical for many military applications, especially for infrared search and tracking systems [1]. Unfortunately, detecting an infrared small target remains a challenging task, primarily due to following reasons: [2] (1) targets are typically small and lack a defined shape or texture, and (2) they are often submerged in a complex background with low signal-to-clutter ratio. Moreover, the background edges exhibiting high contrast and complexity might be falsely detected as targets, which lead to a high false alarm rate [3]. Although numerous methods have been proposed to handle the IR small target detection over the last two decades, it remains an open problem.
Currently, single frame IR small target detection methods can be broadly categorized into four groups: the filter-based model, the human visual system (HVS)-based model, the low-rank and sparse decomposition-based model, and the deep learning-based model. The filter-based approaches, such as the Max-Mean/Max-Median filter [4], the two-dimensional least mean square (TDLMS) filter [5,6], the bilateral filter [7], and the morphological filter [8,9,10,11,12], were developed to suppress background and clutter. They are mostly simple and fast but always enhance the edges of the background and make it difficult to determine whether it is the target or the edge during the detection process [2]. Leveraging the low-rank property of the background and the sparsity of the targets, the infrared patch image (IPI) model [1] pioneered the low-rank and sparse decomposition-based methods. Subsequently, numerous efforts have focused on improving its performance, such as the weighted infrared patch image (WIPI) model [13], the non-convex rank approximation minimization joint l2,1 norm (NRAM) [14], the partial sum of the tensor nuclear norm (PSTNN) [15], target-aware non-local low-rank modeling with saliency filtering regularization (TNLRS) [2], the image patch tensor model (IPT) [16], nonconvex tensor fibered rank approximation (NTFRA) [17], and the edge and corner awareness-based spatial-temporal tensor (ECA-STT) [18] model. Although the previously mentioned methods WIPI [13] and NRAM [14] have made significant progress to remove the edge residuals, they cannot fully eliminate the strong local clutters of various shapes completely by employing only a specific sophisticated norm to replace the nuclear norm [19]. Separately, HVS-based methods have been proposed regarding the dissimilarity between the current location and its neighborhood; these include the local contrast measure (LCM) [20], multi-scale patch based contrast measure (MPCM) [21], relative local contrast measure (RLCM) [22], derivative entropy-based contrast measure (DECM) [23], tri-layer local contrast measure (TLLCM) [24], double-neighborhood gradient method (DNGM) [25], strengthened robust local contrast measure (SRLCM) [26], and adaptive scale patch-based contrast measure (ASPCM) [27]. The above models usually adopt a sliding window to traverse the original image and have a fast-running speed, but most of them fail to maintain high performance when encountering a heterogeneous background exhibiting similar characteristics across local regions [2,28]. Recently, due to its automatic feature-learning ability for image hierarchical features, deep learning has been widely used for small target detection, such as adversarial learning for small object segmentation [29], the robust infrared small target detection network (RISTDnet) [30], the infrared small target detection with generative adversarial network (IRSTD-GAN) [31], the attentional local contrast network (ALCNet) [32], and the cross-connected bidirectional pyramid network (CBP-Net) [33]. However, the learning-based approaches are not universally applicable in IR small target detection, perhaps due to factors such as limited training data or requirements for predictability in novel scenarios.
Specifically, the spatial domain filter-based technique, which seeks to predict the background as accurately as possible, is a simple yet effective tool for infrared small target detection. Within this technique, the background estimation is the most important step, usually conducted as follows: [34] filtering is first applied locally centered on each pixel. The original pixel value is then replaced with the calculated result; this process is repeated for the entire image to generate the background prediction image. An essential consideration for background estimation is the need to preserve various edges in the background, since edges are highly prone to producing false alarms. However, most of the current spatial filter-based methods perform well on smooth backgrounds but yield less satisfactory results when encountering complex scenes containing different kinds of edges, such as the above-mentioned Max-Mean/Max-Median filter [4], Top-Hat filter [8], and TDLMS filter [5]. To overcome this shortcoming, Bae et al. [6,7] proposed edge directional bilateral/TDLMS filter-based methods, utilizing the edge information of surrounding prediction pixels in four directions. These methods can work well when the edges are obvious and strong, while it is still inadequate to cope with sophisticated edges, which are complex and vary in the real scenes.
As mentioned above, a crucial challenge in small target detection is to recover objects as much as possible while distinguishing them from the most representative clutters of the image, such as edges, corners, and other sharp structures. To solve this problem, we proposed a background estimation method for small target detection via fast saliency [35] and WGIF [36]. This method can eliminate or reduce edges of varying degrees and directions effectively and extract real targets under complex natural scenes. In addition, the proposed algorithm is suitable for parallel processing, which is useful for improving the detection speed. To the best of our knowledge, the proposed approach is the first work that combines the advantage of WGIF and fast visual saliency for infrared small target detection. Notably Ref. [37] applied the guided image filter to IR small target detection, but only employed it as a preprocessing step to strengthen the low-rankness of background components and suppress noise components.
The rest of this paper is organized as follows. The second section describes the proposed method in detail, followed by experimental results and analysis in the third section. We demonstrate the performance of the proposed method by comparing it with other SOTA methods. Finally, the paper is concluded in the fourth section.

2. Proposed Algorithm

Figure 1 illustrates the complete flowchart of the proposed infrared small target detection method. The fast saliency map is derived using the facet kernel described in Ref. [35]. Subsequently, the steering kernel [38] is utilized to remove residual strong edges. The modulated saliency map represents the degree of the smoothing effect for each pixel. On the other hand, the edge-preserving smoothed images are acquired from a series of WGIF operations. Then the predicted background image pixel value is computed via fuzzy sets from the smoothed or non-smoothed images according to the previous modulated saliency map. Finally, the target is extracted from the subtracted image by applying simple thresholding.

2.1. Saliency Map Calculation

Since the conventional spatial domain filter is not designed specifically for small target detection, a saliency map representing the probability of small target presence is delicately constructed for the subsequent background prediction process. This enables more accurate region classification.
Firstly, the gradient magnitude map R is obtained by convolving the original infrared image D with the 5 × 5 facet kernel F [35] as follows:
R = D × F , F = 4 1 0 1 4 1 2 3 2 1 0 3 4 3 0 1 2 3 2 1 4 1 0 1 4
The square value for each pixel is computed in the map R to get the enhanced map E:
E = R 2
Typically the most salient point in the enhanced map E is a target. However, it still contains numerous edge interferences as shown in the fast saliency map of Figure 1. Thus, this result should be refined to increase confidence in the small target presence.
The steering kernel is an attractive tool in image processing [38]. Extensive experiments have shown that feature descriptors using SK are robust to brightness variation and noise interference. Unfortunately, the SK descriptor of a small-target patch highly resembles that of a texture clutter patch but differs dramatically from a structural edge region [39]. Hence, we use SK to suppress residual edges in the previous enhanced map E.
The descriptive power of the SK mainly derives from a symmetric gradient covariance matrix C i , which can be estimated as:
C ^ i G i T G i , G i = D k , x 1 D k , x 2 , k ω i
Here, G i is the local gradient matrix, ω i is the square window centered at pixel i, and D k , x 1 and D k , x 2 are the first derivatives along x 1 and x 2 directions at pixel k, which are computed by the second order classic kernel regression method. To improve robustness and stability, the gradient covariance matrix is decomposed into three components as follows [38]:
C i = γ i U θ i Λ i U θ i T
U θ i = cos θ i sin θ i sin θ i cos θ i
Λ i = σ i 0 0 σ i 1
where U θ i is a rotation matrix and Λ i is the elongation matrix. The elongation parameter σ i , scaling parameter γ i , and rotation parameter θ i are all determined by the singular value decomposition (SVD) of the local gradient matrix G i . If the diagonal matrix of SVD is denoted as d i a g ( s 1 , s 2 ) , the elongation parameter σ i corresponding to the energy of the dominant gradient direction is denoted as:
σ i = s 1 + λ s 2 + λ , λ 0
The scaling parameter γ i is given as:
γ i = s 1 s 2 + λ M 1 2
where λ and λ are regularization parameters, and M is the size of local patch. Considering the meanings of the elongation and scaling parameters, similar to Ref. [13], the edge unlikelihood coefficient is constructed as follows:
p i = 1 σ i γ i
Consequently, the SK modulated saliency map S is defined as:
S = E P
Leveraging the structurally informative edge unlikelihood coefficient, the proposed SK modified saliency map S effectively eliminates prominent edges while preserving targets, as shown in Figure 1.

2.2. Weighted Self-Guided Image Filtering

By incorporating an edge-aware weighting into an existing guided image filter [40], the WGIF [36] is introduced to produce images with excellent visual quality and avoid halo artifacts like the existing global smoothing filters. Due to the WGIF’s outstanding edge-preserving smoothing property, we adopted this method to estimate the candidate background for small targets. Since the guided image is the same as the input infrared image D, we refer to this self-guided image filtering.
Let V a r k be the variance of D in the 11 × 11 local window; an edge-aware weighting w k is defined for all pixels as follows:
w k = 1 N k = 1 N V a r k + ε V a r k + ε
where N is the number of pixels; ε is a small constant, and its value is selected as 0.001 × L 2 ; while L is the dynamic range of the image D. It should be pointed out that, taking into account the common scale of the small target, the size of the local window is somewhat larger than the value recommended in Ref. [36].
The linear coefficients a k , b k centered at pixel k assumed to be constant in a square window of a radius r (denoted as ω k ) are computed as follows:
a k = δ k 2 δ k 2 + λ w k b k = ( 1 a k ) μ k
where μ k and δ k 2 are the mean and variance of D in ω k , λ is a regularization parameter, and w k is the weighting mentioned in Equation (11).
After computing a k , b k for all windows ω k in the image, the filtering output q i centered at pixel i is given as:
a ¯ i = 1 ω k ω i a k b ¯ i = 1 ω k ω i b k q i = a ¯ i D i + b ¯ i
Here, a ¯ i and b ¯ i are the average coefficients of all windows overlapping i , and ω is the number of pixels in ω i .
The WGIF is applied repeatedly for the input image D with four different sets of parameters r j , λ j j = 1 , 2 , 3 , 4 , and the edge-preserving images with vary smoothing effect are acquired easily, denoted as G j j = 1 , 2 , 3 , 4 , as shown in Figure 1.

2.3. Background Prediction Using Fuzzy Sets

To enhance WGIF’s performance specifically in small target detection, we can utilize prior knowledge from the saliency map S to facilitate the background estimation process. Note that the value of the saliency map S represents the probability of a target; the mapping relationship between the target prior distribution and the WGIF filtered/non-filtered images could be established to generate a background image B. Partially inspired from Ref. [7], the background reconstruction mechanism is heuristically defined by the following rules:
If a pixel’s saliency is low, then the predicted pixel is chosen from less smoothed filtered image.
If a pixel’s saliency is medium, then the predicted pixel is chosen from medium smoothed filtered image.
If a pixel’s saliency is high, then the predicted pixel is chosen from more smoothed filtered image.
Since these are fuzzy terms, we can express the concepts of saliency and smoothing extent by the membership functions. As shown in Figure 2, the triangular and singleton types of membership functions are defined for input saliency and output pixel gray, respectively. A serial of parameters s j ( j = 0 , 1 , 2 , 3 , 4 ) and g j ( j = 0 , 1 , 2 , 3 , 4 ) are applied for the construction of the membership functions. It should be noted that the value of g 0 is set the same as the original image pixel gray, and others g 1 , g 2 , g 3 , g 4 come from WGIF filtered images G j j = 1 , 2 , 3 , 4 obtained by different smooth parameters r j , λ j j = 1 , 2 , 3 , 4 . Considering the processing speed is an important factor for small target detection, the constant output membership function significantly reduces computational requirements. Because we are dealing with constants in the output membership function, the output predicted background image, B i , to any saliency map input, s i , is given by:
B i = μ l o w s i × g 0 + μ m i d 1 s i × g 1 + μ m i d 2 s i × g 2 + μ m i d 3 s i × g 3 + μ h i g h s i × g 4 μ l o w s i + μ m i d 1 s i + μ m i d 2 s i + μ m i d 3 s i + μ h i g h s i
According to Equation (14), the background image B is constructed pixel by pixel from the corresponding smoothed or non-smoothed images. When the pixel lies in a target region with a high saliency value, the background pixel is computed from the corresponding WGIF filtered images with both high radius r and regularization parameter λ , thus giving a more intensive blurring effect to the potential target region. Conversely, when the pixel lies in flat or edge locations, the predicted pixel is obtained from the less smoothed images or retained directly from the original image. Leveraging WGIF’s excellent edge-preserving smoothing performance, the major image structures and the complex edge clutter are retained in the background component, while the small target is clearly removed. The predicted image is displayed in Figure 1.

2.4. Target Detection

A subtraction operation is adopted to separate the target from the complex background:
T = D B
In the subtracted image T, the target becomes evident while background clutter is significantly suppressed. Subsequently, the binarization is carried out by setting the threshold as α T max , where T max is the maximum grayscale of the subtracted image and α is an adjustment parameter of the threshold. Our experiments show that setting α as 0.3~0.5 is sufficiently high for single-target segmentation since the target is very salient and the clutter is well suppressed.

3. Experimental and Analysis

In this section, we validate the effectiveness of the proposed approach using four real IR image sequences. Seq. 1 belongs to the cloudy sky scenario, and Seq.2~Seq.4 [41] are IR image data sets with complex backgrounds. The descriptions of each sequence used for performance evaluation are listed in Table 1. All the experiments were implemented in MATLABR2014a and run on a computer with 8 GB memory and IntelCorei7-8565U CPU. The filter radius and regularization parameters r j , λ j j = 1 , 2 , 3 , 4 in WGIF procedure were set as 4 , 0.1 2 , 8 , 0.2 2 , 12 , 0.4 2 , 16 , 0.8 2 , respectively. The fuzzy set parameters s j j = 0 , 1 , 2 , 3 , 4 related to background prediction were set empirically as 0.4, 0.5, 0.6, 0.7, 0.8. It is worth noting that the choice of parameters in the above discussion appears to be purely ad hoc; we could also adjust these values carefully to achieve a better detection result.

3.1. Evaluation Metrics and Comparison Methods

To evaluate the performance of the proposed method, the signal-to-clutter ratio gain (SCRG), the background suppression factor (BSF), and the contrast gain (CG) are introduced as:
SCRG = SCR out SCR in , BSF = C in C out , CG = CON out CON in
where SCR in , C in , and CON in are the signal-to-clutter ratio (SCR), the standard deviation of the entire background, and the contrast measure of the input image, respectively; SCR out , C out , and CON out denote the corresponding values for the processed image. The definitions of SCR and CON are given as:
SCR = μ t μ b σ b , CON = μ t μ b
where μ t , μ b , and σ b are the gray average of the target, the gray average of the background, and the standard deviation of the local background, respectively. It is noted that the sequence averages of the above-mentioned metrics calculated in this paper are denoted as SCRG ¯ , BSF ¯ , and CG ¯ . The other two important metrics are the detection probability P d and the false alarm rate F a , which are defined as:
P d = number of true detections number of actural targets
F a = number of false detections number of images
In this paper, the local contrast measure (LCM) [20], multiscale patch-based contrast measure (MPCM) [21], tri-layer local contrast measure (TLLCM) [24], local intensity and gradient (LIG) [42], infrared patch image (IPI) [1], partial sum of the tensor nuclear norm (PSTNN) [15], and nonconvex tensor fibered rank approximation (NTFRA) [17] are chosen for comparisons. The parameter settings of all compared methods are listed in Table 2.

3.2. Experimental Results and Analysis

To manifest the process clearly, the detailed intermediate results obtained by the proposed method are shown in Figure 3. These include the fast saliency map E, the SK modulated saliency map S, the WGIF smoothed images G j j = 1 , 2 , 3 , 4 , the predicted image B, and the subtracted image T. By comparing E and S in the same sequence, we can see most of the strong edge responses produced by fast saliency are removed correctly through the SK module. In addition, the different WGIF filtering results demonstrate both the edge-preserving effect and the smoothing of small targets. It is also clearly evident that the small target is nearly eliminated from the reconstructed background image B. Therefore, the small target can be easily detected through the simple subtraction operation.
To evaluate the contribution of the critical components to the proposed method, an ablation experiment is conducted to illustrate the effectiveness of the steering kernel and edge-aware weighting. Table 3 shows that the results obtained by non-SK modified fast saliency and non-weighted guided image filtering are inferior to those of the full-components method for all the sequences evaluated. Therefore, the combination of SK and WGIF significantly reduces complex edges and enhances target detection while suppressing backgrounds.
There are several key parameters such as the filter radii, regularization parameters, and fuzzy set parameters. To achieve better performance with real datasets, it is necessary to adjust these parameters carefully. Six groups of parameters and the chosen set are listed in Table 4. The average SCRG, BSF, and CG corresponding to different parameters are shown in Table 5 for Sequences 1–4. It should be noted that each parameter was tuned while keeping the others fixed, potentially resulting in suboptimal performance. From the results of Groups 1–4, it is observed that the algorithm was relatively insensitive to the variations in filter parameters, as different filter radii and regularization parameters all performed well on the four sequences. Furthermore, Group 5 achieved the best CG values for all the sequences but yielded the worst results for SCRG and BSF. This phenomenon may be attributed to the larger step size between the fuzzy parameters, which enhances object contrast. Therefore, to achieve an overall balanced performance, we adopt the moderate step size of filter parameters and fuzzy set parameters, as shown in Table 4.
For visual comparison, representative images from the four sequences and detection results, including the corresponding 3D gray distributions obtained by different algorithms, are given in Figure 4 and Figure 5. The results of LCM, MPCM, and NTFRA contain the most residual background, indicating that their background suppression capability is inferior to the other methods. LIG and PSTNN show some progress in background suppression, but both exhibit significant responses to strong edges, which is particularly evident in Seq. 2. Although IPI and TLLCM effectively suppress the background and detect the targets, the false alarms nevertheless occur in Seqs. 2, 3, and 4. In contrast, our method extracts the targets correctly and removes almost all background on the four sequences except some faint residual clutter in Seq. 4. These experimental results visually demonstrate that the proposed method has excellent background suppression capability, especially against edge clutters under complex scenes.
The average SCRG, BSF, CG, and processing time of each method are listed in Table 6. The assessment criteria SCRG and BSF evaluate the performance of local background suppression and target enhancement. The CG is adopted because in some cases the other two metrics may be very large or infinite [43]. Higher values for all three criteria indicate better performance. To be more intuitive, the best result is marked in bold. As shown, the proposed method achieves the best SCRG and BSF for all four sequences, indicating superior background suppression performance. Although no contrast stretching transformation is applied to enhance the subtracted image, our method still maintains the largest CG values in Seq. 3 and Seq. 4, suggesting better target enhancement capability. In terms of processing time, the proposed method is not as efficient as the LCM or MPCM, but it possesses significant potential for parallel processing and could be readily accelerated using the multi-thread technique or even GPU implementation.
To demonstrate the robustness of our method comprehensively, the receiver operating characteristic (ROC) curves of different methods are shown in Figure 6. These curves illustrate the tradeoff between the detection probability and the false alarm rate. The results indicate that our method is advantageous over other methods in most cases and achieves the best overall detection performance among the eight methods. However, it is noteworthy that TLLCM achieves an impressive detection performance in Seq. 4. The main reason lies in the more evident contrast between the target and surrounding background in Seq. 4 compared to the other sequences, combined with the TLLCM’s ability to first enhance the target, thereby making the distinction sufficiently obvious. This explains why TLLCM’s performance is better than that of LCM or MPCM in most cases. Additionally, the ROC curves of TLLCM, IPI, and the proposed method are nearly overlapping in Seq. 3, indicating comparable performance among these three methods.

4. Conclusions

This paper proposes an IR small target detection method based on modified fast saliency and WGIF to achieve enhanced detection performance under cluttered background. First, by calculating the SK modified saliency map, a simple yet powerful image prior is introduced to characterize targets, modeled as the target existence probability. Then, the edge-preserving smoothed images are acquired by utilizing the rapid local spatial filter WGIF. Subsequently, the background is estimated based on a mapping relationship between the saliency map and smoothed/original images, constructed via the fuzzy set technique. Finally, the target is easily extracted by subtraction and thresholding. The experimental results demonstrate that the proposed method not only achieves a good detection performance and edge suppression ability but also outperforms other baseline methods in various scenarios.

Author Contributions

Conceptualization, Y.C.; methodology, Y.C.; software, Y.Z.; investigation, T.L.; data curation, Y.Z., G.Z. and X.H.; writing—original draft preparation, Y.C.; and writing—review and editing, X.H. and G.C. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the Natural Science Foundation of Sichuan Province under Grant 2025ZNSFSC0533 and in part by the CAS “Light of West China” Program.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The original contributions presented in this study are included in the article. Further inquiries can be directed to the corresponding author(s).

Conflicts of Interest

The authors declare that they have no conflicts of interest to report regarding the present study.

References

  1. Gao, C.; Meng, D.; Yang, Y.; Wang, Y.; Zhou, X.; Hauptmann, A.G. Infrared Patch-Image Model for Small Target Detection in a Single Image. IEEE Trans. Image Process. 2013, 22, 4996–5009. [Google Scholar] [CrossRef] [PubMed]
  2. Zhu, H.; Ni, H.; Liu, S.; Xu, G.; Deng, L. TNLRS: Target-Aware Non-Local Low-Rank Modeling With Saliency Filtering Regularization for Infrared Small Target Detection. IEEE Trans. Image Process. 2020, 29, 9546–9558. [Google Scholar] [CrossRef]
  3. Li, L.; Li, Z.; Li, Y.; Chen, C.; Yu, J.; Zhang, C. Small Infrared Target Detection Based on Local Difference Adaptive Measure. IEEE Geosci. Remote Sens. Lett. 2020, 17, 1258–1262. [Google Scholar] [CrossRef]
  4. Deshpande, S.D.; Er, M.H.; Venkateswarlu, R.; Chan, P.; Drummond, O.E. Max-mean and max-median filters for detection of small targets. In Proceedings of the SPIE’s International Symposium on Optical Science, Engineering, and Instrumentation, Denver, CO, USA, 18–23 July 1999; Volume 3809, pp. 74–84. [Google Scholar]
  5. Cao, Y.; Liu, R.; Yang, J. Small Target Detection Using Two-Dimensional Least Mean Square (TDLMS) Filter Based on Neighborhood Analysis. Int. J. Infrared Millim. Waves 2008, 29, 188–200. [Google Scholar] [CrossRef]
  6. Bae, T.-W.; Zhang, F.; Kweon, I.-S. Edge directional 2D LMS filter for infrared small target detection. Infrared Phys. Technol. 2012, 55, 137–145. [Google Scholar] [CrossRef]
  7. Bae, T.-W.; Sohng, K.-I. Small Target Detection Using Bilateral Filter Based on Edge Component. J. Infrared Millim. Terahertz Waves 2010, 31, 735–743. [Google Scholar] [CrossRef]
  8. Tom, V.T.; Peli, T.; Leung, M.; Bondaryk, J.E. Morphology-based algorithm for point target detection in infrared backgrounds. In Proceedings of the Signal and Data Processing of Small Targets, Orlando, FL, USA, 22 October 1993; Volume 1954, pp. 2–11. [Google Scholar]
  9. Bai, X.; Zhou, F. Analysis of new top-hat transformation and the application for infrared dim small target detection. Pattern Recognit. 2010, 43, 2145–2156. [Google Scholar] [CrossRef]
  10. Zhu, H.; Zhang, J.; Xu, G.; Deng, L. Balanced Ring Top-Hat Transformation for Infrared Small-Target Detection With Guided Filter Kernel. IEEE Trans. Aerosp. Electron. Syst. 2020, 56, 3892–3903. [Google Scholar] [CrossRef]
  11. Wang, C.; Wang, L. Multidirectional Ring Top-Hat Transformation for Infrared Small Target Detection. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2021, 14, 8077–8088. [Google Scholar] [CrossRef]
  12. Deng, L.; Xu, G.; Zhang, J.; Zhu, H. Entropy-Driven Morphological Top-Hat Transformation for Infrared Small Target Detection. IEEE Trans. Aerosp. Electron. Syst. 2022, 58, 962–975. [Google Scholar] [CrossRef]
  13. Dai, Y.; Wu, Y.; Song, Y. Infrared small target and background separation via column-wise weighted robust principal component analysis. Infrared Phys. Technol. 2016, 77, 421–430. [Google Scholar] [CrossRef]
  14. Zhang, L.; Peng, L.; Zhang, T.; Cao, S.; Peng, Z. Infrared Small Target Detection via Non-Convex Rank Approximation Minimization Joint l(2,1) Norm. Remote Sens. 2018, 10, 1821. [Google Scholar] [CrossRef]
  15. Zhang, L.; Peng, Z. Infrared Small Target Detection Based on Partial Sum of the Tensor Nuclear Norm. Remote Sens. 2019, 11, 382. [Google Scholar] [CrossRef]
  16. Zhang, X.; Ding, Q.; Luo, H.; Hui, B.; Chang, Z.; Zhang, J. Infrared small target detection based on an image-patch tensor model. Infrared Phys. Technol. 2019, 99, 55–63. [Google Scholar] [CrossRef]
  17. Kong, X.; Yang, C.; Cao, S.; Li, C.; Peng, Z. Infrared Small Target Detection via Nonconvex Tensor Fibered Rank Approximation. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–21. [Google Scholar] [CrossRef]
  18. Zhang, P.; Zhang, L.; Wang, X.; Shen, F.; Pu, T.; Fei, C. Edge and Corner Awareness-Based Spatial–Temporal Tensor Model for Infrared Small-Target Detection. IEEE Trans. Geosci. Remote Sens. 2021, 59, 10708–10724. [Google Scholar] [CrossRef]
  19. Huang, S.; Liu, Y.; He, Y.; Zhang, T.; Peng, Z. Structure-Adaptive Clutter Suppression for Infrared Small Target Detection: Chain-Growth Filtering. Remote Sens. 2020, 12, 47. [Google Scholar] [CrossRef]
  20. Chen, C.L.P.; Li, H.; Wei, Y.; Xia, T.; Tang, Y.Y. A Local Contrast Method for Small Infrared Target Detection. IEEE Trans. Geosci. Remote Sens. 2014, 52, 574–581. [Google Scholar] [CrossRef]
  21. Wei, Y.; You, X.; Li, H. Multiscale patch-based contrast measure for small infrared target detection. Pattern Recognit. 2016, 58, 216–226. [Google Scholar] [CrossRef]
  22. Han, J.; Liang, K.; Zhou, B.; Zhu, X.; Zhao, J.; Zhao, L. Infrared Small Target Detection Utilizing the Multiscale Relative Local Contrast Measure. IEEE Geosci. Remote Sens. Lett. 2018, 15, 612–616. [Google Scholar] [CrossRef]
  23. Bai, X.; Bi, Y. Derivative Entropy-Based Contrast Measure for Infrared Small-Target Detection. IEEE Trans. Geosci. Remote Sens. 2018, 56, 2452–2466. [Google Scholar] [CrossRef]
  24. Han, J.; Moradi, S.; Faramarzi, I.; Liu, C.; Zhang, H.; Zhao, Q. A Local Contrast Method for Infrared Small-Target Detection Utilizing a Tri-Layer Window. IEEE Geosci. Remote Sens. Lett. 2020, 17, 1822–1826. [Google Scholar] [CrossRef]
  25. Wu, L.; Ma, Y.; Fan, F.; Wu, M.; Huang, J. A Double-Neighborhood Gradient Method for Infrared Small Target Detection. IEEE Geosci. Remote Sens. Lett. 2021, 18, 1476–1480. [Google Scholar] [CrossRef]
  26. Li, Z.; Liao, S.; Zhao, T. Infrared Dim and Small Target Detection Based on Strengthened Robust Local Contrast Measure. IEEE Geosci. Remote Sens. Lett. 2022, 19, 7506005. [Google Scholar] [CrossRef]
  27. Qiu, Z.; Ma, Y.; Fan, F.; Huang, J.; Wu, M. Adaptive Scale Patch-Based Contrast Measure for Dim and Small Infrared Target Detection. IEEE Geosci. Remote Sens. Lett. 2022, 19, 7000305. [Google Scholar] [CrossRef]
  28. Zhang, C.; He, Y.; Tang, Q.; Chen, Z.; Mu, T. Infrared Small Target Detection via Interpatch Correlation Enhancement and Joint Local Visual Saliency Prior. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5001314. [Google Scholar] [CrossRef]
  29. Wang, H.; Zhou, L.; Wang, L. Miss Detection vs. False Alarm: Adversarial Learning for Small Object Segmentation in Infrared Images. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Republic of Korea, 27 October–2 November 2019; pp. 8508–8517. [Google Scholar]
  30. Hou, Q.; Wang, Z.; Tan, F.; Zhao, Y.; Zheng, H.; Zhang, W. RISTDnet: Robust Infrared Small Target Detection Network. IEEE Geosci. Remote Sens. Lett. 2022, 19, 7000805. [Google Scholar] [CrossRef]
  31. Zhao, B.; Wang, C.; Fu, Q.; Han, Z. A Novel Pattern for Infrared Small Target Detection With Generative Adversarial Network. IEEE Trans. Geosci. Remote Sens. 2021, 59, 4481–4492. [Google Scholar] [CrossRef]
  32. Dai, Y.; Wu, Y.; Zhou, F.; Barnard, K. Attentional Local Contrast Networks for Infrared Small Target Detection. IEEE Trans. Geosci. Remote Sens. 2021, 59, 9813–9824. [Google Scholar] [CrossRef]
  33. Bai, Y.; Li, R.; Gou, S.; Zhang, C.; Chen, Y.; Zheng, Z. Cross-Connected Bidirectional Pyramid Network for Infrared Small-Dim Target Detection. IEEE Geosci. Remote Sens. Lett. 2022, 19, 7506405. [Google Scholar] [CrossRef]
  34. Zhao, M.; Li, W.; Li, L.; Hu, J.; Ma, P.; Tao, R. Single-Frame Infrared Small-Target Detection: A survey. IEEE Geosci. Remote Sens. Mag. 2022, 10, 87–119. [Google Scholar] [CrossRef]
  35. Qi, S.; Xu, G.; Mou, Z.; Huang, D.; Zheng, X. A fast-saliency method for real-time infrared small target detection. Infrared Phys. Technol. 2016, 77, 440–450. [Google Scholar] [CrossRef]
  36. Li, Z.; Zheng, J.; Zhu, Z.; Yao, W.; Wu, S. Weighted Guided Image Filtering. IEEE Trans. Image Process. 2015, 24, 120–129. [Google Scholar]
  37. Liu, Y.; Peng, Z. Infrared Small Target Detection Based on Resampling-Guided Image Model. IEEE Geosci. Remote Sens. Lett. 2022, 19, 7002405. [Google Scholar] [CrossRef]
  38. Takeda, H.; Farsiu, S.; Milanfar, P. Kernel regression for image processing and reconstruction. IEEE Trans. Image Process. 2007, 16, 349–366. [Google Scholar] [CrossRef]
  39. Li, Y.; Zhang, Y. Robust infrared small target detection using local steering kernel reconstruction. Pattern Recognit. 2018, 77, 113–125. [Google Scholar] [CrossRef]
  40. He, K.; Sun, J.; Tang, X. Guided Image Filtering. IEEE Trans. Pattern Anal. Mach. Intell. 2013, 35, 1397–1409. [Google Scholar] [CrossRef]
  41. A Dataset for Dim-Small Target Detection and Tracking of Aircraft in Infrared Image Sequences. Available online: http://www.csdata.org/p/387/1 (accessed on 23 June 2020).
  42. Zhang, H.; Zhang, L.; Yuan, D.; Chen, H. Infrared small target detection based on local intensity and gradient properties. Infrared Phys. Technol. 2018, 89, 88–96. [Google Scholar] [CrossRef]
  43. Gao, C.; Wang, L.; Xiao, Y.; Zhao, Q.; Meng, D. Infrared small-dim target detection based on Markov random field guided noise modeling. Pattern Recognit. 2018, 76, 463–475. [Google Scholar] [CrossRef]
Figure 1. Illustration of the proposed method (targets are shown in red rectangles).
Figure 1. Illustration of the proposed method (targets are shown in red rectangles).
Sensors 25 04405 g001
Figure 2. Input (a) and output (b) membership functions for background prediction.
Figure 2. Input (a) and output (b) membership functions for background prediction.
Sensors 25 04405 g002
Figure 3. The results of the fast saliency map E, the SK modulated saliency map S, the WGIF smoothed images Gj, the predicted image B, and the subtracted image T on four real IR sequences (red rectangle marks a true target in the original images).
Figure 3. The results of the fast saliency map E, the SK modulated saliency map S, the WGIF smoothed images Gj, the predicted image B, and the subtracted image T on four real IR sequences (red rectangle marks a true target in the original images).
Sensors 25 04405 g003
Figure 4. Original images and experimental results of different methods on four real IR sequences (red rectangle marks a true target, and green ellipse denotes a false alarm location).
Figure 4. Original images and experimental results of different methods on four real IR sequences (red rectangle marks a true target, and green ellipse denotes a false alarm location).
Sensors 25 04405 g004
Figure 5. Comparison of 3D gray distributions by different methods on four real IR sequences (red rectangle marks a true target).
Figure 5. Comparison of 3D gray distributions by different methods on four real IR sequences (red rectangle marks a true target).
Sensors 25 04405 g005
Figure 6. ROC curves of four sequences. (a) Seq. 1. (b) Seq. 2. (c) Seq. 3. (d) Seq. 4.
Figure 6. ROC curves of four sequences. (a) Seq. 1. (b) Seq. 2. (c) Seq. 3. (d) Seq. 4.
Sensors 25 04405 g006
Table 1. Description of four sequences.
Table 1. Description of four sequences.
SequenceSize/PixelsLength/FramesTarget DescriptionBackground Description
Seq.1256 × 20030Single, relatively large, moving along cloud edgesHeavy cloud sky
Seq.2256 × 25680Single, tiny, low nonlocal contrast Complex road and forest
Seq.3256 × 25680Single, tiny, varying sizeMuch target-like clutter
Seq.4256 × 25680Single, a little long strip, varying sizeMountain and artificial structures
Table 2. Parameters of 8 methods.
Table 2. Parameters of 8 methods.
No.MethodParameter Settings
1LCMLargest scale S = 4 size: 3 × 3, 5 × 5, 7 × 7, 9 × 9
2MPCMMean filter size: 3 × 3, N = 3,5,7,9
3TLLCMCore layer size: 3 × 3, Reserve layer size: 5 × 5, 7 × 7, 9 × 9
4LIGSliding window size: 11 × 11, k = 0.2
5IPIPatch size: 50 × 50, step:10
λ = 1 / min ( n 1 , n 2 ) , ε = 10 7
6PSTNNPatch size: 40 × 40, step:40
λ = 0.7 / min ( n 1 , n 2 ) n 3 , ε = 10 7
7NTFRAPatch size: 40 × 40, step:40
λ = 1 / min ( n 1 , n 2 ) n 3 , β = 0.01 , μ = 200
8ProposedFuzzy set parameters: 0.4,0.5,0.6,0.7,0.8
Filter parameters: 4 , 0.1 2 , 8 , 0.2 2 , 12 , 0.4 2 , 16 , 0.8 2
Table 3. The impact of different components on the performance of the proposed method.
Table 3. The impact of different components on the performance of the proposed method.
MethodsWithout SKWithout WeightingProposed
SCRG ¯


Seq 1InfInfInf
Seq 2InfInfInf
Seq 3InfInfInf
Seq 4InfInfInf
BSF ¯


Seq 1InfInfInf
Seq 29.33924.959Inf
Seq 3InfInfInf
Seq 436.31248.39358.503
CG ¯


Seq 12.3682.7283.198
Seq 20.7681.4551.510
Seq 30.9040.8301.767
Seq 40.9911.1631.488
Table 4. Six groups of experimental parameters and the finally adopted ones.
Table 4. Six groups of experimental parameters and the finally adopted ones.
No.Filter ParametersFuzzy Set Parameters
Group 1 4 , 0.1 2 , 6 , 0.2 2 , 8 , 0.4 2 , 10 , 0.8 2
0.4,0.5,0.6,0.7,0.8
Group 2 4 , 0.1 2 , 10 , 0.2 2 , 16 , 0.4 2 , 22 , 0.8 2
0.4,0.5,0.6,0.7,0.8
Group 3 4 , 0.1 2 , 8 , 0.15 2 , 12 , 0.2 2 , 16 , 0.25 2
0.4,0.5,0.6,0.7,0.8
Group 4 4 , 0.1 2 , 8 , 0.5 2 , 12 , 1.0 2 , 16 , 1.5 2 0.4,0.5,0.6,0.7,0.8
Group 5 4 , 0.1 2 , 8 , 0.2 2 , 12 , 0.4 2 , 16 , 0.8 2 0.1,0.3,0.5,0.7,0.9
Group 6 4 , 0.1 2 , 8 , 0.2 2 , 12 , 0.4 2 , 16 , 0.8 2 0.5,0.55,0.6,0.65,0.7
Adopted 4 , 0.1 2 , 8 , 0.2 2 , 12 , 0.4 2 , 16 , 0.8 2 0.4,0.5,0.6,0.7,0.8
Table 5. Average SCRG, BSF, and CG by different parameters of the proposed method.
Table 5. Average SCRG, BSF, and CG by different parameters of the proposed method.
ParametersGroup 1Group 2Group 3Group 4Group 5Group 6Adopted
SCRG ¯


Seq 1InfInfInfInfInfInfInf
Seq 2InfInfInfInf173.22InfInf
Seq 3InfInfInfInfInfInfInf
Seq 4InfInfInfInf106.05InfInf
BSF ¯


Seq 1InfInfInfInfInfInfInf
Seq 2114.7390.617129.0867.43221.977InfInf
Seq 3InfInfInfInf107.01InfInf
Seq 4162.5149.1445.647380.5138.99Inf58.503
CG ¯


Seq 13.1673.2353.0062.9673.3463.0993.198
Seq 21.4691.4931.4861.5181.6811.4181.510
Seq 31.6911.7391.7171.7871.8671.6621.767
Seq 41.3111.4511.5411.4681.5661.4071.488
Table 6. Average SCRG, BSF, CG, and processing time obtained by 8 methods on four sequences.
Table 6. Average SCRG, BSF, CG, and processing time obtained by 8 methods on four sequences.
MethodsLCMMPCMTLLCMLIGIPIPSTNNNTFRAProposed
SCRG ¯


Seq 11.5631.656Inf52.3259.130Inf8.875Inf
Seq 21.6381.49617.972131.880InfInfInfInf
Seq 30.5332.131Inf23.6497.656Inf4.330Inf
Seq 40.3201.6954.39438.3845.106Inf6.140Inf
BSF ¯


Seq 10.7062.13513.13210.07016.186Inf10.875Inf
Seq 20.8894.0978.1513.89514.7713.3231.309Inf
Seq 31.6126.90429.25010.14134.61515.6664.999Inf
Seq 41.9847.81842.53512.85019.53910.3912.64558.503
CG ¯


Seq 13.3061.1722.2642.8572.5913.3284.3203.198
Seq 23.9720.8881.8371.2191.2351.4821.9111.510
Seq 31.4210.6611.3091.4641.2611.5751.7291.767
Seq 41.4750.7751.3321.3061.1611.4081.4071.488
Time ¯ /s Seq 10.07860.08372.0411.2706.3110.06341.2110.715
Seq 20.09020.09012.7731.6478.5530.2821.8270.894
Seq 30.08770.08992.8641.6679.0070.2291.9020.968
Seq 40.08990.09252.7521.67410.0550.2531.7930.901
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Cui, Y.; Lei, T.; Chen, G.; Zhang, Y.; Zhang, G.; Hao, X. Infrared Small Target Detection via Modified Fast Saliency and Weighted Guided Image Filtering. Sensors 2025, 25, 4405. https://doi.org/10.3390/s25144405

AMA Style

Cui Y, Lei T, Chen G, Zhang Y, Zhang G, Hao X. Infrared Small Target Detection via Modified Fast Saliency and Weighted Guided Image Filtering. Sensors. 2025; 25(14):4405. https://doi.org/10.3390/s25144405

Chicago/Turabian Style

Cui, Yi, Tao Lei, Guiting Chen, Yunjing Zhang, Gang Zhang, and Xuying Hao. 2025. "Infrared Small Target Detection via Modified Fast Saliency and Weighted Guided Image Filtering" Sensors 25, no. 14: 4405. https://doi.org/10.3390/s25144405

APA Style

Cui, Y., Lei, T., Chen, G., Zhang, Y., Zhang, G., & Hao, X. (2025). Infrared Small Target Detection via Modified Fast Saliency and Weighted Guided Image Filtering. Sensors, 25(14), 4405. https://doi.org/10.3390/s25144405

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop