Next Article in Journal
Obstacle Avoidance Drone by Deep Reinforcement Learning and Its Racing with Human Pilot
Previous Article in Journal
Systematic Literature Review of Predictive Analysis Tools in Higher Education
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Small Infrared Target Detection via a Mexican-Hat Distribution

College of Computer Science and Technology, Harbin Engineering University, Harbin 150001, China
*
Author to whom correspondence should be addressed.
Appl. Sci. 2019, 9(24), 5570; https://doi.org/10.3390/app9245570
Submission received: 27 November 2019 / Revised: 10 December 2019 / Accepted: 13 December 2019 / Published: 17 December 2019

Abstract

:
Although infrared small target detection has been broadly used in airborne early warning, infrared guidance, surveillance and tracking, it is still an open issue due to the low signal-to-noise ratio, less texture information, background clutters, and so on. Aiming to detect a small target in an infrared image with complex background clutters, this paper carefully studies the characteristics of a target in an IR image filtered by the difference of Gaussian filter, concluding that the intensity of the adjacent region around a small infrared target roughly has a Mexican-hat distribution. Based on such a conclusion, a raw infrared image is sequentially processed with the modified top-hat transformation and the difference of Gaussian filter. Then, the adjacent region around each pixel in the processed image is radially divided into three sub-regions. Next, the pixels that distribute as the Mexican-hat are determined as the candidates of targets. Finally, a real small target is segmented out by locating the pixel with the maximum intensity. Our experimental results on both real-world and synthetic infrared images show that the proposed method is so effective in enhancing small targets that target detection gets very easy. Our method achieves true detection rates of 0.9900 and 0.9688 for sequence 1 and sequence 2, respectively, and the false detection rates of 0.0100 and 0 for those two sequences, which are superior over both conventional detectors and state-of-the-art detectors. Moreover, our method runs at 1.8527 and 0.8690 s per frame for sequence 1 and sequence 2, respectively, which is faster than RLCM, LIG, Max–Median, Max–Mean.

1. Introduction

Small infrared (IR) target detection has been widely used in airborne early warning, IR guidance, surveillance and tracking, and others. Usually, the earlier we detect a small IR target, the more time we get for dealing with it, and thus the more suitable decision can be made. However, factors such as low signal-to-noise ratio (SNR), variable target sizes, variable target intensity, less shape and texture information, blurred edges, and serious background clutters, cause small IR target detection to be a challenging task.
So far, many IR target detection algorithms have been developed ranging from recursive estimation techniques [1] to partial sum of the tensor nuclear norm [2]. Among them, morphology filtering such as the top-hat transform and its variants plays an important role [3,4]. Top-hat transforms are also combined with other techniques such as genetic algorithm [5] to improve the detection performance. Though these methods can detect targets to a certain degree, their performance is greatly degraded under complex background clutters.
Over the last few years, it is widely shared that algorithms derived from visual attention mechanics work well for detecting a small target. The early work in 1998 made intellectuals pour attention into visual attention distribution [6], and then saliency estimation methods corresponding to visual attention became popular in small IR target detection [7]. Wang et al. are the early researchers to use the Difference of Gaussian (DoG) filters to get a saliency map for detecting real targets [8]. However, due to problems like signal interference, a small target does not strictly obey the 2D Gaussian distribution, and thus cannot be successfully detected by merely adopting DoG filters. Further studies on visual attention focus on the relationship between a target and its neighborhoods. The local contrast method [9,10], the multiscale patch-based contrast measure [11], weighted local difference measure [12], human visual mechanism detectors [13,14], variance difference detector [15], entropy-based contrast measure [16], weighted local contrast detector [17], local intensity and gradient detector [18], visual saliency guided detector [19], local difference adaptive measure [20], the adaptive local measurement contrast, and salient region extraction and gradient vector processing [21] are reported successively, demonstrating that the contrast between a target and its neighbor is helpful for small IR target detection.
Though many improved algorithms have been reported, it is still challenging to detect a small IR target in images with low SNR, low contrast, and serious background clutters. This paper presents an effective detection method based on the Mexican-hat distribution. A raw IR image is first processed by the modified top-hat transformation [22] and the DoG filter, getting a filtered IR image. Then, the adjacent region around a pixel of the filtered image is radially divided into three sub-regions. Next, the pixels whose adjacent regions have the Mexican-hat distribution are determined as candidate targets. Finally, a small target is segmented out by locating the brightest pixel. Experimental results show that the adoption of the Mexican-hat distribution benefits our method with higher detection rate, lower false alarm rate, and faster detection speed than existing detectors. Though the existing detector in [15] is resemble to our method, it performs on raw IR images and only compares the relationship between two regions. On contrast, we operate on the filtered image and detect an IR target based on the Mexican-hat distribution, and our experimental results show that such strategy can effectively enhance the small IR target and improve the detection rate.
The rest of the paper is organized as follows. Section 2 analyzes the characteristics of the adjacent region of a target in a DoG-filtered image. Section 3 describes the details of the proposed method. Section 4 presents experimental results along with some analysis. Finally, Section 5 gives some conclusions.

2. Mexican-Hat Distribution of the Adjacent Region around an IR Target

Let I(x, y) be an image. Then its DoG-filtered image, I D o G , is formed by (1) [23]:
I D o G ( x , y ) = G ( x , y , σ 1 ) I ( x , y ) G ( x , y , σ 2 ) I ( x , y )
where σ 1 and σ 2 are the standard deviations of the two Gaussian kernels, ‘*’ is the convolution operation given by (2).
G ( x , y , σ ) = 1 2 π σ 2 e x 2 + y 2 2 σ 2
After careful observation, we find the pixel intensity of the adjacent region around an IR target shows a regular bright-dark-bright pattern along the radial direction. Specifically, Figure 1 illustrates such bright-dark-bright pattern. The center pixel of a target patch is the brightest, and then the intensity decreases gradually along the radial direction until reaching the lowest value. After that, the intensity increases again. As shown in Figure 1c, the profile of such bright-dark-bright pattern is like a Mexican hat, and thus we call it Mexican-hat distribution in this paper.
Based on the above observations, we propose to detect small targets in IR images via the Mexican-hat distribution. As shown in Figure 2, we first divide the adjacent region of a small target into three sub-regions along the radial direction that are respectively termed as R 0 , R 1 , and R 2 . For a small target positioned at (r, c), R 0 is a (2L + 1) × (2L + 1) square image region centered at (r, c), and R 1 is one-pixel square border just outside R 0 , and R 2 is one-pixel square border without four corner pixels outside R 1 . According to the Mexican-hat distribution, we know that the intensity relationship between the three sub-regions of a target roughly meets the bright-dark-bright pattern, and thus have the following rule:
Rule 1: If the mean intensity of R 0 is both larger than R 1 and R 2 , and the mean intensity of R 1 is smaller than R 2 , then the center of R 0 can be a candidate target center.

3. The Proposed Method

As shown in Figure 3, the proposed four-step method for detecting a small IR target is based on the Mexican-hat distribution. To increase the contrast between small targets and the background, the modified top-hat transformation [22] is first applied to a raw IR image, followed by DoG filtering. Then, the Mexican-hat-distribution based Rule 1 is applied to the DoG image, getting candidate targets. Finally, a small IR target is detected by locating the brightest pixel.
Here, to deal with the variable size of small IR targets, the following iteration strategy is adopted when applying Rule 1 to a DoG image. First, let the half-width, L, of region R 0 ranges from 1 to 5, implying the size of region R 0 varies from 3 × 3 to 11 × 11 that is consistent to the recognition that a small target usually occupies less than 80 pixels [24]. Then for each L, Rule 1 is applied to the adjacent of each pixel of the DoG-filtered image. Those pixels that meet Rule 1 are labeled as candidates, and their intensity is emphasized by the mean intensity of R 1 . Meanwhile, the corresponding size of R 0 , i.e., (2L + 1) × (2L + 1) are recorded as the target size.
Algorithm 1 presents the details for finding candidate targets as well as their sizes. Considering small IR targets are usually brighter than the background, Rule 1 is merely applied to the patches whose center pixel is brighter than t h = μ 0 + σ 0 with μ 0 and σ 0 being the mean intensity and the standard variance of the DoG-filtered image.
Algorithm 1 Candidate target detection.
1:
Input: I D o G —the DoG-filtered image
   H,W—the height and the width of I D o G
   th—threshold
2:
Output: I c a n —the image with enhanced candidate targets
    ( r c a n , c c a n )—the coordinates of candidates
    s—the size of candidates
    inten—the mean intensity of candidates
3:
for r = 1 to H do
4:
for c = 1 to W do
5:
  if I D o G (r, c) > th then
6:
   for L = 1 to 5 do
7:
    Locate sub-regions R 0 , R 1 and R 2 according to Figure 2;
8:
     m i = mean intensity of R i , with i = 0 , 1 and 2.
9:
    if m 0 > m 1 , m 0 > m 2 , and m 1 < m 2 then
10:
     ( r c a n , c c a n ) = (r, c)
     s( r c a n , c c a n ) = ( 2 L + 1 ) × ( 2 L + 1 )
     inten( r c a n , c c a n ) = m 0
      I c a n ( r c a n , c c a n ) = I D o G ( r c a n , c c a n ) − m 1
11:
    else
12:
     break;
13:
    end if
14:
   end for
15:
  end if
16:
end for
17:
end for
Next, to further remove false targets, the statistics of sub-regions R 0 , R 1 and R 2 are calculated. For real small IR targets, their intensity might not strictly meet the Mexican-hat distribution. For example, the first row of Figure 4 shows that the center pixel of a target is not the brightest, and the second row shows that nearly half of pixels in region R 1 and in region R 2 don’t meet the Mexican-hat distribution. Therefore, for a candidate target located at (r, c), (3) is adopted to calculate the ratio of the number of pixels in R 1 being darker than the mean intensity of R 0 to the total pixels of R 1 .
Rt 10 ( r , c ) = C a r d ( ( x , y ) R 1 | I ( x , y ) < m 0 ) C a r d ( R 1 )
where m 0 is the mean intensity of R 0 , and C a r d ( . ) represents the cardinality of a set.
Then the pixel by pixel differences between R 1 and R 2 are calculated along the row direction and the column direction, respectively, followed by the computation of the ratio of the number of pixels in R 1 being darker than the corresponding pixels in R 2 to the total pixels with (4) and (5).
Rt 12 ( r , c ) = N 12 C a r d ( R 1 ) + 4
N 12 = Card ( ( x l , y ) R 1 | I D o G ( x l 1 , y ) > I D o G ( x l , y ) ) + Card ( ( x r , y ) R 1 | I D o G ( x r + 1 , y ) > I D o G ( x r , y ) ) + Card ( ( x , y u ) R 1 | I D o G ( x , y u 1 ) > I D o G ( x , y u ) ) + Card ( ( x , y d ) R 1 | I D o G ( x , y d + 1 ) > I D o G ( x , y d ) )
where ( x L , y ) and ( x R , y ) represent pixels of the first and the last columns in R1, and ( x , y U ) and ( x , y D ) are the pixels of the first and the last rows in R 1 . Equation (5) indicates that each corner element in R 1 has been compared two times, so the total number of pixels in (4) is given by Card( R 1 ) + 4 rather than Card( R 1 ).
Having the ratios of Rt 10 and Rt 12 , we compare them with two predetermined thresholds th 10 and th 12 , and use Rule 2 to detect a true small target. Finally, Rule 3 is applied to both target regions and non-target regions to further enhance targets while suppressing non-targets.
Rule 2: If Rt 10 > th 10 and Rt 12 > th 12 , then the corresponding pixel position (r, c) is the center of a small target.
Rule 3: If (r, c) is the center of a small target, then use (6) to enhance it; otherwise use (7) to suppress it.
I F i n a l ( r , c ) = max ( Rt 10 , Rt 12 ) × I ( r , c ) × I D o G ( r , c )
I F i n a l ( r , c ) = ( min ( Rt 10 , Rt 12 ) ) 2 × I ( r , c ) × I D o G ( r , c )
where I F i n a l ( r , c ) , I ( r , c ) and I D o G ( r , c ) represent the final enhanced IR image, the raw IR image, and the DoG-filtered image, max(.) and min(.) are the maximum and the minimum functions.
Since small targets in both a raw IR image and DoG-filtered image are usually brighter than their surroundings, the term I ( r , c ) × I D o G ( r , c ) in (6) and (7) can enhance targets while suppressing backgrounds. Moreover, since both Rt 10 and Rt 12 are positive numbers smaller than 1, the term max ( Rt 10 , Rt 12 ) in (6) is always bigger than the term ( min ( Rt 10 , Rt 12 ) ) 2 in (7). Therefore, by applying (6) and (7) to targets and non-targets, we can get a final image with salient targets. As a result, the target is easily detected by assigning the pixel with the maximum intensity as the target center. Algorithm 2 gives the details of our method for detecting small IR target.
Algorithm 2 Our method for detecting small IR target.
1:
Input: I, I D o G —raw IR image and the corresponding DoG image
     th 10 , th 12 —thresholds
    H,W—the height and the width of I
2:
Output: ( r t a r , c t a r )—target center
3:
for r = 1 to H do
4:
for c = 1 to W do
5:
  if ( r , c ) is the center of a candidate target then
6:
   compute R 10 and R 12 with (3)–(5).
7:
   if R 10 > th 10 and R 12 > th 12 then
8:
    use (6) to get I F i n a l ( r , c )
9:
   else
10:
    use (7) to get I F i n a l ( r , c )
11:
   end if
12:
  end if
13:
end for
14:
end for
15:
[ r t a r c t a r ] = max 1 r H , 1 c W ( I F i n a l ( r , c ) )

4. Experimental Results and Analysis

The proposed method for detecting a small IR target has been implemented in MATLAB using a PC with 2.6 GHz Intel Core i7 processor and 8 GB RAM. To evaluate the proposed method, different quantitative and qualitative evaluations have been carried out on IR images with complex backgrounds. The test data consists of 100 synthetic IR images with resolution of 584 × 468 pixels (termed as Seq.1) and 32 real-world IR images with resolution of 240 × 320 pixels (termed as Seq.2). Each image of Seq.1 has only one target. So does Seq.2. Here, the 100 synthetic images are made by adding targets with a Gaussian distribution to real-world cloud backgrounds.

4.1. Influences of the Modified Top-Hat Transformation and the DoG

This experiment adopts Seq.2 that has low target-background contrast to evaluate the influence of the modified top-hat transformation. According to [22] as well as our experimental results, we set the parameters of the modified top-hat transformation as L t o p - h a t = 5 , δ = 1, α = 1, β = 1.2, γ = 1.2, and ε = 5. Here L t o p - h a t is the window size. α , β and γ are the coefficients of target enhancement operation. For the two thresholds of our method, we set t h 10 = 0.9 and t h 12 = 0.5 . Figure 5, Figure 6 and Figure 7 show some results with and without modified top-hat transformation. The raw IR images of these figures include a small target located at the edge of clouds, a small target located at the sky background without clouds, and a small target near the edge of clouds, respectively. We can see that the modified top-hat transformation benefits to increasing the contrast between the target and the background. Therefore, the targets with low contrast can be detected easier with such a transformation.

4.2. Step-By-Step Results

As shown in Figure 3, our proposed method consists of four major steps. This experiment aims to present the influence of each step on the detection. Figure 8 and Figure 9 show some representative results. Generally speaking, the proposed method is able to well enhance the IR targets under different backgrounds and can suppress much of the background clutters. Specifically, we can see that by applying the modified top-hat transformation, dark sky backgrounds are darkened further, and cloud clutters become smoother (Figure 8b and Figure 9b). Then, with DoG filtering, a small IR target is emphasized, while the background is greatly suppressed, and thus the target becomes more prominent (Figure 8c and Figure 9c). Next, by using the Mexican-hat distribution, candidate targets including the real target as well as some cloud edges are enhanced (Figure 8d and Figure 9d). And then by using Rule 2 and Rule 3, the real target becomes the most prominent one (Figure 8e,f and Figure 9e,f). As a result, the target can be easily detected by locating the brightest pixel.
Figure 8c shows that DoG filtering is good at dealing with a synthetic target whose intensity strictly meets the Gaussian distribution. Generally speaking in this case the DoG itself is enough to detect the real target. But Figure 9c indicates that the DoG cannot well deal with real targets which may not have a strict Gaussian distribution. In this case, background clutters like cloud edges may have a similar intensity to a small target after DoG, and thus the simple DoG cannot detect targets correctly. From Figure 9f we can see that in the case of real targets our method still performs well. This is due to the rules based on the Mexican-hat distribution play an important role in enhancing real targets.

4.3. Comparison with Other Methods

This experiment compares the proposed method with other nine popular small IR target detectors, i.e. Relative Local Contrast Measure (RLCM) [25], Patch-based Contrast Measure (PCM) [11], Variance difference (Var-diff) [15], Absolute Average Gray Difference (AAGD) [25], Laplacian of Gaussian (LoG) [26], Local Intensity and Gradient (LIG) [18], Max-median [27], Max–Mean [27] and the traditional Top-Hat filter. Some of them are traditional methods like LoG, Max-median, Max–Mean, and the traditional Top-Hat filter, and others are state-of-the-art methods like RLCM, PCM, Var-diff, AAGD, and LIG. Most parameters of the nine methods are the same as the original references, except the following parameters. The standard deviation of the Gaussian kernel is 0.5 in LoG, and the sliding window size of LIG is 19 × 19. The patch sizes of Max-median and Max–Mean is 5. The structure element is a rectangle of 3 × 3 in the Top-Hat filter. We determine these parameters to make the corresponding method perform best on both Seq.1 and Seq.2.
Figure 10 and Figure 11 visually illustrate the detection results corresponding to each method. The final detected target is circled in yellow. An image without any yellow circle means the method fails to detect any target. Here zoomed in images have been presented to see the target clearly.
From the 3D intensity plots shown in Figure 10 and Figure 11, we can see the target enhanced by our Mexican-hat distribution based method is more prominent than other compared methods, leading our method a satisfied detection performance. Particularly, RLCM can enhance the target successfully, but as shown in Figure 10 and Figure 11, it cannot suppress the high bright backgrounds successfully. For the methods of LoG, Max-median, Max-mean, and Top-Hat, they sometimes fail in suppressing clutters, leading to a false detection or failing to detect any target. For PCM, Var-diff, AAGD, and LIG, they apt to fail to detect any target in the images with complex backgrounds. In fact, for the simple methods such as Max-median and Max-mean, they seem to be more effective on a synthetic target, because of the non-strict Gaussian distribution of real small IR targets as well as the low contrast between target and background.
Besides visual comparisons, quantitative metrics have also been evaluated. We choose the Signal to Clutter Ratio Gain (SCRG) [28], Background Suppression Factor(BSF) [28], the true detection rate P d , the false detection rate P f [29], and time cost as the quantitative metrics. The results are listed in Table 1 and Table 2.
Table 1 and Table 2 show that for both synthetic and real-world targets, our method gets the highest SCRG and BSF, indicating it performs better than other compared methods in terms of both clutter suppression and target enhancement. Though for a synthetic target, P d and P f of our method are slightly worse than the Max-median and the Max-median, they are much better than the other seven methods. Moreover, among all of the compared methods, our method gets the highest P d and the lowest and P f on Seq.2 that has a real-world target. In terms of time cost, our method has a medium performance. The average processing time of our method is about 1.86 s for an image of 468 by 584 pixels, and 0.87 s for an image of 240 by 320 pixels. Our method is slower than PCM, Var-diff, AAGD, LoG, and Top-Hat, but faster than LIG and Max-median, and much faster than RLCM.

4.4. Multi-Target Detection Results

Although our method is originally designed for detecting single target, our method can also be directly applied to multi-target detection. To evaluate the performance of our method on multi-target detection, two sequences respectively derived from Seq.1 and Seq.2 are adopted. Two Gaussian-distributed synthetic small targets are added to the first thirty images of Seq.1 (Seq.2) and we named it as Seq.3 (Seq.4). Here, we only compare our method to LoG, LIG, Max-median, Max-mean, Top-Hat due to their good performance on Seq.1 and Seq.2. Table 3 and Table 4 list the quantitative results, while Figure 12 and Figure 13 show some detection results visually. Once again, we can see from Table 3 and Figure 12 that our method achieves the best P d and P f . However, our detector performs worse than LoG in the case of complex backgrounds as presented in Table 4 and Figure 13. Form Figure 12 and Figure 13, we can also see that our method still has strong ability for removing background clutters while reserving targets.
We also compare our method to those detectors mentioned in Section 4.3. As is shown in Figure 12, compared to other methods, our method can detect multi targets more accurately due to its stronger ability than RLCM, LoG, LIG, Max- median, Max-mean, and Top-hat for removing background clutters while reserving targets.

5. Conclusions

In this paper, we have proposed a new method for detecting small IR targets. Our method is based on that the adjacent region of a small target in a DoG-filtered image roughly holds the Mexican-hat distribution. Our experimental results on both real-world and synthetic IR images show that our method is quite effective in enhancing small IR targets while suppressing background clutters. In terms of SCRG, BSF, P d , and P f , our method outperforms both the traditional and state-of-the-art methods. Moreover, it runs faster than the RLCM, the LIG, the Max–median, and the Max–mean.
Our experimental results show that our method performs rather well in single target detection and can also be directly used for detecting multi-targets in simple background. However, such direct use degrades the detection performance when the background is very complicated. Thus, our future work will focus on detecting multi-targets in complex background via such a method.

Author Contributions

Conceptualization, Y.Z. (Yubo Zhang) and L.Z.; Data curation, Y.Z. (Yubo Zhang) and L.Z.; Formal analysis, Y.Z. (Yubo Zhang) and L.Z.; Funding acquisition, L.Z.; Investigation, Y.Z. (Yubo Zhang); Methodology, Y.Z. (Yubo Zhang); Project administration, L.Z.; Resources, L.Z.; Software, Y.Z. (Yubo Zhang); Supervision, L.Z.; Visualization, Y.Z. (Yubo Zhang); Writing—original draft, Y.Z. (Yubo Zhang); Writing—review & editing, Y.Z. (Yubo Zhang), L.Z. and Y.Z. (Yanbo Zhang).

Funding

This work has been supported by the National Natural Science Foundation of China (No. 61771155).

Conflicts of Interest

The authors declare no conflict of interest and there is no copyright issue in all of figures.

References

  1. Soni, T.; Zeidler, J.R.; Ku, W.H. Recursive estimation techniques for detection of small objects in infrared image data. In Proceedings of the ICASSP-92: 1992 IEEE International Conference on Acoustics, Speech, and Signal Processing, San Francisco, CA, USA, 23–26 March 1992; Volume 3, pp. 581–584. [Google Scholar]
  2. Zhang, L.; Peng, Z. Infrared Small Target Detection Based on Partial Sum of the Tensor Nuclear Norm. Remote Sens. 2019, 11, 382. [Google Scholar] [CrossRef] [Green Version]
  3. Bai, X. Morphological center operator for enhancing small target obtained by infrared. Opt.-Int. J. Light Electron Opt. 2014, 125, 3697–3701. [Google Scholar] [CrossRef]
  4. He, Q.; Mo, B.; Liu, F.; He, Y.; Liu, S. Small infrared target detection utilizing Local Region Similarity Difference map. Infrared Phys. Technol. 2015, 71, 131–139. [Google Scholar]
  5. Deng, L.; Zhu, H.; Zhou, Q.; Li, Y. Adaptive top-hat filter based on quantum genetic algorithm for infrared small target detection. Multimed. Tools Appl. 2017, 77, 10539–10551. [Google Scholar] [CrossRef]
  6. Itti, L.; Koch, C.; Niebur, E. A model of saliency-based visual attention for rapid scene analysis. IEEE Trans. Pattern Anal. Mach. Intell. 1998, 11, 1254–1259. [Google Scholar] [CrossRef] [Green Version]
  7. Li, W.; Pan, C.; Liu, L.X. Saliency-based automatic target detection inforward looking infrared images. In Proceedings of the 2009 16th IEEE International Conference on Image Processing (ICIP), Cairo, Egypt, 7–10 November 2009; Volume 153, pp. 327–333. [Google Scholar]
  8. Wang, X.; Lv, G.; Xu, L. Infrared dim target detection based on visual attention. Infrared Phys. Technol. 2014, 55, 513–521. [Google Scholar] [CrossRef]
  9. Chen, C.L.P.; Li, H.; Wei, Y.; Xia, T.; Tang, Y.Y. A local contrast method for small infrared target detection. IEEE Trans. Geosci. Remote Sens. 2014, 52, 574–581. [Google Scholar] [CrossRef]
  10. Yang, J.; Gu, Y.; Sun, Z.; Cui, Z. A Small Infrared Target Detection Method Using Adaptive Local Contrast Measurement. In Proceedings of the 2019 IEEE International Instrumentation and Measurement Technology Conference (I2MTC), Auckland, New Zealand, 20–23 May 2019; pp. 1–6. [Google Scholar]
  11. Wei, Y.; You, X.; Li, H. Multiscale patch-based contrast measure for small infrared target detection. Pattern Recognit. 2016, 58, 216–226. [Google Scholar] [CrossRef]
  12. Deng, H.; Sun, X.; Liu, M.; Ye, C.; Zhou, X. Small infrared target detection based on weighted local difference measure. IEEE Trans. Geosci. Remote Sens. 2016, 10, 4204–4214. [Google Scholar] [CrossRef]
  13. Han, J.; Ma, Y.; Huang, J.; Mei, X.; Ma, J. An infrared small target detecting algorithm based on human visual system. IEEE Geosci. Remote Sens. Lett. 2016, 13, 452–456. [Google Scholar] [CrossRef]
  14. Chen, Y.; Song, B.; Wang, D.; Guo, L. An effective infrared small target detection method based on the human visual attention. J. Abbr. 2018, 95, 128–135. [Google Scholar] [CrossRef]
  15. Nasiri, M.; Chehresa, S. Infrared small target enhancement based on variance difference. Infrared Phys. Technol. 2017, 10, 107–119. [Google Scholar] [CrossRef]
  16. Bai, X.; Bi, Y. Derivative entropy-based contrast measure for infrared small target detection. IEEE Trans. Geosci. Remote Sens. 2018, 56, 2452–2466. [Google Scholar] [CrossRef]
  17. Liu, J.; He, Z.; Chen, Z.; Shao, L. Tiny and dim infrared target detection based on weighted local contrast. IEEE Geosci. Remote Sens. Lett. 2018, 15, 1780–1784. [Google Scholar] [CrossRef]
  18. Zhang, H.; Zhang, L.; Yuan, D.; Chen, H. Infrared small target detection based on local intensity and gradient properties. Infrared Phys. Technol. 2018, 89, 88–96. [Google Scholar] [CrossRef]
  19. Yi, X.; Wang, B.; Zhou, H.; Qin, H. Dim and small infrared target fast detection guided by visual saliency. Infrared Phys. Technol. 2019, 97, 6–14. [Google Scholar] [CrossRef]
  20. Li, L.; Li, Z.; Li, Y.; Chen, C.; Yu, J.; Zhang, C. Small Infrared Target Detection Based on Local Difference Adaptive Measure. IEEE Geosci. Remote Sens. Lett. 2019, 1–5. [Google Scholar] [CrossRef]
  21. Zhang, J.; Zhang, B.; Liu, P. Infrared small target detection based on salient region extraction and gradient vector processing. In Proceedings of the 2019 International Conference on Robotics, Intelligent Control and Artificial Intelligence, Shanghai, China, 20–22 September 2019; pp. 422–426. [Google Scholar]
  22. Bai, X.; Zhou, F. Infrared small target enhancement and detection based on modified top-hat transformations. Comput. Electr. Eng. 2010, 36, 1193–1201. [Google Scholar] [CrossRef]
  23. Lowe, D.G. Distinctive image features from scale-invariant keypoints. Int. J. Comput. Vis. 2004, 60, 91–110. [Google Scholar] [CrossRef]
  24. Nilufar, S.; Ray, N.; Zhang, H. Object detection with dog scale-space: A multiple kernel learning approach. J. Abbr. 2012, 21, 3744–3756. [Google Scholar] [CrossRef]
  25. Moradi, S.; Moallem, P.; Sabahi, M.F. A false-alarm aware methodology to develop robust and efficient multi-scale infrared small target detection algorithm. Infrared Phys. Technol. 2018, 89, 387–397. [Google Scholar] [CrossRef]
  26. Kim, S.; Yang, Y.; Lee, J.; Park, Y. Small target detection utilizing robust methods of the human visual system for IRST. J. Infrared Millimeter Terahertz Waves 2009, 30, 994–1011. [Google Scholar] [CrossRef]
  27. Deshpande, S.D.; Er, M.H.; Venkateswarlu, R.; Chan, P. Max-mean and max-median filters for detection of small targets. Signal Data Process. Small Targets 1999, 3809, 74–83. [Google Scholar]
  28. Hilliard, C.I. Selection of a clutter rejection algorithm for real-time target detection from an airborne platform. Signal Data Process. Small Targets 2000, 4048, 74–84. [Google Scholar]
  29. Gu, Y.; Wang, C.; Liu, B.; Zhang, Y. A kernel-based nonparametric regression method for clutter removal in infrared small-target detection applications. IEEE Geosci. Remote Sens. Lett. 2010, 10, 469–473. [Google Scholar] [CrossRef]
Figure 1. The bright-dark-bright pattern of the adjacent region of an IR target.
Figure 1. The bright-dark-bright pattern of the adjacent region of an IR target.
Applsci 09 05570 g001
Figure 2. Three sub-regions of a small image patch.
Figure 2. Three sub-regions of a small image patch.
Applsci 09 05570 g002
Figure 3. The proposed method for detecting small IR targets.
Figure 3. The proposed method for detecting small IR targets.
Applsci 09 05570 g003
Figure 4. The intensity distribution of two real small IR targets in a DoG-filtered image.
Figure 4. The intensity distribution of two real small IR targets in a DoG-filtered image.
Applsci 09 05570 g004
Figure 5. The twenty-first image in Seq.2. (x,y) in each sub-caption is the coordinate of the target center, and the detected target is circled in yellow.
Figure 5. The twenty-first image in Seq.2. (x,y) in each sub-caption is the coordinate of the target center, and the detected target is circled in yellow.
Applsci 09 05570 g005
Figure 6. The twenty-fifth image in Seq.2.
Figure 6. The twenty-fifth image in Seq.2.
Applsci 09 05570 g006
Figure 7. The thirty-second image in Seq.2.
Figure 7. The thirty-second image in Seq.2.
Applsci 09 05570 g007
Figure 8. The thirty-seventh image in Seq.1.
Figure 8. The thirty-seventh image in Seq.1.
Applsci 09 05570 g008
Figure 9. The twentieth image in Seq.2.
Figure 9. The twentieth image in Seq.2.
Applsci 09 05570 g009
Figure 10. The forty-ninth image in Seq.1.
Figure 10. The forty-ninth image in Seq.1.
Applsci 09 05570 g010
Figure 11. The first image in Seq.2.
Figure 11. The first image in Seq.2.
Applsci 09 05570 g011
Figure 12. Detection results on the second image in Seq.3.
Figure 12. Detection results on the second image in Seq.3.
Applsci 09 05570 g012
Figure 13. Detection results on the second image in Seq.4.
Figure 13. Detection results on the second image in Seq.4.
Applsci 09 05570 g013
Table 1. Quantitative assessment on Seq.1 (time is the average per image processing time).
Table 1. Quantitative assessment on Seq.1 (time is the average per image processing time).
MethodOur MethodRLCMPCMVar-DiffAAGDLoGLIGMax–MedianMax–MeanTop-Hat
P d 0.99000.060000.260000.35000.810010.99000.9000
P f 0.01000.050000.130000.23000.0900000.0400
SCRG203.91310.39300.00250.44810.00550.02750.02750.35770.21930.1267
BSF113.81300.00041.18720.23590.07030.00180.00180.000040.00010.0004
Time (second)1.857240.35410.47090.13620.15370.11447.55866.857315.45460.1023
Table 2. Quantitative assessment of Seq.2.
Table 2. Quantitative assessment of Seq.2.
MethodOur MethodRLCMPCMVar-DiffAAGDLoGLIGMax–MedianMax–MeanTop-Hat
P d 0.96880.343800.031300.90630.53130.87500.90630.9063
P f 00.562600.875000.06250.31250.12500.06250.0625
SCRG4920.86110.00290.02520.06320.00300.01720.14820.25150.09010.1482
BSF612.03570.00057.85532.19690.13170.00190.00090.00020.00040.0009
Time (second)0.869014.68530.18610.10570.11020.10402.13912.01894.60560.1294
Table 3. Quantitative assessment of Seq.3.
Table 3. Quantitative assessment of Seq.3.
MethodOur MethodLoGLIGMax–MedianMax–MeanTop-Hat
P d 0.96670.80000.86670.71110.73330.8444
P f 0.03330.20000.13330.28890.26670.1556
Table 4. Quantitative assessment of Seq.4.
Table 4. Quantitative assessment of Seq.4.
MethodOur MethodLoGLIGMax–MedianMax–MeanTop-Hat
P d 0.65560.80300.73960.57290.69790.7369
P f 0.34440.19690.26040.42710.30210.2604

Share and Cite

MDPI and ACS Style

Zhang, Y.; Zheng, L.; Zhang, Y. Small Infrared Target Detection via a Mexican-Hat Distribution. Appl. Sci. 2019, 9, 5570. https://doi.org/10.3390/app9245570

AMA Style

Zhang Y, Zheng L, Zhang Y. Small Infrared Target Detection via a Mexican-Hat Distribution. Applied Sciences. 2019; 9(24):5570. https://doi.org/10.3390/app9245570

Chicago/Turabian Style

Zhang, Yubo, Liying Zheng, and Yanbo Zhang. 2019. "Small Infrared Target Detection via a Mexican-Hat Distribution" Applied Sciences 9, no. 24: 5570. https://doi.org/10.3390/app9245570

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop