Next Article in Journal
A Self-Collision Detection Algorithm of a Dual-Manipulator System Based on GJK and Deep Learning
Previous Article in Journal
Design and Test of a New Dielectric-Loaded Resonator for the Accurate Characterization of Conductive and Dielectric Materials
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Single-Pixel Background Modeling Algorithm for Strong Sky Scenes Based on Local Region Spatial Bases

1
School of Computer Sciences and Engineering, Guilin University of Aerospace Technology, Jinji Road, Guilin 541004, China
2
Institute of Optics and Electronics, Chinese Academy of Sciences, Guangdian Avenue, Chengdu 610209, China
3
University of Chinese Academy of Sciences, Yuquan Road, Beijing 100049, China
*
Author to whom correspondence should be addressed.
Sensors 2023, 23(1), 522; https://doi.org/10.3390/s23010522
Submission received: 19 November 2022 / Revised: 18 December 2022 / Accepted: 19 December 2022 / Published: 3 January 2023
(This article belongs to the Section Sensing and Imaging)

Abstract

:
In the dim-small target detection field, background suppression is a key technique for stably extracting the target. In order to effectively suppress the background to enhance the target, this paper presents a novel background modeling algorithm, which constructs base functions for each pixel based on the local region background and models the background of each pixel, named single pixel background modeling (SPB). In SPB, the low-rank blocks of the local backgrounds are first obtained to construct the background base functions of the center pixel. Then, the background of the center pixel is optimally estimated by the background bases. Experiments demonstrate that in the case of extremely low signal-to-noise ratio (SNR < 1.5 dB) and complex motion state of targets, SPB can stably and effectively separate the target from the strongly undulant sky background. The difference image obtained via SPB background modeling has the characters: the non-target residual could be white noise, and the target is significantly enhanced. Compared with the other typical five algorithms, SPB remarkably outperforms other algorithms to detect the target of a low signal-to-noise ratio.

1. Introduction

It is difficult to stably detect a dim-small target whose size is no larger than 9   ×   9 pixels and whose energy is weak. The dim-small target detecting ability is a critical performance index of a remote detection system. Therefore, dim-small target detection has always been a research hotspot in the field of target detection. In complex scenes, due to the interference of environmental clutters such as an undulant sky background, illumination, clouds, and man-made artifacts, the contrast between the target and the local background is extremely weak. Therefore, in complex scenes, in order to highlight and stably detect the target, it is necessary to suppress or eliminate the background.
In the field of dim-small target detection, there are two main kinds of classical background modeling algorithms, which are based on two different image models. The first kind considers the image as the composition of background, noise, and target. Typical algorithms of this kind include the Gaussian filtering algorithm [1,2], the top hat transformation algorithm [3], the median filtering algorithm [4,5], and so on. Their basic idea is that the local neighborhood pixels of the background are statistically correlated, and the background of the target could be properly fitted by its neighborhood background. This kind of algorithm directly estimates the background but may severely damage the target signal. The other kind of algorithm regards the image data as the composition of a low-rank background matrix and sparse matrix from the perspective of the mathematical matrix. Recently, many algorithms have been proposed to estimate the background based on the low-rank matrix, such as the robust PCA algorithm [6,7,8,9], IPI [10], and PSTNN [11]. They mine the matrix structure information of image data, so that the integrity of the target signal is better. From the perspective of the global characteristics of the algorithms, the common point of these two kinds of algorithms is that they use a global function with fixed parameters to estimate the background image without fully considering the change in local characteristics. The first kind of algorithm uses a global structural element (or sliding window) to estimate the background of the central pixel. The second kind of algorithm uses principal component analysis (PCA) to construct the global background estimation function and obtain the low-rank component of the image data matrix.
The high-altitude dim-small target detection scene under strong sunlight is called the strong sky scene. Unlike the ground scene and low altitude scene whose backgrounds are mainly composed of vegetation, water flow, artificial objects, clouds, and other natural environments, the background of the strong sky scene is mainly composed of noise and non-uniform vignetting. When the non-uniform vignetting is very strong, the SNR of the target will be severely weakened, and the target will almost be submerged in the strong vignetting background. In this case, the geometric feature and gray feature of the dim-small target almost completely disappear, which makes the algorithm of the first kind fail to achieve good results. Similarly, the strong vignetting makes the image excessively sparse, that is, the sparsity of the target and noise is greatly weakened, and the target becomes a part of the background. As a result, the second kind of algorithm cannot effectively separate the background and the target. Therefore, in order to solve the difficult problem that is how to effectively separate the dim-small target from the strong vignetting background, this paper presents a background modeling algorithm for a single pixel based on the local region background base (SPB).
The contribution of this paper is that it is a new pixel-level background estimation algorithm, which provides an effective background suppression method under the condition of extremely low SNR (SNR < 1.5 dB) for the strong vignetting sky scene. The characteristics of SPB are as follows:
  • It is a background suppression algorithm based on a single frame. The background value of a single pixel is determined by the base of the neighborhood background space. Therefore, the background estimation of an image can be realized by using a single-frame image.
  • Good robust performance can be achieved when the dim-small target with different characteristics. In the SPB algorithm, as the neighborhood background fluctuates, the weight coefficient of the background base for linear combination will be adjusted instead of adopting a global function with fixed parameters. Therefore, the algorithm can adapt to the fluctuation of the background and the change of SNR, velocity, and trajectory of the target.
  • Residual noise has a high degree of white noise. Although the characteristics of each region of the image are different, the residual noise of the difference images has evident white noise characteristics, including equal distribution and spectrum fluctuation around the expected value.
  • Good effect can be stably obtained for strong sky scenes under the condition of extremely low SNR (SNR < 1.5 dB). In the case of extremely low SNR, in terms of energy, there is almost no difference between the target gray value and that of the neighborhood; in terms of scale, there is almost no difference between the target scale and the noise scale. However, for these scenes, the SPB algorithm can well separate the background and the target, and the target can be significantly enhanced.
In the rest of this article, Section 2 mainly discusses the background modeling algorithms based on image composition and matrix mathematical composition analysis. In the Section 3, based on discussing the image characteristics of the strong sky scene, the model-building method and background estimation method of the SPB algorithm are given. Background suppression experiments and quality assessment were carried out in the Section 4. The Section 5 is a brief conclusion.

2. Related Works

For the dim-small target scene, the geometric and grayscale features of the target are not obvious, making the target almost indistinguishable from noise and background texture. Therefore, deep learning algorithms are generally not used to estimate the background. The focus of background estimation research is mainly on image structure and matrix mathematical component analysis.
From the perspective of image structure, the image is composed of a background, noise, and a target. Algorithms based on this cognitive perspective can fall into two main categories. The first category is to estimate the background according to the good continuity and correlation of the background region and then subtract the estimated background from the original video frame image to obtain the difference image that the target had enhanced. Furthermore, this kind of algorithm can be divided into three subclasses: the filtering algorithm, the statistical model algorithm, and the algorithm based on the classification problem. The filtering algorithm uses a structure element (sliding window) to obtain the neighborhood pixel grayscale information to estimate the background value of the center pixel. Common algorithms include anisotropy algorithm [12], top hat transformation algorithm [3], median filtering algorithm [4,5], mean filtering algorithm [1,4], bilateral filtering algorithm [13], two dimensional least mean square error algorithm [14], and so on. While suppressing the background, these algorithms also make the target information lose. For example, the target is smoothed, the target edge information is lost, etc. They are less adaptable when the background is complex, such as rich textures, diverse background types, and small areas. The statistical model algorithm establishes appropriate statistical models for various pixels to obtain better-estimated values for complex backgrounds. Gaussian mixture model algorithm [1,2] is the typical representative of this kind of algorithm. However, the real-time performance of this kind of algorithm is poor, mainly for the following reasons: The background model initialization requires sequence images, and the model learning needs to be redone when the background changes significantly; with the increase of the number of statistical models that are involved in mixing, the complexity of updating the model parameters will increase sharply. The background modeling algorithm based on the classification problem has a good real-time performance. The typical representative algorithms are VIBE [15,16] algorithm and its derivative algorithms. Such as PBAS [17], VIBE+ [18], etc. These algorithms only need a single frame of the image to build the sample set of each pixel, so the background model initialization can be done quickly. However, the most obvious disadvantages of this kind of algorithm are the presence of ghost areas, the absorption of stationary or slow-moving targets, shadow foreground, and incomplete moving targets. These shortcomings are closely related to the sample set update strategy for pixels. Different from the first category, which directly estimates the background, the second category is to make use of the singularity characteristics of the target to match, enhance and extract, so as to achieve the separation of target and background. There are two typical subclasses of this category: algorithms based on the human visual system and dictionary-matching lookup algorithms. The algorithm based on the human visual system obtains the target matching scale chart according to the salience of the contrast between the target and the neighborhood background [11]. Such as the difference between the Gaussian algorithm (DOG) [19], Laplacian of Gaussian algorithm (LOG) [20], RLCM [21], etc. The dictionary matching look-up algorithm is to establish the target dictionary according to the salient characteristics of the target, and the target can be confirmed when it matches the dictionary. Such as the works of the literature [22,23,24], etc. The most important deficiency of the second category algorithms is that the target must have good salience characteristics, that is, the target SNR must be high, otherwise the algorithm effect will be affected.
Analyzing the data components of the image matrix from the perspective of the mathematical matrix is a modeling method to explore the intrinsic law of image data. A typical representative of this approach is the low-rank background modeling algorithm based on principal component analysis (PCA) [7]. In order to improve the robustness of the principal component analysis algorithm and effectively apply it to high-dimensional data, Candès et al. [6] proposed a novel robust PCA algorithm (RPCA) and successfully applied it to signal processing and computer vision. This work has made PCA - based low-rank background modeling problems a hot topic in the recent decade. According to different applications, many algorithms have been proposed. Typically, a real data sequence will last a long time and may change slowly. In this case, the robust subspace tracing (RST) [8,9] model was proposed to reduce the computational complexity and achieve continuous processing power. The RST model assumes that long data sequences reside in low-dimensional subspaces. This model is beneficial to solving the problem of insufficient memory space and realizing continuous real-time processing. The previous algorithms obtain the observation data matrix by converting the image into the column or row vector of the matrix, which will destroy the original structural characteristics of the image. Different from the algorithms that were mentioned above, in order to maintain the structure of the image and make good use of it, an algorithm based on tensor (TRPCA) [25] to obtain observation matrix data was proposed. The aforementioned robust PCA algorithms, that is, RPCA, RST, and TRPCA use sequence images to form data matrix in background modeling applications, that is, background initialization needs to be realized by using sequential frame images. When the background changes fast or the target moves rapidly, the effect of the background estimation will be affected. Therefore, in order to adapt to the rapid change of background or target, the robust PCA background modeling algorithm for a single frame image based on patch-image has had more attention in the recent years. Such as IPI [10,26,27], IPT [28], PSTNN [11], etc. These algorithms assume that the background of a single frame image is composed of the same low-rank subspace and the observation data matrix is obtained by region segmentation of a single frame image.
For dim-small target detection, the biggest challenge is carring out background modeling under the condition of extremely low SNR (SNR < 1.5 dB), so that the background and target can be effectively separated. The main reason for this is that in the case of extremely low SNR, the energy and geometric features of the target are not prominent so the target is almost completely integrated into the background and almost indistinguishable from the noise. Therefore, most of the current algorithms mainly focus on the input SNR ≥ 1.5 dB and rarely involve the scene with very low SNR. In order to effectively separate the background and target of a strong sky scene under very low SNR, the SPB algorithm is proposed in this paper. SPB takes advantage of the robust PCA algorithms’ feature that makes good use of the image structure. The method of estimating single-pixel background values based on background bases that were proposed by the SPB algorithm can adapt to scenes with fluctuating backgrounds and achieve a good target enhancement effect under the condition of extremely low SNR.

3. SPB Algorithm

This section will introduce the SPB Algorithm. Firstly, according to the characteristics of strong sky scene images, the main difficulties that affect dim-small target detection were analyzed. Then, the low-rank and sparse properties of strong sky scene images in global and local regions were discussed. A single-pixel background modeling model based on the background space base of the local region was presented. Finally, the optimal estimation of the single-pixel background value was discussed in detail.

3.1. Image Characteristic Analysis

In the sky dim-small target detection, the main factors that affect the detection effect are the non-target photon noise and the dynamic random non-uniform strong vignetting caused by the system aperture effect. Figure 1a is the case where there is no non-uniformity phenomenon. Apart from noise, there is no obvious vignetting on the image, and the target is prominent. There is a serious non-uniformity phenomenon in Figure 1b. The target is drowned by the random strong vignetting, and the size and grayscale features of the target are completely impossible to be directly recognized by vision. Three successive frames are randomly selected from the sequence frame images, and then their row and column pixel grayscale values are summed respectively. The results of the summation of each row and column are shown as the curves in Figure 1c,d, respectively. As shown in the summation curves, the pixel values of the corresponding columns and rows between the previous and subsequent frames fluctuate significantly, which intuitively reflects the strong dynamic changes of the vignetting. From these analyses, it can be seen that there are three main difficulties affecting dim-small target detection. First, because of the wide distribution area and high gray value of the random non-uniform strong vignetting, the target and noise are submerged, making the sparsity component of the video frame image not obvious. In this case, if the algorithm of global robust PCA is used to carry out background estimation, the result is that the target is included in the low-rank background, that is, the difference image does not effectively contain the target. For example, Figure 2a is the original video frame image, on which there is a target as shown in Figure 2d. Figure 2b,c are the low-rank backgrounds and difference images that were obtained by the algorithm of global robust PCA, respectively. Obviously, due to the weak sparsity of the target in the original image, the target was integrated into the strong vignetting background, so that the algorithm of global robust PCA could not effectively separate the target from the background. Second, the distribution of vignetting has the characteristics of randomicity. The distribution of vignetting may be circular, elliptical, or other shapes depending on the working environment and the mode of movement of the imaging system. At the same time, the strength of each local area of vignetting is also random. The random distribution of vignetting will affect the adaptability of the background modeling algorithm to different scenes. Third, the random non-uniform strong vignetting is dynamic change. This dynamic change will weaken the correlation of the background image in the time domain and affect the effect of the background modeling algorithm based on the sequential frame image.
The rank of a matrix is equal to the number of non-zero singular values of the matrix, and the sparsity of a matrix can be measured by the sparse factor (SF). The definition of sparse factor is shown in Equation (1).
SF = N t Q
where t is the set of elements whose values are greater than the threshold value T, t = { x i | x i > T ,   x i D } , x i is the ith element of D. T takes the mean value of matrix D. N t and Q are the total number of elements of t and D, respectively. The low-rank and sparsity features of global (the whole image) and local regions are discussed below. Figure 3 is the global and local low-rank characteristic analysis diagram of random non-uniform strong vignetting images. In Figure 3a, A, B, and C are three randomly selected local regions. To enhance readability, the logarithm of the singular value gradient is taken to plot the curve. Figure 3 shows that the global singular value gradient becomes flat after the fifth singular value, and the local singular value gradient curve becomes flat after the second or third singular value. It shows that they all have good low-rank characteristics. Figure 4 is the sparse factor curve reflecting the global and local sparse characteristics. Table 1 shows the data of the sparse factors and their corresponding regional scale. The data in Figure 4 and Table 1 shows that as the local regional scale decreases, the sparse factor increases. This reflects that for strong vignetting images, the local region is more obvious sparsity than the global one, that is, the proportion of sparse elements in the local region is higher than that in the global one. Therefore, the analysis data of low-rank characteristics and sparse characteristics shows that the local region maintains appropriate sparsity, that is, avoids the occurrence of over-sparsity phenomenon, and still retains good low-rank characteristics.
The threshold value T of each local region takes the mean value of their grayscale.

3.2. SPB Background Model Modeling

3.2.1. Model Analysis

Figure 5 is the schematic diagram of the single-pixel background modeling algorithm (SPB). In practice, the background is always disturbed. Figure 5a is the background model of region B in the presence of interference. In order to make good use of the characteristic information of the background region to estimate the background value of the single pixel, region B is divided into several smaller local blocks. Thus, these local blocks have better consistent properties than region B. Assume that the background basis function of the ith local block A i is φ i . As a result of the continuity of the background region, the basic function Φ of the real background of the local block where pixel O is located can be estimated by the background basis functions of all the local blocks, i.e.,
Φ = i = 1 k ω i φ i
ω i is the weight coefficient, 0 ω i 1 , i = 1 k ω i = 1 , i = 1 ,   ,   k .
Figure 5. The schematic diagram of SPB algorithm. (a) The background model of region B in the presence of interference. (b) The schematic diagram of strong vignetting background modeling.
Figure 5. The schematic diagram of SPB algorithm. (a) The background model of region B in the presence of interference. (b) The schematic diagram of strong vignetting background modeling.
Sensors 23 00522 g005
Figure 5b is the schematic diagram of strong vignetting background modeling. The previous analysis of the data in Figure 3 and Figure 4 and Table 1 shows that the vignetting local block has a low rank and sparsity characteristics. Therefore, a low-rank background model can be established for the local block A i , that is
minmize L i ,   P i     Rank ( A i )
subject   to   A i = L i + P i
  Rank ( A i ) r ,     P i 0 z
where, L i and P i are low-rank background matrix and sparse component matrix, respectively. 0 is the zero norm, and Rank ( ) is the rank of a matrix. Both r and z are positive real numbers. The low-rank background L i can be obtained from Equation (3), and then the basis function φ i of L i can be found. Thus, as Equation (4), the optimal estimation model of the background value of the central pixel O can be established.
arg min     f ( x , y ) f O ( x , y ) 2
subject   to     f O ( x , y ) = θ ( Φ ) = θ ( ω φ T )
0 ω i 1 ,   i = 1 k ω i = 1
where, f ( x , y ) is the grayscale value of pixel O, and f O ( x , y ) is its optimal background value. θ ( ) is a nonnegative function of Φ . ω = { ω i ,   i = 1 ,     ,   k } , φ = { φ i ,   i = 1 ,     ,   k } .

3.2.2. Background Estimation

For the strong vignetting background, when the frame image has the over-sparsity phenomenon, the previous analysis shows that the appropriate sparse feature of the local region can be used to obtain the low-rank background. Then, according to the continuity of the local background, the background value of the central pixel is estimated by using the background space basis function of the local region.
For simplicity, A i , L i , and P i will be written in terms of A, L, and P for brevity when there is no ambiguity. Here, rewrite Equation (3) as follows
minmize L ,   P     Rank ( A )
subject   to   A = L + P
  Rank ( A ) r ,     P 0 z
Equation (5) is a typical principal component analysis model. According to RPCA theory [6,7,8,29], it can be transformed into the principal component tracking model of the following equation for discussion
minimize     L * + λ P 1
subject   to   A = L + P
where, * is the nuclear norm, and 1 is the 1 norm. λ is the parameter that balances the terms L and P, 0 <   λ   <   1 . A is the local block. L is the background matrix. P is the sparse matrix, which is composed of noise and potential target. According to the optimization theory [30,31,32], in order to improve the universality, the augmented Lagrange optimization algorithm is used to find the optimal solution of (6), i.e.,
l ( L , P , Y ) = L * + λ P 1 + Y , A L P + μ 2 A L P F 2
where Y is the Lagrange multiplier. μ is the penalty parameter, μ > 0 . By direct repeated iteration and update, the optimal solution of Equation (7) can be obtained, but the computing complexity is high. In order to reduce the complexity, the optimization method uses the soft-thresholding operator [30] and the singular value thresholding operator [33], i.e.,
T t [ x ] = sgn ( x ) max ( | x | t , 0 )
H t ( X ) = U T t ( Λ ) V *
where, x R , t > 0 . For two matrices, G and M, with the same scale, a new matrix R is obtained through the soft threshold operation by traversing all the elements, i.e., R = T G [ M ] = T g i j [ m i j ] . Here, g i j , m i j are the elements of the matrices G and M, respectively. X = U Λ V * is singular value decomposition of matrix X. Thus, the optimal solution of Equation (7) is obtained. That is
arg min L     l ( L ,   P ,   Y ) = H 1 μ ( A P + μ 1 Y ) arg min P     l ( L ,   P ,   Y ) = T λ μ ( A L + μ 1 Y )
Y k + 1 = Y k + μ ( A L k + 1 P k + 1 )
It is well known that the local regional background has strong consistency. Therefore, in order to estimate the background value of the center pixel of the region, the operations as follow are carried out: the eigenvector ε i of the L i local block background matrix composes the basis functions, then the background value f O ( x , y ) of the center pixel O can be modeled on these basis functions, i.e.,
f O ( x , y ) = θ ( Φ ) = 1 n j = 1 n ω ε T
Here, the nonnegative function θ ( ) is the mean function. ω = { ω i ,   i = 1 ,     ,   k } is the weighting coefficient, 0 ω i 1 , i = 1 k ω i = 1 . ε = { ε i ,   i = 1 ,     ,   k } is the basis function vector. ε i is the eigenvector of the background matrix L i . n is the number of all pixels in a local block.
Now, the specific expression of the SPB can be written as Formula (12)
arg min     f ( x , y ) f O ( x , y ) 2
subject   to     f O ( x , y ) = 1 n j = 1 n ω ε T
0 ω i 1 ,   i = 1 k ω i = 1
For conciseness, in the case of without causing ambiguity, f ( x , y ) and f O ( x , y ) are abbreviated as f and f O . Then Equation (12) can be written as following
arg min     f f O 2
subject   to     f O = 1 n j = 1 n ω ε T
0 ω i 1 ,   i = 1 k ω i = 1
The Lagrange multiplier method [32] is used to find the optimal solution of Equation (13). First, Equation (13) is transformed into
l ( ω , γ ) = f 1 n j = 1 n ω ε T 2 + γ ( i = 1 k ω i 1 )
Then, the derivative method is used to find the optimal solution of ω and γ
arg min ω i     l ( ω , γ ) = l ( ω , γ ) ω i
arg min γ     l ( ω , γ ) = l ( ω , γ ) γ
γ is the Lagrange multiplier.
In order to facilitate the understanding of the SPB algorithm, according to the above analysis, the pseudo-code and the flow diagram of the SPB algorithm are given out in Algorithm 1 and Figure 6.
Algorithm 1 The pseudo code of SPB algorithm
Input: Original video frame images I R o × p .
Output: Background images I L and difference images I P .
For i0 = 1:num//num is the total number of frames of the video.
  1. Read in the i0th frame image I, and get its size m and n.
  2. Sliding the window W on the image I pixel by pixel to intercept the local region B.
  3. B is divided into (k + 1) local blocks A i , i = 0 , 1 ,     ,   k .
  4. The low-rank background L i of local block A i is calculated by using the Equation (10).
  5. Find the eigenvector ε i of the background matrix L i .
  6. The optimal weight coefficient ω i is calculated according to the Equation (15).
  7. The optimal background value f O of the central pixel O is estimated out by using the Equation (11).
  8. Obtain the difference image: I P = I I L .
End for
  9. Output I L , I P .

4. Experiments

In this part, experiments were carried out to verify the performance of SPB in suppressing the background and enhancing the target. Based on the experimental results, the robustness of the SPB algorithm is analyzed for the targets with different characteristics, the characteristics of the SPB residual noise are discussed, and a comparison with five other state-of-art algorithms is described. Furthermore, the experimental results are used to quantitatively analyze the operating range (OR), SNR gain (SNRG), and background suppress factor (BSF) of the SPB algorithm.

4.1. Experimental Environment

The experiments were carried out with OCTAVE on a computer running a 32-bit Windows 7 system, with a Core-i5 CPU, and 3 GB of RAM. For the strong sky scenes, in order to demonstrate the SPB algorithm performance of the background suppression and the dim-small target enhancement, three typical strong sky scenes with strong random vignetting are used for the experiments. The center area of the vignetting of scene A is large, scene B has random textures floor, and the vignetting of scene C changes significantly all the time. Five robust PCA algorithms: PCP [6], FPCP [34], OSTD [35], IPI [10], and PSTNN [11] were used as comparison algorithms. The experimental parameters of the comparison algorithms are shown in Table 2, where D is the data matrix or tensor. The evaluation indexes mainly include the Spectrum diagram, energy distribution diagram, SNR, SNRG, BSF, and operating range (OR). The definitions of SNR, SNRG [36], BSF, and OR are as follows
SNR = 10 lg ( α t α l δ l )
SNRG = 10 lg ( SNR out SNR in )
BSF = δ in δ out
OR min ( SNR in )
where, α t and α l are the grayscale mean values of the target and the target neighborhood region, respectively. δ l is the grayscale standard deviation value of the target neighborhood region. SNR in and SNR out are the input and output SNR of the images before and after the background suppression. δ in , δ out are the grayscale standard deviation values of the images before and after the background suppression. Here, the target neighborhood region scale is 3 times the target scale. The operating range (OR) is used to measure the input signal-to-noise ratio (SNR) range in which the algorithm can stably and effectively detect the dim-small targets. For dim-small target detection, generally, the algorithm can effectively detect the target with a high input SNR, and the detection effect of the algorithm is affected by a low input SNR. Therefore, the OR is determined by the minimum input SNR of the target that the algorithm can detect.

4.2. Quality Estimate

This section mainly discusses the performances of the background suppression and target enhancement of the SPB algorithm for the strong sky scenes from three aspects: the robustness of SPB, the suppression effect of SPB on the vignetting background, and the comparison of the experimental results with other five state-of-the-art methods.

4.2.1. Robustness to the Targets with Different Characteristics

For the strong sky scenes, the characteristics of dim-small targets mainly include the ratio of target signal energy to its neighborhood background energy (signal-to-noise ratio, SNR), target velocity, and target trajectory. The robustness of these characteristics is an important factor to measure the applicability of an algorithm to different strong sky scenes. The characteristics of the dim-small targets in these three scenes are listed in Table 3. The mean values of input SNR ( SNR in ) range from −0.2066 dB to 3.0237 dB. The SNR in values of a single frame image vary from −12.0820 dB to 11.1423 dB. The target speed ranges from 0 pixels per frame to 2 pixels per frame. The target in scene A first moves quickly up and then slowly to the lower right. In scene B, the target moves slowly all the time. In scene C, the target reciprocating moves quickly up and down, then left and right. In Figure 7, the target movement tracks of scenes A, B, and C are shown from left to right. The target trajectory in scene A is close to an inverted V. In scene B, the target swings near the original position. In scene C, the target reciprocating moves with a complex trajectory.
Figure 8 is the difference images when the target SNR in each scene is minimum and maximum, respectively. Figure 9 shows the difference images when the target velocity in each scene is minimum and maximum, respectively. For each scene, one frame was randomly selected when the target was moving on the trajectory, and their difference images are showed in the Figure 10. As demonstrated in Figure 8, Figure 9 and Figure 10, the SPB algorithm can effectively suppress the background and enhance the target when the target is in an extremely low SNR, moving at different speeds and different trajectories, that is, the SPB algorithm is robust to targets with different characteristics, and is adaptive to different strong sky scenes.

4.2.2. Analysis of Residual Noise Characteristics

For strong sky scenes, the purpose of the SPB algorithm is to suppress the background and whiten the residual noise of the difference image, that is, the statistical characteristics of the residual noise conform to the characteristics of white noise, and the target is significantly enhanced. One frame image was randomly selected from each scene, and the spectrum characteristics and energy distribution of their difference images obtained by SPB are shown in Figure 11. In this figure, the method to obtain the spectrum of the difference images is to randomly select a row and a column, and then draw the spectrum curves of this row and column, respectively. As can be seen from Figure 11, the spectrum of the difference images of these three scenes fluctuates around the mean line, which is approximate to the spectrum of white noise. The residual noise in the difference images is evenly distributed, and the target energy is significantly stronger than that of the noise. Therefore, it can be seen that the SPB algorithm whitens the residual noise, and significantly enhances the target.
The first row is the original video frame images. The difference images are in the second row. The third and fourth rows are the spectrum images of a row and a column that were randomly selected from the difference images. The fifth row is the three-dimensional energy distribution diagrams of the difference images.

4.2.3. Visual Comparison with Other Algorithms

Figure 12, Figure 13 and Figure 14 show the visual comparison of the background suppression effect between the SPB algorithm and the PCP, FPCP, OSTD, IPI, and PSTNN algorithms. These three sets of pictures correspond to scenes A, B, and C respectively, and one frame image is randomly selected from each scene. The first row of each set is the original video frame image. In each set, from the second row to the seventh row, the background image is on the left, the difference image is in the middle, and the three-dimensional energy distribution image is on the right. For the strong sky scene, the experimental results demonstrate that the six algorithms all have the ability to suppress the background, but the separation effect of sparse components and the target enhancement ability is very different. The difference images of SPB, PCP, and FPCP all contain sparse components, but only the difference image of SPB could effectively obtain the target that is significantly enhanced. OSTD, IPI, and PSTNN could not effectively separate the sparse components, that is, there is no effective sparse component in their difference images.

4.3. Quantitative Evaluation

In addition to the SPB algorithm proposed in this paper, which can effectively separate dim-small targets from the strong vignetting background, the other five algorithms are unable to effectively separate targets from the background, that is, the difference images of these five algorithms cannot contain effective target information. Therefore, only a quantitative analysis of the SPB algorithm was carried out in this section.
The input SNR ( SNR in ) directly affects the difficulty of separating the dim-small target from the background. The smaller the SNR in , the more difficult it is to separate the target from the background. Therefore, whether a good target enhancement effect can be achieved under different SNR in can be used to measure the operating range (OR) of an algorithm. The lower the effective SNR in , the wider the OR of the algorithm. As shown in Table 3 and Table 4 and Figure 8, when SNR in 12.0820 dB , the SPB algorithm can effectively separate the target from the background. Thus, the operating range of the SPB is OR 12.0820   dB . The stronger the enhancement ability of an algorithm to the target, the higher the value of SNRG. In Table 4, the mean SNRG values of scenes A, B, and C are 3.0103 dB, 5.5784 dB, and 6.6343 dB, respectively, indicating that the SPB algorithm can effectively enhance the target in the working range of OR 12.0820   dB . BSF measures the ability of an algorithm to suppress the background. The better the ability of an algorithm to suppress the background, the more evenly the residual noise distribution of the difference image, the smaller the variance of the difference image, and the larger the BSF value. In Table 4, the mean BSF values of the SPB algorithm for scene A, B, and C are 2.8061, 1.7039, and 13.8459, respectively, indicating that the algorithm performs effectively to suppress the background.

5. Conclusions

SPB is a novel single-frame background modeling algorithm based on the background space basis function. The algorithm achieves good background suppression and target enhancement performance, which is mainly reflected in the following aspects. Firstly, SPB is robust to targets with different SNR, velocity, and trajectory. Experiments demonstrate that the SPB could effectively separate the dim-small target from the background under extremely low SNR ( SNR < 1 . 5 dB ), moving speed up to 2 pixels/frame, and complex trajectory. Secondly, the residual noise could be well whitened, whose spectrum is evenly distributed. Finally, the target has been significantly enhanced. When the averages of the input SNR of Scene A, B, and C were 3.0237 dB, − 0.2066 dB, and 2.8393 dB, respectively, the SNR gain mean values obtained by SPB are 3.0103 dB, 5.5784 dB, and 6.6343 dB, respectively. Meanwhile, the corresponding mean values of background suppress factor (BSF) are 2.8061, 1.7039, and 13.8459, respectively.
At present, SPB only investigates matrix features from the perspective of the spatial domain, while future works can be carried out from the perspective of the time domain and spatial domain fusion. The main shortcoming of the SPB algorithm is a bit complex and time-consuming due to the fact that to estimate the background, it needs to traverse the entire video frame image pixel by pixel.

Author Contributions

Conceptualization, Z.X.; Methodology, B.L.; Software, B.L. and Q.Z.; Formal analysis, Q.Z.; Investigation, Q.Z.; Data curation, Z.X.; Writing—original draft, B.L.; Writing—review & editing, J.Z.; Supervision, Z.X.; Project administration, B.L.; Funding acquisition, J.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research was partly supported by Science and Technology Program of Guilin (grant number: 20220107-3), Doctoral Fund of the Guilin University of Aerospace Technology (grant number: KX202201101), and the West Light Foundation of the Chinese Academy of Sciences (grant number: ya18k001).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data used to support the findings of this study are included within the article.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Sobral, A.; Vacavant, A. A comprehensive review of background subtraction algorithms evaluated with synthetic and real videos. Comput. Vis. Image Underst. 2014, 122, 4–21. [Google Scholar] [CrossRef]
  2. Goyal, K.; Singhai, J. Review of background subtraction methods using Gaussian mixture model for video surveillance systems. Artif. Intell. Rev. 2017, 50, 241–259. [Google Scholar] [CrossRef]
  3. Bai, X.; Zhou, F. Infrared small target enhancement and detection based on modified top-hat transformations. Comput. Electr. Eng. 2010, 36, 1193–1201. [Google Scholar] [CrossRef]
  4. Garcia-Garcia, B.; Bouwmans, T.; Silva, A.J.R. Background subtraction in real applications: Challenges, current models and future directions. Comput. Sci. Rev. 2019, 35, 100204. [Google Scholar] [CrossRef]
  5. Villar, S.A.; Acosta, G.G. Median Filtering: A New Insight. J. Math. Imaging Vis. 2016, 58, 130–146. [Google Scholar] [CrossRef]
  6. Candès, E.J.; Li, X.; Ma, Y.; Wright, J. Robust principal component analysis? J. ACM 2011, 58, 1–37. [Google Scholar] [CrossRef]
  7. Bouwmans, T.; Javed, S.; Zhang, H.; Lin, Z.; Otazo, R. On the Applications of Robust PCA in Image and Video Processing. Proc. IEEE 2018, 106, 1427–1457. [Google Scholar] [CrossRef] [Green Version]
  8. Vaswani, N.; Bouwmans, T.; Javed, S.; Narayanamurthy, P. Robust Subspace Learning: Robust PCA, Robust Subspace Tracking, and Robust Subspace Recovery. IEEE Signal Process. Mag. 2018, 35, 32–55. [Google Scholar] [CrossRef] [Green Version]
  9. Guo, H.; Qiu, C.; Vaswani, N. An Online Algorithm for Separating Sparse and Low-Dimensional Signal Sequences From Their Sum. IEEE Trans. Signal Process. 2014, 62, 4284–4297. [Google Scholar] [CrossRef] [Green Version]
  10. Gao, C.; Meng, D.; Yang, Y.; Wang, Y.; Zhou, X.; Hauptmann, A.G. Infrared Patch-Image Model for Small Target Detection in a Single Image. IEEE Trans. Image Process. 2013, 22, 4996–5009. [Google Scholar] [CrossRef]
  11. Zhang, L.; Peng, Z. Infrared Small Target Detection Based on Partial Sum of the Tensor Nuclear Norm. Remote Sens. 2019, 11, 382. [Google Scholar] [CrossRef] [Green Version]
  12. Zhou, H.; Zhao, Y.; Qin, H.; Yin, S.; Liu, G.; Zhao, D.; Yan, X. Rong Infrared dim and small target detection algorithm based on multi-scale anisotropic diffusion equation. Acta Photon-Sin. 2015, 44, 0910002. [Google Scholar] [CrossRef]
  13. Bae, T.-W.; Sohng, K.-I. Small Target Detection Using Bilateral Filter Based on Edge Component. J. Infrared Millim. Terahertz Waves 2010, 31, 735–743. [Google Scholar] [CrossRef]
  14. Bae, T.-W.; Kim, Y.-C.; Ahn, S.-H.; Sohng, K.-I. An Efficient Two-Dimensional Least Mean Square (TDLMS) Based on Block Statistics for Small Target Detection. J. Infrared Millim. Terahertz Waves 2009, 30, 1092–1101. [Google Scholar] [CrossRef]
  15. Barnich, O.; Van Droogenbroeck, M. ViBE: A powerful random technique to estimate the background in video sequences. In Proceedings of the 2009 IEEE International Conference on Acoustics, Speech and Signal Processing, Taipei, Taiwan, 19–24 April 2009; pp. 945–948. [Google Scholar] [CrossRef] [Green Version]
  16. Barnich, O.; Van Droogenbroeck, M. ViBe: A Universal Background Subtraction Algorithm for Video Sequences. IEEE Trans. Image Process. 2010, 20, 1709–1724. [Google Scholar] [CrossRef] [Green Version]
  17. Hofmann, M.; Tiefenbacher, P.; Rigoll, G. Background segmentation with feedback: The pixel-based adaptive segmenter. In Proceedings of the 2012 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops, Providence, RI, USA, 16–21 June 2012; pp. 38–43. [Google Scholar] [CrossRef] [Green Version]
  18. Van Droogenbroeck, M.; Paquot, O. Background subtraction: Experiments and improvements for ViBe. In Proceedings of the 2012 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops, Providence, RI, USA, 16–21 June 2012; pp. 32–37. [Google Scholar] [CrossRef] [Green Version]
  19. Dong, X.; Huang, X.; Zheng, Y.; Shen, L.; Bai, S. Infrared dim and small target detecting and tracking method inspired by Human Visual System. Infrared Phys. Technol. 2014, 62, 100–109. [Google Scholar] [CrossRef]
  20. Singh, M.; Mandal, M.K.; Basu, A. Gaussian and Laplacian of Gaussian weighting functions for robust feature based tracking. Pattern Recognit. Lett. 2005, 26, 1995–2005. [Google Scholar] [CrossRef]
  21. Han, J.; Liang, K.; Zhou, B.; Zhu, X.; Zhao, J.; Zhao, L. Infrared Small Target Detection Utilizing the Multiscale Relative Local Contrast Measure. IEEE Geosci. Remote Sens. Lett. 2018, 15, 612–616. [Google Scholar] [CrossRef]
  22. Liu, D.; Li, Z.; Liu, B.; Chen, W.; Liu, T.; Cao, L. Infrared small target detection in heavy sky scene clutter based on sparse representation. Infrared Phys. Technol. 2017, 85, 13–31. [Google Scholar] [CrossRef]
  23. Wan, M.; Gu, G.; Qian, W.; Ren, K.; Chen, Q. Robust infrared small target detection via non-negativity constraint-based sparse representation. Appl. Opt. 2016, 55, 7604–7612. [Google Scholar] [CrossRef]
  24. Li, Z.-Z.; Chen, J.; Hou, Q.; Fu, H.-X.; Dai, Z.; Jin, G.; Li, R.-Z.; Liu, C.-J. Sparse Representation for Infrared Dim Target Detection via a Discriminative Over-Complete Dictionary Learned Online. Sensors 2014, 14, 9451–9470. [Google Scholar] [CrossRef] [PubMed]
  25. Lu, C.; Feng, J.; Chen, Y.; Liu, W.; Lin, Z.; Yan, S. Tensor robust principal component analysis: Exact recovery of corrupted low-rank tensors via convex optimization. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 5249–5257. [Google Scholar] [CrossRef] [Green Version]
  26. Dai, Y.; Wu, Y.; Song, Y.; Guo, J. Non-negative infrared patch-image model: Robust target-background separation via partial sum minimization of singular values. Infrared Phys. Technol. 2017, 81, 182–194. [Google Scholar] [CrossRef]
  27. Guo, J.; Wu, Y.; Dai, Y. Small target detection based on reweighted infrared patch-image model. IET Image Process. 2018, 12, 70–79. [Google Scholar] [CrossRef]
  28. Dai, Y.; Wu, Y. Reweighted Infrared Patch-Tensor Model With Both Nonlocal and Local Priors for Single-Frame Small Target Detection. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2017, 10, 3752–3767. [Google Scholar] [CrossRef] [Green Version]
  29. Chandrasekaran, V.; Sanghavi, S.; Parrilo, P.; Willsky, A.S. Rank-Sparsity Incoherence for Matrix Decomposition. SIAM J. Optim. 2011, 21, 572–596. [Google Scholar] [CrossRef]
  30. Lin, Z.; Liu, R.; Su, Z. Linearized alternating direction method with adaptive penalty for low rank representation. arXiv 2011, arXiv:11090367. Available online: https://arxiv.org/pdf/1109.0367.pdf (accessed on 20 December 2020).
  31. Yuan, X.; Yang, J. Sparse and low-rank matrix decomposition via alternating direction method. Pac. J. Optim. 2009, 9, 11. Available online: http://www.optimization-online.org/DB_HTML/2009/11/2447.html (accessed on 15 October 2019).
  32. Boyd, S.; Vandenberghe, L. Convex Optimization; Cambridge University Press: Cambridge, UK, 2004. [Google Scholar]
  33. Cai, J.; Candès, E.J.; Shen, Z. A Singular Value Thresholding Algorithm for Matrix Completion. SIAM J. Optim. 2010, 20, 1956–1982. [Google Scholar] [CrossRef]
  34. Rodriguez, P.; Wohlberg, B. Fast principal component pursuit via alternating minimization. In Proceedings of the 2013 IEEE International Conference on Image Processing, Melbourne, Australia, 15–18 September 2013; pp. 69–73. [Google Scholar] [CrossRef]
  35. Sobral, A.; Javed, S.; Jung, S.K.; Bouwmans, T.; Zahzah, E.-H. Online stochastic tensor decomposition for background subtraction in multispectral video sequences. In Proceedings of the 2015 IEEE International Conference on Computer Vision Workshop (ICCVW), Santiago, Chile, 7–13 December 2015; pp. 946–953. [Google Scholar] [CrossRef]
  36. Dan, B.; Li, M.; Tang, T.; Qi, X.; Zhu, Z.; Ouyang, Y. Spatial–Temporal Stochastic Resonance Model for Dim-Small Target Detection. IEEE Geosci. Remote Sens. Lett. 2022, 19, 7507805. [Google Scholar] [CrossRef]
Figure 1. Analysis of vignetting characteristics. (a) There is no obvious vignetting on the image. (b) There is obvious vignetting on the image. (c) The curves of the summation of row pixel grayscale values of three random successive frames. (d) The curves of the summation of column pixel grayscale values of three random successive frames.
Figure 1. Analysis of vignetting characteristics. (a) There is no obvious vignetting on the image. (b) There is obvious vignetting on the image. (c) The curves of the summation of row pixel grayscale values of three random successive frames. (d) The curves of the summation of column pixel grayscale values of three random successive frames.
Sensors 23 00522 g001
Figure 2. The sparse target cannot be separated effectively. (a) The original video frame image. (b) The low-rank background image. (c) The difference image. (d) The target image.
Figure 2. The sparse target cannot be separated effectively. (a) The original video frame image. (b) The low-rank background image. (c) The difference image. (d) The target image.
Sensors 23 00522 g002
Figure 3. The global and local low-rank analysis of random non-uniform strong vignetting image. (a) Original video frame image. A, B, and C are three randomly selected local regions, and their scale is 15 × 15. (b) The singular value gradient logarithmic curve of the whole image. (ce) are the singular value gradient logarithmic curves of the regions A, B, and C, respectively.
Figure 3. The global and local low-rank analysis of random non-uniform strong vignetting image. (a) Original video frame image. A, B, and C are three randomly selected local regions, and their scale is 15 × 15. (b) The singular value gradient logarithmic curve of the whole image. (ce) are the singular value gradient logarithmic curves of the regions A, B, and C, respectively.
Sensors 23 00522 g003
Figure 4. The sparse factor curve.
Figure 4. The sparse factor curve.
Sensors 23 00522 g004
Figure 6. The flow diagram of the SPB methods.
Figure 6. The flow diagram of the SPB methods.
Sensors 23 00522 g006
Figure 7. The target movement tracks of scenes A, B, and C.
Figure 7. The target movement tracks of scenes A, B, and C.
Sensors 23 00522 g007
Figure 8. The background suppression effect of SPB when the target SNR is minimum and maximum of scenes A, B, and C.
Figure 8. The background suppression effect of SPB when the target SNR is minimum and maximum of scenes A, B, and C.
Sensors 23 00522 g008
Figure 9. The background suppression effect of SPB when the target velocity is minimum and maximum of scenes A, B, and C.
Figure 9. The background suppression effect of SPB when the target velocity is minimum and maximum of scenes A, B, and C.
Sensors 23 00522 g009
Figure 10. The background suppression effect of SPB under the different motion trajectories of the targets of scenes A, B, and C.
Figure 10. The background suppression effect of SPB under the different motion trajectories of the targets of scenes A, B, and C.
Sensors 23 00522 g010
Figure 11. Whitening effect of the residual noise.
Figure 11. Whitening effect of the residual noise.
Sensors 23 00522 g011
Figure 12. The visual effect of background suppresses of the six algorithms for scene A.
Figure 12. The visual effect of background suppresses of the six algorithms for scene A.
Sensors 23 00522 g012
Figure 13. The visual effect of background suppresses of the six algorithms for scene B.
Figure 13. The visual effect of background suppresses of the six algorithms for scene B.
Sensors 23 00522 g013
Figure 14. The visual effect of background suppresses of the six algorithms for scene C.
Figure 14. The visual effect of background suppresses of the six algorithms for scene C.
Sensors 23 00522 g014
Table 1. The data of the sparse factors and their corresponding regional scale.
Table 1. The data of the sparse factors and their corresponding regional scale.
Scales24720110191817161514131
SF0.44810.47880.49190.49030.48930.48880.48880.48930.49020.4921
Scales211917151311975
SF0.49470.49520.49560.49590.49620.49650.49650.49670.4965
Table 2. The parameters of 6 algorithms.
Table 2. The parameters of 6 algorithms.
AlgorithmsParameters
SPBLocal region size: 3 a × 3 a , local block size: a × a , λ = 1 / max ( size ( D ) ) μ = 0.25 / mean ( D ) ,   φ = 10 7 . Here, a is the target scale.
PCP λ = 1 / max ( size ( D ) ) , μ = 0.25 / mean ( D ) ,   φ = 10 7
FPCP λ = 1 / max ( size ( D ) ) , lambdaFactor = 1.0; initial rank is r = 1, rankThreshold = 0.01
OSTD λ 1 = 1 / max ( size ( D ) ) , λ 2 = 10 λ 1
IPIPatch size: 50 × 50 ,   sliding   step : 10 ,   λ = 1 / min ( size ( D ) ) , φ = 10 7
PSTNNPatch size: 40 × 40 , sliding step:40, λ = 0.6 / max ( size ( D ) ) ,   φ = 10 7
Table 3. The energy, velocity, and trajectory characteristics of the target.
Table 3. The energy, velocity, and trajectory characteristics of the target.
ScenesFramesSizes SNR in (dB) Velocities (Pixels/Frame)Trajectories
MinMaxMeanMinMaxMean
A298250 × 250−0.28294.57803.0237010.0369Close to an inverted V.
B487120 × 120−1.73091.0886−0.2066010.0021Swinging near the original position
C178250 × 250−12.082011.14232.8393020.3599Reciprocating moves with a complex trajectory
Table 4. The SNRG and BSF values of SPB algorithm.
Table 4. The SNRG and BSF values of SPB algorithm.
Scenes BSF SNRG
MinMaxMeanMinMaxMean
A1.99164.38972.80611.85304.76223.0103
B1.27752.10211.70394.15736.88825.5784
C8.608437.711313.84592.099125.52066.6343
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Li, B.; Xu, Z.; Zhang, J.; Zhao, Q. Single-Pixel Background Modeling Algorithm for Strong Sky Scenes Based on Local Region Spatial Bases. Sensors 2023, 23, 522. https://doi.org/10.3390/s23010522

AMA Style

Li B, Xu Z, Zhang J, Zhao Q. Single-Pixel Background Modeling Algorithm for Strong Sky Scenes Based on Local Region Spatial Bases. Sensors. 2023; 23(1):522. https://doi.org/10.3390/s23010522

Chicago/Turabian Style

Li, Biao, Zhiyong Xu, Jianlin Zhang, and Quanyou Zhao. 2023. "Single-Pixel Background Modeling Algorithm for Strong Sky Scenes Based on Local Region Spatial Bases" Sensors 23, no. 1: 522. https://doi.org/10.3390/s23010522

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop