Next Article in Journal
Highly Coupled Seven-Core Fiber for Ratiometric Anti-Phase Sensing
Previous Article in Journal
An Unsupervised Machine Learning Approach for Monitoring Data Fusion and Health Indicator Construction
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Dim and Small Target Detection with a Combined New Norm and Self-Attention Mechanism of Low-Rank Sparse Inversion

1
Institute of Optics and Electronics, Chinese Academy of Sciences, Chengdu 610209, China
2
Key Laboratory on Adaptive Optics, Chinese Academy of Sciences, Chengdu 610209, China
3
School of Automation, Guangxi University of Science and Technology, Liuzhou 545006, China
4
Guangxi Collaborative Innovation Centre for Earthmoving Machinery, Guangxi University of Science and Technology, Liuzhou 545006, China
*
Author to whom correspondence should be addressed.
Sensors 2023, 23(16), 7240; https://doi.org/10.3390/s23167240
Submission received: 25 June 2023 / Revised: 11 August 2023 / Accepted: 15 August 2023 / Published: 18 August 2023
(This article belongs to the Section Remote Sensors)

Abstract

:
Methods for detecting small infrared targets in complex scenes are widely utilized across various domains. Traditional methods have drawbacks such as a poor clutter suppression ability and a high number of edge residuals in the detection results in complex scenes. To address these issues, we propose a method based on a joint new norm and self-attention mechanism of low-rank sparse inversion. Firstly, we propose a new tensor nuclear norm based on linear transformation, which globally constrains the low-rank characteristics of the image background and makes full use of the structural information among tensor slices to better approximate the rank of the non-convex tensor, thus achieving effective background suppression. Secondly, we construct a self-attention mechanism in order to constrain the sparse characteristics of the target, which further eliminates any edge residuals in the detection results by transforming the local feature information into a weight matrix to further constrain the target component. Finally, we use the alternating direction multiplier method to decompose the newly reconstructed objective function and introduce a reweighted strategy to accelerate the convergence speed of the model. The average values of the three evaluation metrics, SSIM, BSF, and SNR, for the algorithm proposed in this paper are 0.9997, 467.23, and 11.72, respectively. Meanwhile, the proposed detection method obtains a higher detection rate compared with other algorithms under the same false alarm rate.

1. Introduction

The weak target detection algorithm in infrared is an important technology for infrared warning and guidance. In real-world scenarios, the imaging distance of the target is far, the imaging area is small, and the radiation energy is low, which makes the target easily affected by noise and clutter interference. Therefore, how to separate weak targets from complex backgrounds is a challenging research topic, and many scholars have conducted in-depth research on this issue [1,2].
Traditional spatial–temporal filtering methods are one of the mainstream weak target detection algorithms in infrared, which separate the target and the background by local feature information of the image in the spatial–temporal domain. There are two methods of temporal filtering: single-frame detection algorithms and multi-frame detection algorithms. Multi-frame detection algorithms are mainly three-dimensional matching filtering algorithm, projection transformation algorithm, dynamic programming algorithm, pipeline filtering algorithm, etc. Common single-frame detection algorithms include top-hat filtering algorithm, gradient reciprocal filtering algorithm, etc. The principle of three-dimensional matching filtering is to set a three-dimensional filter on all possible motion trajectories of the target, and then analyze and judge the collection results of the filter. The three-dimensional filter with the highest signal-to-noise ratio is determined as the weak target and motion trajectory. However, for motion-mismatched targets, multiple velocity filters need to be set to capture the target trajectory, which leads to an increase in the computational complexity of the detection algorithm [3]. The projection transformation projects multidimensional data into a low-dimensional space, and obtain the motion trajectory of the target from the low-dimensional data. Since this algorithm appears to reduce the dimensionality of the data during the calculation, the data of the samples to be processed and the suspicious target signals are greatly reduced, which simplifies the computation. Compared with the multidimensional matching filtering, reduced computation is accompanied by poorer target detection result [4,5]. In the study of small target detection, Wang et al. [6] first used feature triangles to perform registration on infrared images and reduced the computation by using the star point coordinate matrix method. Then, the maximum value projection transformation method was used to process the sequence images and, therefore, obtain the result. The idea based on dynamic programming is to equivalently treat the accumulated energy of the detected small target on a certain trajectory as the decision function in the theory of dynamic programming. The motion range of the target at different stages is regarded as the decision space. Then, by recursive methods, the motion trajectory that can make the decision function achieve global optimization is found [7]. Guo et al. [8] proposed a dynamic programming method based on parallel computing, which performs a parallel search on the suspicious candidate target area to obtain the candidate target points. This method effectively improves the detection efficiency, but its detection performance decreases as the target signal-to-noise ratio gradually decreases. The principle of the pipeline filtering algorithm is that when the target signal is sampled at an appropriate frequency, the motion characteristics of the target signal in space are continuous in time, while random noise does not have continuity. Dong et al. [9] proposed a method that combines a modified visual attention model (VAM) with anti-jitter pipeline filtering algorithm. In this method, the best mode is adaptively selected to calculate the saliency map. Then, a local saliency singularity evaluation strategy is designed to automatically extract suspicious targets and suppress background clutter. Finally, the real target is distinguished by anti-vibration pipe filtering. This method can improve the efficiency of target detection in infrared imaging under different weather conditions. Although many improved pipeline methods have been proposed, it is difficult to achieve good detection performance when the motion of small targets has complex and varied characteristics. The top-hat filtering algorithm models the background of the image by using a structural element as a template for opening and closing operations on the image, and obtains the final difference map. Bai et al. [10] combined multi-scale operations with structural elements, which predict the image background by different scale structural elements. The top-hat filtering algorithm causes a loss of target energy intensity during opening and closing operations, resulting in poor detection in scenes with low target energy. Li et al. [11] established a correlation coefficient function based on the local feature information of the image, and introduced it into the gradient reciprocal filter algorithm, which can obtain an improved background estimation. Due to the complex and variable gray-scale characteristics of the image background, the filtering template of the gradient reciprocal filter algorithm cannot completely match the background features, resulting in a large number of edge residues in the detection results. Most real-world scenarios contain a lot of noise and clutter, and the noise signal is only slightly different from the weak target in terms of texture and energy. It is difficult to fully distinguish noise from the target based on only local feature information in the image. Therefore, traditional spatio-temporal filtering algorithms are easily interfered by noise and clutter, leading to low detection rates.
In the study of infrared faint target detection with deep learning, in order to effectively improve the generalization ability of the deep neural network detection model, Dai et al. proposed a special ACM detection model based on the construction of a high-quality annotation data set. There is an asymmetric upper and lower information modulation model for small infrared target detection [12]. This model not only realizes forward and backward information transfer to fusion, but also supplements a top-down information modulation module based on point-oriented channel attention, which enhances the fusion of target signals. The DNAnet network proposed by Li et al. completes the feature fusion of the target through the repeated use of image features [13], making full use of the weak information of the target. In addition, the model also includes the constructed dense interaction module and channel-attention module, which realizes the fusion of target features and the enhancement of adaptive features, and has achieved remarkable results in the detection of small infrared targets. In order to solve the problems of detection loss (MD) and false alarm (FA) more effectively, Wang et al. proposed an adversarial network consisting of two generators and one discriminator to extract target feature information [14]. This model distributes detection tasks to different training models, which improves the overall detection efficiency of the model. To sum up, the weak and small target detection model of deep learning has achieved remarkable results, but considering the limitations of the generalization ability of the model brought about by the multiple changes in the air and space detection environment, it is necessary to start from the data set construction and model construction. From the perspective of research, increase the practical application of deep learning detection models in the detection of small and weak infrared targets.
In recent years, scholars have paid more attention to the information about the entire image. They have transformed the problem of detecting targets into the matrix reconstruction problem through machine learning theories, and solved this problem using the overlapping direction multiplier method to obtain the detect result. Gao et al. [15] proposed an infrared patch-image (IPI) model that utilizes the correlation characteristics of non-local image background. Wang et al. [16] suppressed strong background edge contours by combining anisotropy and L1 norm on the basis of the IPI model. Xue et al. [17] suppressed strong background edge contours by combining anisotropy and the L1 norm. Xue et al. used the weighted L1 norm to constrain the target and added constraints on the noise using L1,1 and L2,1 norms to eliminate the background residue. However, in a complex background, the simple nuclear norm cannot fully constrain the background. These methods are unable to fully recover the background image, causing unsatisfactory detection results. Compared with two-dimensional data, high-dimensional data have more structural information, making it easier for us to explore hidden relationships between data. Therefore, Dai et al. [18] proposed an infrared patch tensor (IPT) model to extend the two-dimensional image matrix to three-dimensional, fully exploring the spatial correlation of non-local areas of the image, and using the prior information to constrain sparse targets through re-weighting. Zhang et al. [19] further suppressed the background using partial sum of the tensor nuclear norm (PSTNN), and then used the corner descriptor function and the background descriptor function to obtain prior information of the image to jointly constrain the target using the L1 norm. Wang et al. [20] constructed a new spatio-temporal tensor model, non-overlapping patch spatial–temporal tensor (NPSTT) model, based on the IPT model to avoid redundant information. The method also obtains better target constraint results with the help of tensor upper bound nuclear norm. Kong et al. [21] approximated the non-convex tensor fiber rank using Log tensor fibered nuclear norm (LogTFNN) on the basis of the IPT model, and then used hypertotal variation (HTV) to suppress noise. Cao et al. [22] defined a mode-k1k2 extension tensor tubal rank (METTR) and fully preserve the spatial correlation of non-local areas in images by METTR-norm.
During the actual process of target detection, the background is complex, often accompanied by strong edge contours and strong noise. Merely using the information about the entire image cannot completely suppress the strong edges, leading to residual background in the detection results. The IPT model constructs background weight function based on the characteristics of the structural tensor, but ignores the information of the target, leading to excessive contraction of the target. The PSTTN model proposes a structural tensor descriptor function for both the background and the corner points based on the RIPT model, while retaining the prior information of both the background and the target. However, because the prior information related to the background is only based on the maximum value of the eigenvalues, it cannot completely suppress the contours of strong edge contours, resulting in some edge contours being retained in the final local feature prior information.
In summary, in order to solve the limitations of existing detection models, we propose a low-rank sparse inversion method for detecting small targets by combining a new norm with self-attention mechanisms.
  • To better constrain the low-rank tensor, we approximate the tensor rank using a new tensor nuclear norm that is non-convex. The new tensor nuclear norm preserves the structural information between the various slices of the tensor and can accurately recover low-rank backgrounds under certain interference. Moreover, the high-order tensor is reduced during the operation, reducing the computational complexity of the algorithm.
  • We construct a self-attention mechanism model to constrain the sparse characteristics of the target. This model obtains local feature information of the image to suppress strong edge contours and strong noise and transforms the local feature information into a weight matrix to introduce the constraint of the sparse component of the infrared small target detection model.
  • We design an overlapping direction multiplier method to solve the infrared small target detection model and obtain the target image. In the solving process, we use a re-weighting strategy to accelerate the convergence speed and accuracy of the model.

2. Introduction to the Principle of Tensor Singular Value Decomposition (T-SVD) [23]

We define a invertible linear transformation L : R n 1 × n 2 × n 3 R n 1 × n 2 × n 3 , for any invertible transformation L, we have L ( 0 ) = L 1 ( 0 ) = 0 , and obtaining A ¯ by linearly transforming A along the third dimension, as shown in:
A ¯ = L ( A ) = A × 3 L .
The symbol × 3 in the formula represents the mode-3 product. The linear transformation L R n 3 × n 3 is any invertible matrix. Under any invertible linear transformation L, the singular value decomposition of A ¯ R n 1 × n 2 × n 3 is given by:
U ¯ i , S ¯ i , V ¯ i = S V D ( A ¯ i ) .
In the above equation U ¯ i , S ¯ i , V ¯ i , B ¯ i R n 1 × n 2 denote the i-th i = 1 n 3 frontal slice of U , S , V , A R n 1 × n 2 × n 3 , respectively. Where U R n 2 × n 2 × n 3 and V R n 2 × n 2 × n 3 are the orthogonal tensor, S R n 2 × n 2 × n 3 is the diagonal tensor. The inverse mapping of U ¯ i , S ¯ i , V ¯ i are given by
U = L 1 U ¯ = U ¯ × 3 L 1 S = L 1 S ¯ = S ¯ × 3 L 1 V = L 1 V ¯ = V ¯ × 3 L 1 .
Figure 1 and Algorithm 1 demonstrate the process of tensor singular value decomposition. As S and S ¯ are F-diagonal tensors, we can define the tensor tube rank using their non-zero tube counts [24].
Algorithm 1 T-SVD under linear transform L
Input: A R n 1 × n 2 × n 3 and invertible linear transform L;
Step 1. Compute A ¯ = L ( A ) ;
Step 2. Frontal slicing of image A ¯ and singular value decomposition of each frontal slice combined with Equation (2) to obtain U ¯ , S ¯ and V ¯
Step 3. The inverse linear transformation of the three tensors is calculated by Equation (3) to obtain U , S and V
Output: U , S , V

3. NNSAM Model

3.1. Constructing New Tensor Nuclear Norm Constrained Background Low Rank Properties

In the absence of noise, the infrared weak small target detection model can be transformed into a Tensor Robust Principal Component Analysis (TRPCA) model, which expression is:
arg min A , E A + λ E 1 s . t . D = A + E ,
where D , A , E represent the original image tensor of the sequence, the background image tensor of the sequence, and the target image tensor of the sequence, respectively. λ is a positive parameter. Considering that the TRPCA model mentioned above is disrupted by the strong edge contours and noise in the environment with many strong noise points, causing the bad detection result, we have made improvements to the above model. Firstly, we introduce the tensor nuclear norm based on linear transformation to replace the original tensor nuclear norm, which constrains the image background globally. We use Algorithm 1 based on linear transformations of tensor singular value decomposition and tensor t-product to define tensor spectral norm, and then derive the new tensor nuclear norm by the property of tensor spectral norm as the dual norm of tensor nuclear norm. The definition of the new tensor nuclear norm is as follows [25,26]:
A = 1 l A ¯ ,
where l > 0 is a constant. A R n 1 × n 2 × n 3 , A ¯ R n 1 n 3 × n 2 n 3 are the block diagonal matrix, and its i-th block on the diagonal is the i-th slice A ¯ ( i ) of A ¯ . The new tensor nuclear norm is able to approximate non-convex tensor rank well, while preserving the structural information between each tensor slice, and achieving dimensionality reduction for high-order tensors with low time complexity and high computational efficiency. Therefore, this paper adopts the new tensor nuclear norm to build the detection model. The model expression of the new tensor nuclear norm is as follows,
arg min A , E 1 l A ¯ + λ E 1 s . t . D = A + E ,

3.2. Constructing Self-Attention Mechanisms to Constrain the Sparsity of the Target Feature

The detection model constructed based on non-local feature information of images can well recognize complex edge contours in the background of images. However, it cannot completely suppress some strong edge contours and strong noise points, lead to residual edge contours and noise in the detection results. Therefore, many scholars use traditional filtering or human visual saliency methods to obtain local feature information of images, and then convert the local feature information into a weight matrix and introduce it into the infrared weak small target detection model constructed by non-local feature information. Scholars have found that in infrared images containing weak targets, the high-order singular values obtained through SVD decomposition usually reflect the noise in the image, the middle-order singular values reflect the targets in the infrared image, and the low-order singular values reflect the background in the infrared image [27]. Based on these characteristics, the singular value decomposition of the image matrix in this article can achieve background suppression. Assuming that the rank of matrix X R m × n is r min ( m , n ) , the singular value decomposition of matrix t can be expressed as follows
X = U V T .
In the above equation, U R m × m and V R n × n are orthogonal matrices, which are, respectively, called left and right singular value vector matrices, and R m × n is the singular value matrix. Since the high-order and low-order singular values reflect the noise and background of the infrared image, respectively, the target image can be obtained by directly truncating the high-order and low-order singular values of the singular value matrix through the double threshold method [27]. Assuming that the truncated singular value matrix is , the SVD filtering result can be calculated by the following equation
X = U V T .
The result of the singular value filtering obtained through the above thoughts is shown in Figure 2b, and its global three-dimensional diagram is shown in Figure 2c.
From Figure 2, we can see that using the double threshold rule alone cannot accurately determine the high-order and low-order singular values of the infrared image matrix, which leads to the retention of a large number of edge contours and noise residues in the detection results. Therefore, we have developed a target sparse constraint algorithm based on self-attention mechanism to further suppress the edge contours and noise residues in the SVD filtering results.
In cognitive science, due to the bottleneck of information processing, humans often focus on important information while ignoring unimportant information, a mechanism called the self-attention mechanism. The Self-attention mechanism usually focuses on important information with high weights and ignores irrelevant information with low weights.
Moreover, the weights can be continuously adjusted to obtain important information under different circumstances [28]. The core formula of the self-attention mechanism is:
A t t e n t i o n ( Q , K , V ) = S o f t m a x ( Q K T d k ) V ,
where Q , K , V represent the Query, Key, and Value matrices, respectively. d k denotes the dimension size of matrix Key, and S o f t m a x is the normalization function. From analyzing the information of the filtered image, it can be seen that there are only a few edge contours and a small amount of noise residues present in the image after filtering. We replace Q , K , V with non-overlapping infrared block image matrices X t , the transpose of non-overlapping infrared block image matrices X t T , and non-overlapping infrared block image matrices X t , respectively. Since the highest correlation between X t and X t T is in the target area, we can achieve target extraction and suppression of edge contours through the idea of self-attention mechanism. The formula established in this paper based on the self-attention mechanism for non-overlapping infrared image blocks is as follows:
A t t e n t i o n ( X t ) = S o f t m a x ( X t X t T ) X t .
where X t denotes the non-overlapping infrared block image matrix, * and ⊙ denote the multiplication of the matrix and the hadamard product of the matrix, respectively, and S o f t m a x ( ) is the normalized function. We found that the background residues in the filtered results always appear as stripes, while the targets appear as dots. Using block image matrices X t X t T , we can enhance the energy of the targets and suppress edge contours. We then normalize the results of the block matrix X t X t T to use them as weight matrices to constrain the original block image matrix, thereby suppressing the background edge residues. The process of constructing a self-attention mechanism to constrain the sparse characteristics of the target is shown in Algorithm 2.
Algorithm 2 Target sparse constraint algorithm based on self-attention mechanism
Input: The result X R m × n of SVD filtering;
Step 1. The image matrix X is divided into block image matrices X p using a sliding window;
Step 2. The operation result X p of the self-attention mechanism is obtained by calculating each block image matrix using Equation (10);
Step 3. The final result matrix X is obtained by concatenating the operation result block matrices X p of the self-attention mechanism;
Output: The result X of the target sparse constraint algorithm after applying the self-attention mechanism.
We construct the tensor W x R m × n × t by stacking together the results X of t-frame sequence images after the self-attentive mechanism, and then take the inverse of all elements of the tensor W x to obtain W l . To accelerate the convergence speed and improve the accuracy of the model, paper [29] proposed a re-weighting strategy which adds a coefficient weight in the model. The expression for this coefficient weight is:
W s w k + 1 = c E k + v ,
where c represents a positive constant and vs. represents a very small constant to prevent the denominator from being 0. By combining the re-weighting strategy with local feature tensors, we obtain the final local feature tensor as follows:
W = W l W s w ,
the objective function of the final model is:
arg min A , E 1 A ¯ + λ W E 1 s . t . D = A + E .

3.3. Solving of the Model

For the model represented by Equation (13), we usually use the overlapping direction multiplier method to solve it [30]. The augmented Lagrangian function for Equation (13) is:
L ( A , E , Y , μ ) = A + λ W E 1 + Y , A + E D + μ 2 A + E D F 2 ,
where Y denotes the Lagrangian multiplier, denotes the inner product of the tensor, and μ > 0 denotes the penalty factor. We divide the augmented Lagrangian function into several sub-problems for solving. With regard to the sub-problem of A , in k + 1 iterations,
A k + 1 = arg min A A + μ k 2 A + E k D + Y k μ k F 2 .
For the above sub-problem, we use the Tensor Singular Value Thresholding (T-SVT) operator [25] for solving. Let D E k Y k μ k = U L S L V T be tensor SVD of D E k Y k μ k ,
A k + 1 = U L S 1 L V T ,
where S 1 = L 1 L S 1 + , here t + = max t , 0 . With regard to the sub-problem of E , in k + 1 iterations,
E k + 1 = arg min E λ E W 1 + μ k 2 A k + 1 + E D + Y k μ k F 2 ,
For the above sub-problem, we use the threshold shrinkage operator for solving:
E k + 1 = S λ W k μ k D A k + 1 Y k μ k .
where S τ X = s i g n ( X ) max ( X τ , 0 ) , and s i g n ( ) denotes the sign function. With regard to the sub-problem of Y , μ , in k + 1 iterations,
Y k + 1 = Y k + μ k ( A k + 1 + E k + 1 D ) μ k + 1 = min ( ρ μ k , μ max ) ,
where ρ , μ max are given constants. Finally, Algorithm 3 shows the entire process of the ADMM algorithm, and Figure 3 shows the entire process of the model.
Algorithm 3 Solve proposed model by ADMM
Input: tensor data D , W , parameter λ ;
Initialization: A 0 = Y 0 = E 0 = 0 , max _ i t e r = 500 , ρ = 1.1 , μ 0 = 1 × 10 3 , μ max = 1 × 10 10 , δ = 1 × 10 8 , k = 0 ;
while not converge do
Step 1. Update A k + 1 by Equation (16);
Step 2. Update E k + 1 by Equation (17);
Step 3. Update Y k + 1 , μ k + 1 by Equation (19);
Step 4. Check the convergence conditions:
A k + 1 A k δ , E k + 1 E k δ , A k + 1 + E k D δ , k max _ i t e r
Step 5.  k : k = k + 1
end while
Output: B , E

4. Experiment and Analysis

In this section, we will further showing the advantages of our model from target detection ability, robustness to noise, and adaptability to multiple scenes. We will compare our algorithm with nine advanced algorithms in the weak object detection field on six sequences representing image detection effects [31]. We will also evaluate our algorithm using performance metrics such as SSIM, BSF, SNR, and ROC curves [32]. Figure 4 shows the six sequential images used in the experiments.

4.1. Parametric Analysis

In our proposed methods, the size of image patches in the sparse target-constrained algorithm based on the self-attention mechanism, as well as the regularization factor λ and μ in the infrared weak small target detection model, have a significant impact on the algorithm proposed in this paper. Therefore, in order to demonstrate the impact of parameters on the algorithm, we conducted separate analysis on these three parameters. The analysis results are as follows.

4.1.1. Patch Size

In the background-constrained algorithm based on self-attention mechanism, the size of image patches has a significant impact on the constraint effect of edge contours. Moreover, the size of image patches is related to the size of the target. When the target is larger, the size of image patches should also be larger. Figure 5a shows the relationship curve between the size of image patches and the global signal-to-noise ratio under sequence (a). In this paper, the size of image patches is 5.

4.1.2. Regularization Factor λ

λ is a regularization factor used to balance the sparse target components and low-rank background components in the model. When λ is too small, it can lead to incomplete separation of the low-rank components, resulting in more background edge contours in the detection results. When λ is too large, it can cause the targets to shrink excessively, affecting the detection performance of the model. Therefore, the value of λ has a significant impact on the model, and finding the best balance is a challenge. Figure 5b shows the relationship curve between the value of λ and the detection rate under sequence (a). In this paper, the regularization factor λ is 0.1.

4.1.3. Penalty Factor μ

The penalty factor μ is used to balance the T-SVT operator and the threshold shrinkage operator. The penalty factor can affect the convergence speed and solution accuracy of the model. When the penalty factor is too small, it will lead to excessive shrinkage of the target, resulting in a low detection rate. When the penalty factor is too large, it will cause inaccurate solution results, leading to false positives and false negatives. Figure 5c shows the relationship curve between the value of μ and the detection rate under sequence (a). In this paper, the penalty factor μ is 3 × 10 4 .

4.2. Results of Background Suppression Experiments

To demonstrate the sparsity constraint effect of the self-attention mechanism algorithm proposed in this paper, we conducted experiments on the eight scenes and the experiments result are shown in Figure 6. The results suggested that the self-attention mechanism-based sparsity constraint algorithm can effectively suppress edge residual and noise in the background by dividing the image matrix into blocks and performing the operation in Equation (10), achieving good background suppression effect. Figure 6a,c,e, respectively, represent the original image, SVD filtering result and the result after being constrained by the self-attention mechanism, while Figure 6b,d,f, respectively, represent the three-dimensional view of the original image, the three-dimensional view of the SVD filtering result, and the three-dimensional view of the result after being constrained by the self-attention mechanism.

4.3. Visual Comparison with Baselines

We compare the visual detection results of the algorithm in this paper with nine algorithms in eight different scenarios and the result shown in Figure 7, Figure 8, Figure 9, Figure 10, Figure 11, Figure 12, Figure 13 and Figure 14. The Aadwcdd [33] algorithm uses the cumulative directional derivative to weight the mean absolute gray difference and eliminate the disadvantage of the decrease in target detection rate when facing strong structured edge backgrounds. However, we can find from Figure 7, Figure 8, Figure 9, Figure 10, Figure 11, Figure 12, Figure 13 and Figure 14 that this algorithm cannot completely suppress the contours and noise in the complex backgrounds, resulting in more background contours and noise in the detection results. The ASTTV [34] algorithm proposes a non-convex approximation tensor rank method that adaptively distributes singular value weights with the help of the Laplace function, and further constrains the background by combining asymmetric spatial–temporal total variation (ASTTV) as a regularization term. However, we can find from Figure 7, Figure 8, Figure 9, Figure 10, Figure 11, Figure 12, Figure 13 and Figure 14 that this algorithm cannot effectively suppress the background when facing strong edge contours and large areas of strong background only by using the global feature information of images. The DPSRG [35] algorithm extracts targets by peak density search and maximum area growth method and can effectively suppress the background. However, we can find from Figure 11 that the DPSRG algorithm cannot separate the target when facing low-energy backgrounds that are submerged in contours, resulting in poorer detection. The GST [36] algorithm models small targets using the second conjugate symmetric derivative of the two-dimensional Gaussian on the directional field. We can find from Figure 9 and Figure 10 that this algorithm cannot detect target points under low-energy target scenes. The FKRW [37] algorithm uses the local contrast descriptor (NLCD) to achieve clutter suppression and target enhancement. We can see from Figure 7, Figure 8, Figure 9, Figure 10, Figure 11, Figure 12, Figure 13 and Figure 14 that NLCD cannot completely inhibit the edge contours under strong edge contours and strong noise conditions. This leads to the appearance of edge residues in the detection results. The PSTNN [19] algorithm and the VNTFR [21] algorithm combine the local prior information and global features of the image to achieve weak target detection. We can find from Figure 8, Figure 9, Figure 10, Figure 11, Figure 12, Figure 13 and Figure 14 that there are a small number of edge contours in the difference map, which is due to the incomplete suppression of the background by the local feature prior information of the PSTNN and VNTFR algorithms, making it impossible to completely constrain the target in the detection model, resulting in a small amount of background residues in the detection results. The SRWS [38] algorithm achieves background suppression by overlapping edge information (OEI). However, we can see from Figure 9 that there are still many edge residues in the detection results under the low-energy target condition. The principle of the top-hat algorithm uses a structural element as a template to perform opening and closing operations on the image to obtain the background model. However, simple opening and closing operations cannot suppress complex backgrounds and noise when facing strong edge contours and noise in complex scenes, leading to more background residues in the detection results. From Figure 7, Figure 8, Figure 9, Figure 10, Figure 11, Figure 12, Figure 13 and Figure 14, we can find that our proposed algorithm has certain advancements compared with the remaining nine algorithms. From the 3D view, we can find that our algorithm completely suppresses the background, eliminates false alarms caused by noise, indicating that our algorithm has excellent noise robustness and background suppression ability.

4.4. Quantitative Evaluation

In this section, we quantitatively evaluated our algorithm and the other nine comparative algorithms using SSIM, BSF, SNR, and ROC curves. The results of the calculation of SSIM, BSF, and SNR for our algorithm and nine classic algorithms under six sequential images are shown in Table 1 and Table 2. The bold text in Table 1 and Table 2 represents the maximum value. The parameters used in the experiment were the best values in the corresponding literature. From Table 1 and Table 2, our method can achieve better background estimation compared with other nine advanced algorithms. Moreover, the SNR of our method is the highest in Scene (d), Scene (e), and Scene (f), while in the other scenes, our algorithm can rank in the top three. From Figure 7, Figure 8, Figure 9, Figure 10, Figure 11 and Figure 12, the detection results of SRWS, Aadwcdd, and FKRW have strong target energy, so they have larger SNR values. However, our method is superior to these three algorithms in terms of noise suppression.
We also plotted the ROC curves of each algorithm under the six sequential images, as shown in Figure 15. From Figure 15a–e, we can see that the ROC curve of our algorithm is included in the curves of other algorithms, and the area under the curve is the largest. Therefore, compared with other algorithms, our algorithm has the perfect results.

4.5. Algorithm Complexity and Computational Time

Real-time performance and accuracy are two important performance indicators for infrared weak small target detection systems. However, for weak small target detection algorithms, balancing real-time performance and accuracy has become a significant challenge. Traditional spatio-temporal filtering methods have a great advantage in real-time performance due to their simple computation process. However, they often need to improve the accuracy of detecting targets in complex environments with strong edge contours and noise. On the other hand, weak small target detection algorithms based on low-rank and sparse theory have better accuracy in different complex environments. However, due to the use of alternating direction method of multipliers to solve the detection model and complex operations such as SVD decomposition in the solving process, these algorithms often perform poorly in terms of real-time performance. In this paper, a new tensor nuclear norm based on linear transformation is proposed. Through linear transformation in the process of solving this nuclear norm, the original three-dimensional tensor is transformed into a two-dimensional matrix, achieving data dimensionality reduction and effectively reducing the running time of the algorithm. For a sequence tensor image of size n 1 × n 2 × n 3 , the algorithm proposed in this paper has a complexity of O n 1 n 2 n 3 2 + n 1 n 2 2 n 3 , where n 1 = max ( n 1 , n 2 ) and n 2 = min ( n 1 , n 2 ) . The running time of the proposed algorithm and nine other algorithms on six sequences is shown in Table 3.

5. Conclusions

In response to the shortcomings of existing weak target algorithms in an intricate environment, such as low noise robustness, weak background suppression capability, and a high false alarm rate, we propose a low-rank sparse inversion weak target detection method using a joint new paradigm and attention mechanism. In this method, we use an improved tensor nuclear norm to explore the hidden structural information between various modes of the tensor, which can better approximate the tensor rank while reducing time and computational complexity compared to PSTNN, SNN, TNN, etc. We also establish a target sparse constraint algorithm based on self-attention mechanism as a prior information constraint model for target components to achieve better background and target separation and reduce noise and residual background in detection results. Finally, we use optimization algorithms based on ADMM and linear transformation T-SVD to solve the detection model. Based on extensive experimental results in various scenarios, our method demonstrates robustness and outperforms other algorithms in evaluation metrics such as BSF, SSIM, SNR, ROC, etc.
In the sparse target-constrained algorithm based on self-attention mechanism, the image patch size used in this paper is fixed, which has certain limitations when dealing with targets of different sizes. Therefore, if a method can be established to adaptively adjust the image patch size, it will lead to more exciting results. At the same time, we will explore the application of attention mechanism in the time domain and establish a target sparse constraint method based on image temporal–spatial information, which will be integrated into TRPCA model to further improve target detection performance.

Author Contributions

Conceptualization, L.M. and A.W.; methodology, L.M., A.W. and X.F.; software, L.M. and A.W.; validation, L.M.; formal analysis, L.M., A.W., X.F. and J.L.; investigation, L.M., A.W., X.F. and J.L.; resources, A.W. and X.F.; data curation, L.M. and F.L.; writing—original draft preparation, L.M. and A.W.; writing—review and editing, L.M., A.W., X.F., F.L. and J.L. All authors have read and agreed to the published version of the manuscript.

Funding

This work was partly supported by the National Natural Science Foundation of China under Grant (62261004, 62001129).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The abbreviations are used in this manuscript as following:
AadwcddAbsolute average difference weighted by cumulative directional derivatives
ASTTVAsymmetric Spatial–Temporal Total Variation
DPSRGDensity Peaks Searching and Maximum-Gray Region Growing
GSTGeneralized Structure Tensor
FKRWFacet Kernel and Random Walker
PSTNNPartial sum of tensor nuclear norm
VNTFRvia Nonconvex Tensor Fibered Rank
SRWSSelf-Regularized Weighted Sparse
SSIMStructural Similarity Image
SNRSignal-to-Noise Ratio
BSFBackground Suppression Factor

References

  1. Xia, C.; Li, X.; Zhao, L. Infrared Small Target Detection via Modified Random Walks. Remote Sens. 2018, 10, 2004. [Google Scholar] [CrossRef]
  2. Dai, T.; Zhang, K.; Yin, D. An end-to-end neural network for mobile phone detection in driving scenarios. Opto-Electron. Eng. 2021, 48, 200325. [Google Scholar] [CrossRef]
  3. Reed, I.S.; Gagliardi, R.M.; Stotts, L.B. Optical moving target detection with 3-D matched filtering. IEEE Trans. Aerosp. Electron. Syst. 2002, 24, 327–336. [Google Scholar] [CrossRef]
  4. Chu, P.L. Optimal projection for multidimensional signal detection. IEEE Trans. Acoust. Speech Signal Process. 1988, 36, 775–786. [Google Scholar] [CrossRef]
  5. Gong, J.; Zhang, Y.; Hou, Q.; He, W.; Guan, H.; Zhang, W. Background suppression for cloud clutter using temporal difference projection. Infrared Phys. Technol. 2014, 64, 66–72. [Google Scholar] [CrossRef]
  6. Qin, H.; Han, J.; Yan, X.; Li, J.; Zhou, H.; Zong, J.; Wang, B.; Zeng, Q. Multiscale random projection based background suppression of infrared small target image. Infrared Phys. Technol. 2015, 73, 255–262. [Google Scholar] [CrossRef]
  7. Buzzi, S.; Lops, M.; Venturino, L. Track-before-detect procedures for early detection of moving target from airborne radars. IEEE Trans. Aerosp. Electron. Syst. 2005, 41, 937–954. [Google Scholar] [CrossRef]
  8. Guo, Q.; Li, Z.; Song, W.; Fu, W. Parallel Computing Based Dynamic Programming Algorithm of Track-before-Detect. Symmetry 2018, 11, 29. [Google Scholar] [CrossRef]
  9. Dong, L.; Wang, B.; Ming, Z.; Xu, W. Robust Infrared Maritime Target Detection Based on Visual Attention and Spatiotemporal Filtering. IEEE Trans. Geosci. Remote Sens. 2017, 55, 3037–3050. [Google Scholar] [CrossRef]
  10. Bai, X.; Zhou, F. Analysis of new top-hat transformation and the application for infrared dim small target detection. Pattern Recognit. 2010, 43, 2145–2156. [Google Scholar] [CrossRef]
  11. Li, B.; Xu, Z.; Wang, C. Background suppression for infrared dim small target scene based on adaptive gradient reciprocal filtering. Opto-Electron. Eng. 2021, 48, 210122. [Google Scholar] [CrossRef]
  12. Dai, Y.; Wu, Y.; Zhou, F.; Barnard, K. Asymmetric contextual modulation for infrared small target detection. arXiv 2020, arXiv:2009.14530. [Google Scholar]
  13. Li, B.; Xiao, C.; Wang, L.; Wang, Y. Dense Nested Attention Network for Infrared Small Target Detection. IEEE Trans. Image Process. 2023, 32, 1745–1758. [Google Scholar] [CrossRef] [PubMed]
  14. Wang, H.; Zhou, L.; Wang, L. Miss Detection vs. False Alarm: Adversarial Learning for Small Object Segmentation in Infrared Images. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Republic of Korea, 27 October–2 November 2019; Volume 1, pp. 8508–8517. [Google Scholar]
  15. Gao, C.; Meng, D.; Yang, Y.; Wang, Y.; Zhou, X.; Hauptmann, A. Infrared Patch-Image Model for Small Target Detection in a Single Image. IEEE Trans. Image Process. 2013, 22, 4996–5009. [Google Scholar] [CrossRef] [PubMed]
  16. Wang, X.; Peng, Z.; Kong, D.; Zhang, P.; He, Y. Infrared dim target detection based on total variation regularization and principal component pursuit. Image Vis. Comput. 2017, 63, 1–9. [Google Scholar] [CrossRef]
  17. Xue, W.; Qi, J.; Shao, G.; Xiao, Z.; Zhong, P. Low-Rank Approximation and Multiple Sparse Constraints Modelling for Infrared Low-Flying Fixed-Wing UAV Detection. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2021, 14, 4150–41661. [Google Scholar] [CrossRef]
  18. Dai, Y.; Wu, Y. Reweighted Infrared Patch-Tensor Model with Both Nonlocal and Local Priors for Single-Frame Small Target Detection. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2017, 10, 3752–3767. [Google Scholar] [CrossRef]
  19. Zhang, L.; Peng, Z. Infrared Small Target Detection Based on Partial Sum of the Tensor Nuclear Norm. Remote Sens. 2019, 11, 382. [Google Scholar] [CrossRef]
  20. Wang, G.; Tao, B.; Kong, X.; Peng, Z. Infrared Small Target Detection Using Nonoverlapping Patch Spatial–Temporal Tensor Factorization with Capped Nuclear Norm Regularization. IEEE Trans. Geosci. Remote Sens. 2021, 60, 5001417. [Google Scholar] [CrossRef]
  21. Kong, X.; Yang, C.; Cao, S.; Li, C.; Peng, Z. Infrared Small Target Detection via Nonconvex Tensor Fibered Rank Approximation. IEEE Trans. Geosci. Remote Sens. 2021, 60, 5000321. [Google Scholar] [CrossRef]
  22. Yang, J.; Zhao, X.; Ji, T.; Ma, T.; Huang, T. Low-rank tensor train for tensor robust principal component analysis. Appl. Math. Comput. 2020, 367, 124783. [Google Scholar] [CrossRef]
  23. Kolda, T. Tensor decompositions and applications. SIAM Rev. 2009, 51, 455–500. [Google Scholar] [CrossRef]
  24. Lu, C.; Feng, J.; Chen, Y.; Liu, W.; Lin, Z.; Yan, S. Tensor Robust Principal Component Analysis with a New Tensor Nuclear Norm. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 42, 925–938. [Google Scholar] [CrossRef] [PubMed]
  25. Lu, C.; Zhou, P. Exact Recovery of Tensor Robust Principal Component Analysis under Linear Transforms. arXiv 2019, arXiv:1907.08288. [Google Scholar]
  26. Lu, C.; Feng, J.; Chen, Y.; Liu, W.; Lin, Z.; Yan, S. Tensor Robust Principal Component Analysis: Exact Recovery of Corrupted Low-Rank Tensors via Convex Optimization; IEEE Computer Society: Washington, DC, USA, 2017. [Google Scholar]
  27. Hu, M.F.; Dong, W.J.; Wang, S.H.; Chen, Z.P. Singular Value Decomposition Band-Pass-Filter for Image Background Suppression and Denoising. Acta Electonica Sin. 2008, 36, 111. [Google Scholar]
  28. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, L.; Polosukhin, I. Attention Is All You Need. arXiv 2017, arXiv:1706.03762. [Google Scholar]
  29. Candès, E.J.; Wakin, M.B.; Boyd, S.P. Enhancing Sparsity by Reweighted L1 Minimization. J. Fourier Anal. Appl. 2008, 14, 877–905. [Google Scholar] [CrossRef]
  30. Lu, C.; Feng, J.; Yan, S.; Lin, Z. A Unified Alternating Direction Method of Multipliers by Majorization Minimization. IEEE Trans. Pattern Anal. Mach. Intell. 2018, 40, 527–541. [Google Scholar] [CrossRef]
  31. Hui, B.; Song, Z.; Fan, H. A dataset for infrared image dim-small aircraft target detection and tracking under ground/air background. Sci. Data Bank 2019, 5, 12. [Google Scholar]
  32. Wu, A.; Fan, X.; Chen, H.; Min, L.; Xu, Z. Infrared Dim and Small Target Detection Algorithm Combining Multiway Gradient Regularized Principal Component Decomposition Model. IEEE Access 2022, 10, 36057–36072. [Google Scholar] [CrossRef]
  33. Aghaziyarati, S.; Moradi, S.; Talebi, H. Small infrared target detection using absolute average difference weighted by cumulative directional derivatives. Infrared Phys. Technol. 2019, 101, 78–87. [Google Scholar] [CrossRef]
  34. Liu, T.; Yang, J.; Li, B.; Xiao, C.; Sun, Y.; Wang, Y.; An, W. Non-Convex Tensor Low-Rank Approximation for Infrared Small Target Detection. IEEE Trans. Geosci. Remote Sens. 2021, 60, 5614718. [Google Scholar]
  35. Huang, S.; Peng, Z.; Wang, Z.; Wang, X.; Li, M. Infrared Small Target Detection by Density Peaks Searching and Maximum-Gray Region Growing. IEEE Geosci. Remote Sens. Lett. 2019, 16, 1919–1923. [Google Scholar] [CrossRef]
  36. Gao, C.Q.; Tian, J.W.; Wang, P. Generalised-structure-tensor-based infrared small target detection. Electron. Lett. 2008, 44, 1349–1351. [Google Scholar] [CrossRef]
  37. Qin, Y.; Bruzzone, L.; Gao, C.; Li, B. Infrared Small Target Detection Based on Facet Kernel and Random Walker. IEEE Trans. Geosci. Remote Sens. 2019, 57, 7104–7118. [Google Scholar] [CrossRef]
  38. Tz, A.; Zp, A.; Hao, W.A.; Yh, A.; Cl, B.; Cy, A. Infrared small target detection via self-regularized weighted sparse model. Neurocomputing 2021, 420, 124–148. [Google Scholar]
Figure 1. An illustration of the t-SVD under linear transform L of n 1 × n 2 × n 3 .
Figure 1. An illustration of the t-SVD under linear transform L of n 1 × n 2 × n 3 .
Sensors 23 07240 g001
Figure 2. (ac) represent the original image, the result of SVD filtering, and the global three-dimensional diagram of the SVD filtering result, respectively.
Figure 2. (ac) represent the original image, the result of SVD filtering, and the global three-dimensional diagram of the SVD filtering result, respectively.
Sensors 23 07240 g002
Figure 3. Flow chart of the proposed model.
Figure 3. Flow chart of the proposed model.
Sensors 23 07240 g003
Figure 4. The representative image of the six sequential images. (af) represent the images of six different sequence scenes, respectively.
Figure 4. The representative image of the six sequential images. (af) represent the images of six different sequence scenes, respectively.
Sensors 23 07240 g004
Figure 5. (a) represents the relationship between image patch size and SNR, (b) represents the relationship between λ and detection rate, and (c) represents the relationship between μ and detection rate.
Figure 5. (a) represents the relationship between image patch size and SNR, (b) represents the relationship between λ and detection rate, and (c) represents the relationship between μ and detection rate.
Sensors 23 07240 g005
Figure 6. The result of target sparse constraint algorithm based on self-attention mechanism. (a,c,e) represent the original image, SVD filtering result and the result after being constrained by the self-attention mechanism, respectively. (b,d,f) represent the three-dimensional view of the original image, the three-dimensional view of the SVD filtering result, and the three-dimensional view of the result after being constrained by the self-attention mechanism, respectively.
Figure 6. The result of target sparse constraint algorithm based on self-attention mechanism. (a,c,e) represent the original image, SVD filtering result and the result after being constrained by the self-attention mechanism, respectively. (b,d,f) represent the three-dimensional view of the original image, the three-dimensional view of the SVD filtering result, and the three-dimensional view of the result after being constrained by the self-attention mechanism, respectively.
Sensors 23 07240 g006aSensors 23 07240 g006b
Figure 7. Visual detection effects of compared methods in scene 1.
Figure 7. Visual detection effects of compared methods in scene 1.
Sensors 23 07240 g007
Figure 8. Visual detection effects of compared methods in scene 2.
Figure 8. Visual detection effects of compared methods in scene 2.
Sensors 23 07240 g008
Figure 9. Visual detection effects of compared methods in scene 3.
Figure 9. Visual detection effects of compared methods in scene 3.
Sensors 23 07240 g009
Figure 10. Visual detection effects of compared methods in scene 4.
Figure 10. Visual detection effects of compared methods in scene 4.
Sensors 23 07240 g010
Figure 11. Visual detection effects of compared methods in scene 5.
Figure 11. Visual detection effects of compared methods in scene 5.
Sensors 23 07240 g011
Figure 12. Visual detection effects of compared methods in scene 6.
Figure 12. Visual detection effects of compared methods in scene 6.
Sensors 23 07240 g012
Figure 13. Visual detection effects of compared methods in scene 7.
Figure 13. Visual detection effects of compared methods in scene 7.
Sensors 23 07240 g013
Figure 14. Visual detection effects of compared methods in scene 8.
Figure 14. Visual detection effects of compared methods in scene 8.
Sensors 23 07240 g014
Figure 15. ROC curves of compared methods in six sequences. (af) represent the ROC curves for each of the six sequence scenarios.
Figure 15. ROC curves of compared methods in six sequences. (af) represent the ROC curves for each of the six sequence scenarios.
Sensors 23 07240 g015
Table 1. SSIM, BSF, and SNR of nine algorithms under scenarios 1–3.
Table 1. SSIM, BSF, and SNR of nine algorithms under scenarios 1–3.
MethodScene 1Scene 2Scene 3
SSIMBSFSNRSSIMBSFSNRSSIMBSFSNR
GST0.978048.616.090.9978145.0212.580.981851.48−5.09
Top_hat0.905615.551.370.956820.698.060.914211.11−0.69
DPSR0.9960111.316.740.9984177.757.090.9963115.60−20.61
ASTTV0.994597.763.620.991077.808.630.9970131.252.21
VATF0.9994296.657.720.993689.587.740.994193.00−14.30
FKWR0.9966122.3011.560.991777.92−0.200.925225.75−11.50
SRWS0.9985179.7111.600.9993258.9212.970.9980156.98−21.35
Aadcdd0.973444.554.560.9970130.6813.600.981853.59−1.70
PSTNN0.9993272.2911.210.9996374.6912.680.9971132.05−13.96
Proposed0.9996352.6811.580.9998487.5813.350.9999750.1510.10
Table 2. SSIM, BSF and SNR of nine algorithms under scenarios 4–6.
Table 2. SSIM, BSF and SNR of nine algorithms under scenarios 4–6.
MethodScene 4Scene 5Scene 6
SSIMBSFSNRSSIMBSFSNRSSIMBSFSNR
GST0.993987.87−0.030.9989214.0711.520.993085.345.86
Top_hat0.90840.912.450.955422.755.940.968730.432.18
DPSR0.993587.835.840.9986188.347.010.9979153.166.84
ASTTV0.9970130.103.170.9975143.757.060.9964120.515.37
VATF0.9990228.44−18.200.9987192.8211.400.991577.261.67
FKWR0.952332.393.600.9985183.7913.020.9984177.4212.01
SRWS0.9978151.572.230.9993270.5911.990.9994297.3111.23
Aadcdd0.985358.84−2.330.9983173.459.520.994696.826.74
PSTNN0.9980160.274.150.9993264.6011.640.9983171.517.97
Proposed0.9997405.9410.910.9997443.8213.570.9996363.2310.82
Table 3. Comparison of the running time (in seconds) of the ten methods.
Table 3. Comparison of the running time (in seconds) of the ten methods.
MethodsSequence 1Sequence 2Sequence 3Sequence 4Sequence 5Sequence 6
Aadcdd0.0317750.0259400.0253430.0265900.0258960.025677
ASTTV2.0253966671.813175671.806571.7308591.740839331.752016
DPSR0.3606310.2664060.2592330.2564140.2687860.260065
GST0.0345310.0213350.0219000.0215370.0231580.022639
FKWR0.1225410.1084350.0945640.0817140.0680380.113238
PSTNN0.4067640.2276690.2296370.2099950.2281070.209519
SRWS1.9264101.9561921.7776001.6593141.8595831.714700
Top_hat0.2488810.1554340.2890180.2823460.2827830.289122
VATF0.9124770.9261170.9120150.9095930.9185720.882623
Proposed0.0093127270.0097890.017515910.027098090.0268070.019232818
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Min, L.; Wu, A.; Fan, X.; Li, F.; Li, J. Dim and Small Target Detection with a Combined New Norm and Self-Attention Mechanism of Low-Rank Sparse Inversion. Sensors 2023, 23, 7240. https://doi.org/10.3390/s23167240

AMA Style

Min L, Wu A, Fan X, Li F, Li J. Dim and Small Target Detection with a Combined New Norm and Self-Attention Mechanism of Low-Rank Sparse Inversion. Sensors. 2023; 23(16):7240. https://doi.org/10.3390/s23167240

Chicago/Turabian Style

Min, Lei, Anqing Wu, Xiangsuo Fan, Feng Li, and Juliu Li. 2023. "Dim and Small Target Detection with a Combined New Norm and Self-Attention Mechanism of Low-Rank Sparse Inversion" Sensors 23, no. 16: 7240. https://doi.org/10.3390/s23167240

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop