An Improved Iterative Reweighted STAP Algorithm for Airborne Radar

: In recent years, sparse recovery-based space-time adaptive processing (SR-STAP) technique has exhibited excellent performance with insufﬁcient samples. Sparse Bayesian learning algorithms have received considerable attention for their remarkable and reliable performance. Its implementation in large-scale radar systems is however hindered by the overwhelming computational load and slow convergence speed. This paper aims to address these drawbacks by proposing an improved iterative reweighted sparse Bayesian learning algorithm based on expansion-compression variance-components (ExCoV-IIR-MSBL). Firstly, a modiﬁed Bayesian probabilistic model for SR-STAP is introduced. Exploiting the intrinsic sparsity prior of the clutter, we divide the space-time coefﬁcients into two parts: the signiﬁcant part with nontrivial coefﬁcients and the irrelevant part with small or zero coefﬁcients. Meanwhile, we only assign independent hyperparameters to the coefﬁcients in the signiﬁcant part, while the remaining coefﬁcients share a common hyperparameter. Then the generalized maximum likelihood (GML) criterion is adopted to classify the coefﬁcients, ensuring both accuracy and efﬁciency. Hence, the parameter space in Bayesian inference will be signiﬁcantly reduced, and the computational efﬁciency can be considerably promoted. Both theoretical analysis and numerical experiments validate that the proposed algorithm achieves superior performance with considerably improved computational efﬁciency in sample shortage scenarios.


Introduction
Moving target detection is a fundamental function of radar systems for military surveillance and reconnaissance.Space-time adaptive processing (STAP) has become an effective and mature clutter suppression technique for airborne early warning phased array radar systems [1][2][3].The optimal STAP filter is constructed using the ideal clutter plus noise covariance matrix (CNCM).In practice, the ideal CNCM is unknown a priori and is typically estimated using independent and identically distributed (i.i.d.) samples around the cell under test (CUT).According to the well-known Reed-Mallet-Brennan criterion [4], the number of i.i.d.samples with at least twice the system degree of freedom (DOF) is required to ensure a performance loss under 3 dB.Unfortunately, sufficient i.i.d.samples are often challenging to obtain due to various terrains, artificial structures and array configurations.
Over the past few decades, many STAP methods have been developed to improve clutter suppression performance in heterogeneous environments, including classical datadependent reduced-rank (RR) methods [5][6][7][8] and data-independent reduced-dimension (RD) methods [9][10][11][12].Although the required number of training samples is reduced to twice the reduced dimension, the requirement is still difficult to meet under severe non-stationary clutter environments for modern large-scale radar systems.
Convex relaxation substitutes the 0 norm with the 1 norm as the sparse penalty.The 1 norm optimization problem has shown that sparse solutions can be stably obtained under certain conditions [22].It has been extensively applied to the least absolute shrinkage and selection operator (LASSO) and basis pursuit (BP) based STAP methods [23].Most convex relaxation methods require careful tuning of the regularization parameter, and inappropriate parameter selection will jeopardize the performance of clutter suppression as well as slow-moving target detection [14,15].However, choosing an appropriate regularization parameter is quite challenging in practice.
In recent years, sparse Bayesian learning (SBL) has drawn much effort due to its preferable advantages, such as automatic self-regulation [18] and flexibility in exploiting the potential signal structure [24].SBL was first proposed by Tipping in 2001 [25] and introduced to the field of STAP by Duan in 2017, termed M-SBL-STAP [17].Numerous empirical results indicate that the SBL based SR-STAP can provide satisfactory performance and is quite robust to noise and high coherence dictionary [19].However, M-SBL-STAP faces an overwhelming computational burden and large memory requirements, hindering its implementation in large-scale radar systems.Many efficient SBL based STAP algorithms have been developed to address this issue.In [18], Wang proposed a fast-converging SBL algorithm by combining an approximation term, but the global convergence property of the algorithm is not guaranteed.An iterative reweighted based M-SBL (M-SBL-IR 2,1 ) STAP algorithm was proposed by Liu [19], which exhibits better reconstruction accuracy and has a favourable convergence speed.
Our experience with numerous simulation experiments demonstrates that the majority of the space-time coefficients to be recovered are zero or close to zero, and only a few have nontrivial values, owing to the sparse nature of the clutter.Therefore, in this paper, we are inspired to propose an improved Bayesian probabilistic model for SR-STAP, exploiting the aforementioned sparsity feature of the clutter.Instead of assigning a separate hyperparameter to each space-time coefficient in the conventional Bayesian model, only the significant coefficients are assigned independent hyperparameters, while the remaining irrelevant coefficients share a common hyperparameter.As a result, the parameter space to be updated will be dramatically reduced, and the computational efficiency will naturally be promoted.In [26], a heuristic expansion-compression variance-component based method (ExCoV) has been proposed to guide us to classify those hyperparameters.
Specifically, the main contributions of this paper are summarized as follows: 1.
By exploiting the inherent sparsity nature of the clutter, the space-time coefficients are divided into two disjoint groups, i.e., the significant and irrelevant groups, thus reducing redundancy, scaling down the parameter space and yielding an improved Bayesian probabilistic model for SR-STAP with reduced computational complexity and reduced memory requirements.

2.
The space-time coefficients are partitioned into the significant and irrelevant groups according to the generalized maximum likelihood (GML) criterion, which preserves both accuracy and efficiency, unlike the conventional SBL cost function.

3.
We extend and modify the real-value ExCoV method to complex-value STAP applications to approximately maximize the GML objective function.Using the Ex-CoV scheme, it is unnecessary to specify convergence thresholds and maximum iteration times.

4.
Extensive experiments as well as detailed comparative analyses are presented, such as clutter suppression performance and target detection performance, etc.
Notations used in this paper are as follows: vectors, matrices and scalars are denoted by bold lowercase, bold uppercase and italic letters, respectively.(•) * , (•) T and (•) H stand for conjugate, transpose and conjugate transpose.⊗ and are the Kronecker and Hadamard (elementwise) product.trace(•) is the trace operator.C represents the set of complex values.• F and • 2,0 are respectively defined as the Frobenius norm and 2,0 mixed norm, which is the number of non-zero elements of the vector formed by the 2 norm of each row.E[ • ] denotes the expectation operator.
The remainder of the paper is organized as follows.In Section 2, the airborne radar signal model is established.The formulation of SR-STAP model is presented in Section 3. The proposed algorithm is introduced in Section 4. The computational complexity analysis is presented in Section 5. Experiments and analyses are carried out on Section 6.Finally, Section 7 discuss the conclusions.

Signal Model
In this research, assume that an airborne pulsed-Doppler early warning radar system with a side-looking uniform linear array (ULA), as depicted in Figure 1, has N omnidirectional elements with half wavelength interelement spacing d = λ/2 and transmits M coherent pulses over a coherent processing interval (CPI) at the fixed pulse repetition frequency (PRF) f r .The aircraft platform is cruising with a constant velocity v p .The remainder of the paper is organized as follows.In Section 2, the airborne radar signal model is established.The formulation of SR-STAP model is presented in Section 3. The proposed algorithm is introduced in Section 4. The computational complexity analysis is presented in Section 5. Experiments and analyses are carried out on Section 6.Finally, Section 7 discuss the conclusions.

Signal Model
In this research, assume that an airborne pulsed-Doppler early warning radar system with a side-looking uniform linear array (ULA), as depicted in Figure 1, has N omnidirectional elements with half wavelength interelement spacing / 2 and transmits M coherent pulses over a coherent processing interval (CPI) at the fixed pulse repetition frequency (PRF) r f .The aircraft platform is cruising with a constant velocity p v .
... Based on the well-known Ward clutter model [27], the ground can be divided into different range rings.The radar returns are composed of numerous evenly distributed and mutually uncorrelated clutter patches in azimuth angles θ .Considering the impact of range ambiguities, a general space-time snapshot from the CUT can be formulated as Based on the well-known Ward clutter model [27], the ground can be divided into different range rings.The radar returns are composed of numerous evenly distributed and mutually uncorrelated clutter patches in azimuth angles θ.Considering the impact of range ambiguities, a general space-time snapshot x ∈ C N M×1 from the CUT can be formulated as where α r, i and s( f d r, i , f s r, i ) denote the complex amplitude and space-time steering vector from the corresponding clutter patch; N c is the number of independent clutter patches in a iso-range ring; N r represents the number of range ambiguities; the vector n is the thermal noise and modelled as a zero-mean complex Gaussian random process; s( f d r, i , f s r, i ), stand for space-time, temporal and spatial steering vectors, respectively.According to the geometric configuration of the radar platform shown in Figure 1, the normalized spatial frequency f s r,i and normalized Doppler frequency f d r, i are defined as follows: where ϕ r and θ i are the elevation and azimuth angle of the (r, i)-th clutter patch, respectively.Since the clutter patches are mutually independent, the ideal clutter plus noise covariance matrix (CNCM) of the CUT can be expressed as where σ 2 denotes the noise power.Under the linearly constrained minimum variance (LCMV) criterion, the optimum STAP weight vector, which maximizes signal-to interference-plus-noise ratio (SINR), can be given by where s target is the space-time steering vector of the desired target.

Sparse Recovery Based STAP Model Specification
For the grid-based SR-STAP algorithms [23], the whole continuous spatial-temporal plane is uniformly discretized into K = N s M d grids, where N s = ρ s N(ρ s > 1) is the number of normalized spatial frequency bins in the spatial domain and M d = ρ d M(ρ d > 1) is the number of normalized Doppler frequency bins in the temporal domain.ρ s and ρ d are the resolution scales.Then Φ = s( f d, 1 , f s, 1 ), • • • , s( f d, M d , f s, N s ) is defined as the N M × K overcomplete dictionary matrix consisting of K grids.The SR-STAP signal model for the multiple measurement vectors (MMV) case X = [x 1 , x 2 , . . ., x L ] ∈ C N M×L can next be reformulated as where L is the number of IID space-time snapshots; A = [a 1 , a 2 , . . ., a L ] ∈ C K×L denotes the space-time coefficients matrix where the non-zero rows indicate the potential presence of clutter components, and each column has the same sparsity support; where η is a nonnegative regularization parameter controlling the trade-off between the sparse penalty and data fidelity, and thus the choice of η is critical to the recovery performance.Unfortunately, finding the 2,0 norm optimal representation requires a combinatorial search (also known as NP-hard [28]) and, therefore, is difficult to obtain.Many recent alternative tractable approaches have been developed to find sparse solutions efficiently [13][14][15][16][17][18][19][20][21].Motivated by the predominance of the SBL algorithms in SR-STAP, we resort to SBL for accurately recovering A from X in this research.

Proposed Algorithm
Following the conventional sparse Bayesian learning framework, all the unknowns are treated as stochastic variables with assigned independent probability distributions.First of all, the noise matrix is modelled as white complex Gaussian noise with unknown power σ 2 ; thus the observed Gaussian likelihood function of the measurements X for the MMV case can be expressed as.
Since measurements are mutually independent and identically distributed, we suppose each column of the space-time coefficient matrix A •l is assigned with the same zero-mean complex Gaussian prior γ, governing the prior variance of each unknown space-time coefficient.
Further, following the Bayesian theorem, the posterior probability density p(A X, γ, σ 2 ) can be easily calculated by The posterior probability p(A X, γ, σ 2 ) obeys a multivariate complex Gaussian distribution CN (µ, Σ) with mean and covariance respectively given by Thus, with a fixed Γ, the estimated sparse recovery solution A of M-SBL is Accordingly, for γ i → 0 , the corresponding row A i• of the space-time coefficient matrix will be zeros as well.In other words, if γ is sparse, the corresponding space-time coefficient estimation A will also be sparse.
The hyperparameters vector can be estimated by performing the type-II maximum likelihood procedure or evidence maximization in γ space [29].Mathematically, the cost function can be expressed as minimizing the marginal likelihood function L(γ) with respect to γ: After the unknown coefficients A have been integrated out, P stands for the covariance of the measurements X with the hyperparameters γ and σ 2 .
This minimization can be performed by an iterative reweighted SBL based algorithm [19,30], which will be modified and described in the following part of this paper.
As we can observe from ( 11) and ( 12), the computational bottleneck is mainly manifested in large-scale matrix multiplication and matrix inversion operations.The storage requirement is also heavy.Moreover, during each iteration, a hyperparameter space (γ, σ 2 ) of dimension K + 1 needs to be updated.These factors make SBL based STAP algorithms considerably slower than other types of sparse recovery algorithms, even though excellent clutter suppression performance can be achieved with limited training samples.
In fact, due to the intrinsic sparsity of the clutter spectrum on the spatial-temporal plane, only a few rows of A have nontrivial magnitudes, and the remaining elements are strictly zero (or close to zero).It implies that most of the hyperparameters in γ will converge to zeros, and they are redundant.Therefore, we can naturally consider partitioning the space-time coefficients into two parts: the significant elements part and the complementary irrelevant elements part.Instead of assigning independent hyperparameters to all coefficients in the conventional SBL framework, we can only allocate independent hyperparameters to significant coefficients part, while the remaining irrelevant coefficients share an identical hyperparameter.Consequently, the dimension of the hyperparameter space can be reduced to be proportional to the sparsity level of the clutter, which is much smaller than K [23].As a result, the computation complexity will be greatly mitigated, and the memory consumption will be reduced, particularly in large-scale modern radar systems.
Next, define Θ = {1, 2, • • • , K} as the complete index set.The index set of significant coefficients is denoted by Θ α with size K α , and the complementary index set of irrelevant coefficients is denoted by Θ β = Θ\Θ α with cardinality K β = K − K α .Accordingly, we will also divide dictionary matrix Φ and space-time coefficients matrix A into two submatrices, i.e., respectively.For more details: • Φ α is the submatrix of Φ corresponding to the significant coefficients index set Θ α , e.g., if Θ α = {2, 8, 9}, then Φ α = [s 2 , s 8 , s 9 ], where s i is the ith column of Φ.
• A α is the submatrix of A corresponding to the index set Θ α , e.g., if Then (9) can be reformulated as where Γ α and γ Θ β I denote the covariance matrix of the significant part A α and irrelevant part A β , respectively.If all the coefficients belong to the Θ α index set and K α = K, then the above probabilistic model will degenerate to the original SBL full model in (9).Similarly, according to the index sets Θ α and Θ β , ( 11) and ( 12) can be respectively partitioned into submatrices where the covariance matrix P in (15) now can be rewritten as For the sake of expression simplicity, we define the set of all unknowns: Now, let us consider estimating the most superior and efficient coefficient index sets Θ α and Θ β , i.e., the optimal representation.According to [26], the generalized maximumlikelihood (GML) criterion can be used to assign the hyperparameters, and its equivalence with the canonical optimization problem in ( 7) is demonstrated.The GML criterion maximizes: where the first term is the same marginal likelihood function in (14) with Θ α fixed, enforcing the estimate fit the measurements; the second term I(ψ) is the Fisher information matrix (FIM) for the hyperparameters.Based on the well-known FIM for the Gaussian measurement model [31], we can deduce the block-partitioned FIM result for the hyperparameters vector (the detailed derivation is shown in Appendix A).
with each block computed as Since the hyperparameter space is nested, namely any Θ α is a subset of Θ, the more parameters in the set Θ α , the larger the value ln p(X|ψ) will be, which reduces the model mismatch.Therefore, there will always be a tendency to choose the complete set Θ α = Θ [31].However, the second term increases as the set Θ α grows, penalising the growth of the set.The GML criterion is, therefore, able to maintain a balance between underfitting and overfitting of the model, ensuring both accuracy and efficiency.
Nevertheless, directly applying the GML criterion (23) to obtain the optimal parameter index sets still requires an exhaustive search and is hard to implement in practical applications.Subsequently, we will employ an ExCoV based method to maximize the objective function approximately [32].The fundamental idea is interleaving the expansion/compression step and the parameter update step.In each expansion or compression procedure, we modify the current estimate of Θ α by one element per step to obtain a larger GML(ψ).Then followed by the parameter update step, the hyperparameters (γ Θ α , γ Θ β , σ 2 ) are updated for a fixed index set Θ α .Subsequently, we introduce the three main iteration steps involved in ExCoV: expansion step, compression step and parameter update step.
A. Expansion step.In this step, we determine the index q ∈ Θ β (p) corresponding to the row of µ β (p) with the largest 2 norm q = arg max where the superscript (•) (p) represents the iteration number.Then the index q is moved from Θ B. Compression step.Here, we determine the index t ∈ Θ α (p) corresponding to the smallest element of γ Θ α (p) (p) C. Parameters update step.In [26], the original ExCoV algorithm employs the expectationmaximization (EM) method for Bayesian inference to estimate the hyperparameters, which usually requires extensive iterations to converge [18].To overcome this drawback and thus further improve code running speed, an iterative reweighted M-SBL strategy is modified to update hyperparameters.
In this step, we firstly assume that the index sets Θ α and Θ β are known and fixed.We next present the derivation of the parameters update step based on the iterative reweighed 2,1 M-SBL algorithm from [19] modified by the previous proposed Bayesian probabilistic model (details can be seen in Appendix B), which can remarkably accelerate the convergence speed.Then, the hyperparameters γ Θ α (p+1) (p+1) , γ Θ β (p+1) , (σ 2 ) (p+1) can be updated by (30) where We term the proposed algorithm as an improved iterative reweighted sparse Bayesian learning algorithm based on expansion-compression variance components (ExCoV-IIR-MSBL).Figure 2 illustrates the processing flowchart of the proposed algorithm and its procedures are summarized as follows.Step1 Initialization: Calculate the minimum norm solution of the space-time coefficients ( ) Then we can utilize the prior knowledge provided by the inertial navigation system and radar system to initialize where represents the slope of the clutter ridge [33] and ⋅     denotes rounding-up operation; or simply set ( 0 ) 1 K α = , since the ExCoV method is insensitive to the initial value of a K .Hence, the index set (0 ) α Θ is constructed using the rows of (0) μ corresponding to the first Step2 Cycle Initialization: Set the initial values for the hyperparameters 2 ( , , ) ( ) ( ) Step1 Initialization: Calculate the minimum norm solution of the space-time coefficients Then we can utilize the prior knowledge provided by the inertial navigation system and radar system to initialize where β = 2v p /( f r d) represents the slope of the clutter ridge [33] and • denotes roundingup operation; or simply set K (0) α = 1, since the ExCoV method is insensitive to the initial value of K a .Hence, the index set Θ (0) α is constructed using the rows of µ (0) corresponding to the first Step2 Cycle Initialization: Set the initial values for the hyperparameters (γ Generally, these parameters can be initialized to an arbitrary positive vector, but it would be more beneficial if a rough estimate could be given.Then the initial values of the optimal record are set to ψ = γ (0) α and μ = µ (0) .Step3 Expansion: Apply the aforementioned expansion step, yielding the updated Θ α (p+1) ,Θ β (p+1) and γ Θ α (p+1) (p) .
Step6 Termination Condition Check: In this step, we would verify whether the expansion is still required to continue [26].
where T is the length of a sliding average window.This condition will help us to determine if there is still a need for the expansion operation and prevents premature termination.If the inequality is satisfied, then the expansion operation is terminated, and Step7 begins; otherwise, the expansion continues and returns to Step3.
Step10 Termination Condition Check: In this step, we would verify whether the compression is still required to continue and check the same condition (39) in Step6.If the inequality is satisfied, then the compression operation is terminated, and Step11 begins; otherwise, the compression continues and returns to Step7.
Step11 Globally Optimal Estimates Update: Moreover, we keep a record of the globally optimal parameter set ψ and the corresponding coefficients μ in the entire iteration cycles, verifying the condition If the inequality holds, then the globally optimal estimates ψ and μ can be updated with ψ and μ; instead, keep them unchanged.
Step12 Algorithm Iteration Termination Check: If the globally optimal index set Θα is consistent between two consecutive cycles.The proposed algorithm is next terminated and the final globally optimal space-time coefficients μ and noise power σ 2 are output.
Otherwise, the globally optimal index set Θα is updated, then the algorithm continues.We reset the iteration index p = 0, K (0) α = Kα , and calculate the initial values for the space-time coefficients Step13 STAP Weight Calculation: Once the estimated sparse recovery solution A = μ and σ 2 = σ 2 is obtained, the CNCM can be reconstructed from Then the STAP weight w can be obtained by using (5).
Step14 Output: Give the filtered output of the cell under test y = w H x CUT .
From the above steps, it is worth noting that the proposed algorithm is fully automatic and does not require setting any convergence threshold as well as the maximum iteration number.

Computational Complexity Analysis
In this Section, we will analyse the computational complexity of the proposed ExCoV-IIR-MSBL algorithm and compare it with some state-of-the-art SR-STAP algorithms, including M-CVX [34], M-OMP [35], M-FOCUSS [36], M-IAA [21], M-SBL [17], M-FCSBL [18] and M-SBL-IR 2,1 [19].Their computational complexity would be measured by the number of complex multiplications in a single iteration.The results are listed in Table 1, where r s stands for the clutter sparsity level.

Algorithm
Computational Complexity for a Single Iteration In the conventional sparse Bayesian learning probabilistic model, the computational load is greatly centred on the parameter update step, see (11) and (12).The computational complexity is in the order of O 3K 2 N M + 2(N M) 2 K + (NM) 3 + N MKL , with large-scale matrix multiplication and matrix inversion dominating a major part.Moreover, the memory requirement is O K 2 .Therefore, these drawbacks make the MSBL-STAP challenging to implement in modern large-scale airborne radar systems, despite achieving an excellent clutter suppression performance with only a few snapshots.
The primary computational burden of the proposed algorithm is concentrated on the parameter update and GML criterion calculation steps.Since the parameter space dimension is significantly reduced, the scale of matrix multiplication is also significantly decreased, given that K α K.In addition, unlike other SBL algorithms, the proposed algorithm no longer needs to update the variance matrix (12); hence, the computational complexity can be further reduced.As a result, the computational complexity is in the , and the required memory consumption is O(N MK).It can be seen that the computational load and memory consumption of the proposed algorithm have been remarkably reduced.These advantages make the proposed algorithm more suitable for modern large-scale airborne radar systems.
For a more intuitive comparison of the above-mentioned algorithms, we plot Figure 3, illustrating the relationship between the number of complex multiplications and the number of pulses.We assume that ρ s = ρ d = 4, N = 8, L = 8, and K α is set to be the rank of clutter.It can be seen from Figure 3 that the M-SBL-IR 2,1 has nearly the same computational load as M-SBL, and the proposed algorithm has the lowest computational complexity among these compared SBL based algorithms and is close to the M-IAA and M-FOCUSS algorithms.

Numerical Experiment
In this section, numerous experiments are performed with simulated data to evaluate the performance of the proposed algorithm.The simulated data are generated by the well-known Ward clutter model introduced in the previous section.The main simulation parameters of a side-looking ULA are listed in Table 2.The dictionary resolution scales are set to be ρ s = ρ d = 4. Furthermore, all the simulation results are averaged over 100 independent Monte-Carlo trails.First of all, we examine the average code running time of the proposed algorithm and compare it with other state-of-the-art SR-STAP algorithms, as listed in Table 3. M-CVX, M-IAA, M-OMP, M-FOCUSS, M-SBL, M-FCSBL and M-SBL-IR 2,1 are employed as benchmarks.Note that all the simulations are operated on the same workstation with Intel Xeon 4114 CPU @2.2GHz and 128 GB RAM.As seen from Table 3, since the parameter space is dramatically shrunk, the average running time of the proposed ExCoV-IIR-SBL is remarkably faster than the other compared SBL based STAP algorithms.It is even better than M-FOCUSS and M-IAA since the proposed algorithm can reach the steady state with fewer iterations with the help of a modified iterative reweighed parameter update step.Thus, the efficiency of the proposed algorithm is demonstrated.
In the following experiments, we assess the clutter suppression performance via the metric of signal to interference plus noise ratio (SINR) loss [27], defined as the ratio of output SINR to the optimum output SNR in a noise-only environment, i.e., where R is the clairvoyant CNCM of the CUT, and w is the STAP weight.
Next, the clutter suppression performance in the ideal case is considered.The SINR loss performance of the proposed algorithm is evaluated and compared with other stateof-the-art SR-STAP algorithms, including M-CVX, M-IAA, M-OMP, M-FOCUSS, M-SBL, M-FCSBL and M-SBL-IR 2,1 .The number of i.i.d.training samples are set to be L = 9. Figure 4 depicts the curves of the SINR loss against the normalized Doppler frequency.As clearly illustrated in Figure 4, the proposed algorithm achieves the same near-optimal performance as M-IAA and the other three SBL based algorithms, revealing that the novel Bayesian probabilistic model proposed in this paper for STAP can accurately reconstruct the CNCM with a smaller parameter space.The performance of M-FOCUSS is found to be slightly inferior to theirs.Meanwhile, the proposed ExCoV-IIR-MSBL consumes the least running time among these three SBL based algorithms on this basis.To more explicitly demonstrate the clutter suppression performance, Figure 5 plots the spatial-temporal adapted responses, given the STAP weight formed by the abovementioned algorithms.The presumed main lobe is located at a normalized spatial frequency of 0 with a normalized Doppler frequency of 0.2.From Figure 5, it is evident that the proposed method is able to maintain high gain at the presumed target location while precisely forming notches at the clutter ridge, suppressing both main lobe and sidelobe of the clutter component.Poorly performing algorithms have distorted two-dimensional responses and cannot accurately form notches at the clutter ridge.To more explicitly demonstrate the clutter suppression performance, Figure 5 plots the spatial-temporal adapted responses, given the STAP weight formed by the abovementioned algorithms.The presumed main lobe is located at a normalized spatial frequency of 0 with a normalized Doppler frequency of 0.2.From Figure 5, it is evident that the proposed method is able to maintain high gain at the presumed target location while precisely forming notches at the clutter ridge, suppressing both main lobe and sidelobe of the clutter component.Poorly performing algorithms have distorted two-dimensional responses and cannot accurately form notches at the clutter ridge.Note that because of the massive computational complexity of the M-CVX, it will not be discussed in the subsequent experiments.Then, Figure 6  Note that because of the massive computational complexity of the M-CVX, it will not be discussed in the subsequent experiments.Then, Figure 6 illustrates the curves of the average SINR loss versus different a number of training samples.It should be noted that the average SINR loss is calculated by the mean of SINR loss for f d ∈ (−0.5, −0.05) ∪ (0.05, 0.5) in this paper.As demonstrated in Figure 6, all algorithms have a certain degree of performance loss with one snapshot and increase with the number of training samples.ExCoV-IIR-MSBL apparently exhibits promising performance similar to the M-IAA and other SBL algorithms under small sample conditions.The proposed algorithm achieves a nearoptimum performance with merely three training samples.

Figure 3
(g) (h) Note that because of the massive computational complexity of the M-CVX, it will not be discussed in the subsequent experiments.Then, Figure 6    Subsequently, we verify the target detection performance via the probability of detection (P d ) versus the target SNR curves, which are acquired by employing the cell-average constant false alarm rate (CA-CFAR) detector [37].One hundred targets within the boresight are randomly added to the entire Range-Doppler plane, and the false alarm probability (P fa ) is fixed to 10 −3 .Figure 7 shows that the proposed algorithm and other SBL based algorithms have noticeable improvements compared to M-OMP and M-FOCUSS.This result indicates that the proposed algorithm has superior target detection performance.Subsequently, we verify the target detection performance via the probability of detection (Pd) versus the target SNR curves, which are acquired by employing the cell-average constant false alarm rate (CA-CFAR) detector [37].One hundred targets within the boresight are randomly added to the entire Range-Doppler plane, and the false alarm probability (Pfa) is fixed to 10 −3 .Figure 7 shows that the proposed algorithm and other SBL based algorithms have noticeable improvements compared to M-OMP and M-FOCUSS.This result indicates that the proposed algorithm has superior target detection performance.Besides, we also consider the SINR loss performance under the spatial mismatch case in the presence of angle-dependent gain-phase error.Specifically, the gain errors among the antennas are generated by a Gaussian distribution with a standard deviation of 0.03, and the phase errors are uniformly distributed within [0°, 2°]. Figure 8 shows the SINR loss versus normalized Doppler frequency.From Figure 8, the curves show that all the SR-STAP algorithms suffer severe performance degradation due to the model mismatch between the dictionary and the actual clutter component.The M-IAA has the most serious performance loss, despite the fact that it can approach near-optimal as the SBL based algorithms in the ideal case.Besides, we also consider the SINR loss performance under the spatial mismatch case in the presence of angle-dependent gain-phase error.Specifically, the gain errors among the antennas are generated by a Gaussian distribution with a standard deviation of 0.03, and the phase errors are uniformly distributed within [0 • , 2 • ]. Figure 8 shows the SINR loss versus normalized Doppler frequency.From Figure 8, the curves show that all the SR-STAP algorithms suffer severe performance degradation due to the model mismatch between the dictionary and the actual clutter component.The M-IAA has the most serious performance loss, despite the fact that it can approach near-optimal as the SBL based algorithms in the ideal case.
in the presence of angle-dependent gain-phase error.Specifically, the gain errors among the antennas are generated by a Gaussian distribution with a standard deviation of 0.03, and the phase errors are uniformly distributed within [0°, 2°]. Figure 8 shows the SINR loss versus normalized Doppler frequency.From Figure 8, the curves show that all the SR-STAP algorithms suffer severe performance degradation due to the model mismatch between the dictionary and the actual clutter component.The M-IAA has the most serious performance loss, despite the fact that it can approach near-optimal as the SBL based algorithms in the ideal case.
Figure 9 shows the SINR loss versus the number of samples of above referred algorithms in the non-ideal case.As depicted in Figure 9, the proposed algorithm can yield a steady-state with only a few training samples as in the ideal case.Figure 9 shows the SINR loss versus the number of samples of above referred algorithms in the non-ideal case.As depicted in Figure 9, the proposed algorithm can yield a steady-state with only a few training samples as in the ideal case.From the above experiments, it has been demonstrated that the proposed algorithm ensures a satisfactory clutter suppression and target detection performance while keeping a remarkable computational efficiency, thereby making it more suitable for implementation in modern large-scale radar systems.

Conclusions
In this work, to enhance the computational efficiency of the M-SBL based SR-STAP algorithm, we derive a novel algorithm called improved iterative reweighted sparse Bayesian learning based on expansion-compression variance-components (ExCoV-IIR-MSBL).Inspired by the intrinsic sparse prior to the clutter, an improved Bayesian probabilistic model with reduced parameter space is developed for SR-STAP.Furthermore, the GML criterion is utilized to partition the efficient and accurate parameter space.Then we extend the ExCoV method into the SR-STAP application to obtain the probabilistic model with maximized GML objective function and enhance its convergence speed by a modified M-SBL-IR 2,1


under the proposed Bayesian model.With numerical experiments, we show that the proposed algorithm outperforms other existing state-of-the-art algorithms with lower computational complexity.From the above experiments, it has been demonstrated that the proposed algorithm ensures a satisfactory clutter suppression and target detection performance while keeping a remarkable computational efficiency, thereby making it more suitable for implementation in modern large-scale radar systems.

Conclusions
In this work, to enhance the computational efficiency of the M-SBL based SR-STAP algorithm, we derive a novel algorithm called improved iterative reweighted sparse Bayesian learning based on expansion-compression variance-components (ExCoV-IIR-MSBL).Inspired by the intrinsic sparse prior to the clutter, an improved Bayesian probabilistic model with reduced parameter space is developed for SR-STAP.Furthermore, the GML criterion is utilized to partition the efficient and accurate parameter space.Then we extend the ExCoV method into the SR-STAP application to obtain the probabilistic model with

Figure 1 .
Figure 1.Geometric configuration of an airborne surveillance radar system.

Figure 1 .
Figure 1.Geometric configuration of an airborne surveillance radar system.

Figure 2 .
Figure 2. Processing flowchart of the proposed algorithm.

Figure 2 .
Figure 2. Processing flowchart of the proposed algorithm.

Figure 4 .
Figure 4. SINR loss comparison of different algorithms in the ideal case.

Figure 4 .
Figure 4. SINR loss comparison of different algorithms in the ideal case.
illustrates the curves of the average SINR loss versus different a number of training samples.It should be noted that the average SINR loss is calculated by the mean of SINR loss for this paper.As demonstrated in Figure6, all algorithms have a certain degree of performance loss with one snapshot and increase with the number of training samples.ExCoV-IIR-MSBL apparently exhibits promising performance similar to the M-IAA and other SBL algorithms under small sample conditions.The proposed algorithm achieves a near-optimum performance with merely three training samples.

Figure 6 .
Figure 6.The average SINR loss versus the number of training samples in the ideal case.Figure 6.The average SINR loss versus the number of training samples in the ideal case.

Figure 6 .
Figure 6.The average SINR loss versus the number of training samples in the ideal case.Figure 6.The average SINR loss versus the number of training samples in the ideal case.

Figure 7 .
Figure 7. Probability of detection versus signal to noise ratio.

Figure 7 .
Figure 7. Probability of detection versus signal to noise ratio.

Figure 8 .
Figure 8. SINR loss comparison of different algorithms with gain-phase error.Figure 8. SINR loss comparison of different algorithms with gain-phase error.

Figure 8 .
Figure 8. SINR loss comparison of different algorithms with gain-phase error.Figure 8. SINR loss comparison of different algorithms with gain-phase error.

Figure 9 .
Figure 9.The average SINR loss versus the number of training samples with gain-phase error.

Figure 9 .
Figure 9.The average SINR loss versus the number of training samples with gain-phase error.

4 .
Extensive experiments as well as detailed comparative analyses are presented, such as clutter suppression performance and target detection performance, etc.
Notations used in this paper are as follows: vectors, matrices and scalars are denoted by bold lowercase, bold uppercase and italic letters, respectively.() * ⋅ , ( ) T ⋅ and ( ) H ⋅ stand for conjugate, transpose and conjugate transpose.⊗and  are the Kronecker and Hadamard (elementwise) product.( ) trace ⋅ is the trace operator.represents the set of complex values.F ⋅ and E ⋅ denotes the expectation operator.

Table 3 .
Running Time Comparison.