Next Article in Journal
A Simple Scenario for Explaining Asymmetric Deformation Across the Altyn Tagh Fault in the Northern Tibetan Plateau: Contributions from Multiple Faults
Previous Article in Journal
A High-Precision Frequency Synchronization Method Based on a Novel Geostationary Communication Satellite Phase-Locked Transponder
Previous Article in Special Issue
A Novel Azimuth Channel Errors Estimation Algorithm Based on Characteristic Clusters Statistical Treatment
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Novel Deep Unfolding Network for Multi-Band SAR Sparse Imaging and Autofocusing

1
School of Optoelectronic Information and Computer Engineering, University of Shanghai for Science and Technology, Shanghai 200093, China
2
Terahertz Technology Innovation Research Institute, Terahertz Spectrum and Imaging Technology Cooperative Innovation Center, University of Shanghai for Science and Technology, Shanghai 200093, China
3
Shanghai Institute of Intelligent Science and Technology, Tongji University, Shanghai 200092, China
4
School of Intelligent Emergency Management, University of Shanghai for Science and Technology, Shanghai 200093, China
5
State Key Laboratory of Millimeter Waves, Southeast University, Nanjing 210096, China
6
National Key Laboratory of Microwave Imaging Technology, Aerospace Information Research Institute, Chinese Academy of Sciences, Beijing 100094, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2025, 17(7), 1279; https://doi.org/10.3390/rs17071279
Submission received: 25 February 2025 / Revised: 1 April 2025 / Accepted: 1 April 2025 / Published: 3 April 2025
(This article belongs to the Special Issue Microwave Remote Sensing for Object Detection (2nd Edition))

Abstract

:
The sparse imaging network of synthetic aperture radar (SAR) is usually designed end to end and has a limited adaptability to radar systems of different bands. Meanwhile, the implementation of the sparse imaging algorithm depends on the sparsity of the target scene and usually adopts a fixed L 1 regularization solution, which has a mediocre reconstruction effect on complex scenes. In this paper, a novel SAR imaging deep unfolding network based on approximate observation is proposed for multi-band SAR systems. Firstly, the approximate observation module is separated from the optimal solution network model and selected according to the multi-band radar echo. Secondly, to realize the SAR imaging of non-sparse scenes, L p regularization is used to constrain the uncertain transform domain of the target scene. The adaptive optimization of L p parameters is realized by using a data-driven approach. Furthermore, considering that phase errors may be introduced in the real SAR system during echo acquisition, an error estimation module is added to the network to estimate and compensate for the phase errors. Finally, the results from both simulated and real data experiments demonstrate that the proposed method exhibits outstanding performance under 0.22 THz and 9.6 GHz echo data: high-resolution SAR focused images are achieved under four different sparsity conditions of 20%, 40%, 60%, and 80%. These results fully validate the strong adaptability and robustness of the proposed method to diverse SAR system configurations and complex target scenarios.

1. Introduction

Synthetic aperture radar (SAR) is a type of advanced microwave imaging equipment offering all-day, all-weather operation, which has been widely used in military and civilian fields [1,2,3]. Its resolution is generally determined by the range and azimuth bandwidth, but it is also affected by other factors, such as system noise and atmospheric interference, etc. The effect is usually manifested in SAR echoes in the form of phase errors, leading to the defocusing of SAR images and a reduction in image resolution [4,5,6,7,8,9,10,11,12,13]. Moreover, in practical scenarios, SAR systems encounter the issue of sparse sampling due to complex operation environments [14]. Therefore, it is necessary to study the SAR imaging method under the condition of sparse sampling and phase errors.
For the sparse sampling problem, sparsity-driven SAR imaging has been widely studied in recent years. According to the compressed sensing (CS) theory, high-dimensional signals can be recovered from sparse sampling through the use of an effective algorithm [15]. With the development of CS theory in the field of SAR imaging, sparse SAR imaging methods are superior to conventional SAR imaging methods in terms of sampling rate and autofocusing [16,17]. Sparse recovery algorithms can be classified into three categories [18]. The greedy tracking algorithm, represented by Orthogonal Matching Pursuit (OMP) [19,20], has a high computational efficiency and needs to determine sparse prior information of reconstruction results. The non-convex optimization algorithm, represented by the sparse Bayesian algorithm [21,22], determines the sparse prior model of the signal and uses the maximum likelihood criterion to estimate the parameters of the prior model to obtain sparse recovery results. However, it involves more matrix inverse operations and has a low computational efficiency. The convex relaxation optimization algorithm, represented by the Iterative Shrinkage Threshold Algorithm (ISTA) [23], Approximate Message Passing (AMP) [24], and Alternating Direction Method of Multipliers (ADMM) [25], utilizes the relaxation l 1 ( 0 p 1 ) of the normal form of the L 0 paradigm in the CS framework, effectively avoiding NP-hard problems. However, all the above CS-SAR imaging algorithms need to pull two-dimensional imaging scenes and two-dimensional echo data into one-dimensional column vectors [26]. With the increase in observation scenes, the dimension of observation matrix becomes very large, which makes them have great limitations in practical applications.
To reduce the dimension of observation matrix and improve imaging efficiency, some CS-SAR imaging methods are proposed based on an approximate observation matrix. In [27], a sparse SAR imaging method based on approximate observation operators is investigated, which combines the traditional SAR imaging algorithm based on matched filtering with CS theory. In [28], the idea of an approximate observation matrix is introduced into FMCW-SAR and applied to real sparse SAR imaging. In [29], a stripmap SAR sparse imaging and autofocusing method is realized through the combination of approximate observation matrix and L 1 sparse regularization methods. However, all of the above methods are affected by data characteristics such as noise levels and sparsity, and the numerical stability and convergence of the algorithm implementation are dependent on parameter settings, requiring the manual adjustment of parameters for different imaging scenes to ensure quality, especially in cases where a high resolution is required, which relies heavily on manual expertise.
With the popularity of deep learning, deep unfolding methods [30,31,32], which combine the iterative process of traditional algorithms with interpretable deep networks, have become more prevalent in CS fields. In the CS-SAR imaging field, some methods have been preliminary employed [33,34,35,36,37,38]. In [33], a complex-valued ADMM-Net method is proposed to improve the stability of the ADMM and realize sparse aperture ISAR imaging and autofocusing. In [34], an autofocusing network, AF-AMP-Net, is proposed to obtain high cross-range resolution for sparse aperture ISAR. In [35], a non-iterative autofocus scheme based on deep learning and a minimum-entropy criterion is proposed. However, these methods are not based on raw data for imaging processing and do not reduce system complexity or sampling rates. In [36], a deep neural network architecture (SAE-Net) is designed to achieve both SAR imaging and autofocus. However, it is based on the traditional CS model and is difficult to apply to resolve real SAR large-scene imaging problem. In [37], a SAR sparse imaging deep unfolding network based on an approximate observation matrix is proposed, which does not consider motion error and range migration correction. In [38], a deep network based on traditional CS models and a data-driven approach is proposed for 3D millimeter wave radar imaging. The network is a fixed embedded structure, which is not suitable for radar systems of different bands. In general, the combination of deep learning technology and traditional sparse SAR imaging technology can effectively improve the image quality [39,40]. But the traditional sparse SAR imaging methods usually assume that the target scene is sparse, that is, there are fewer target points. Due to the complexity of SAR observation scenes, the effect of deep unfolding network reconstruction based on this assumption is not ideal.
To solve the problem of multi-band SAR sparse imaging and autofocusing in complex scenes, a novel deep unfolding network is proposed. The network combines the approximate observation operator with the ADMM solver to reconstruct the image. The approximate observation operator module and the network model are designed separately, and the model optimization does not depend on the fixed band radar data. Moreover, to realize the SAR imaging of complex scenes, the uncertain transform domain of the target scene is constrained by the non-fixed L p norm. In addition, the network’s parameters are adaptively learned in a data-driven way, avoiding the influence of empirical parameter settings on SAR image quality. The superiority of the proposed method is verified by 0.22 THz and 9.6 GHz bands of airborne SAR simulated and real imaging experiments. The main contributions of our work can be summarized as follows:
  • For complex scenes where targets and background are difficult to distinguish, and where sparse sampling, information loss, or significant interference from factors such as vegetation and buildings exist, a deep unfolding network based on multi-band SAR sparse imaging and autofocus is proposed. By decoupling the approximate observation module from the optimal solution network, the proposed network can dynamically adapt to the characteristics of SAR echoes across different bands, significantly enhancing multi-band imaging capabilities;
  • A non-fixed L p regularization is used to constrain the uncertain transform domain of the target scene. The L p parameter and network’s parameters are trained in a data-driven way, which overcomes the difficulties of parameter adaptation and the sparse transformation of the target scene in traditional iterative methods;
  • An error estimation module is added to the network to estimate and compensate for the phase errors introduced in the raw echo data for real SAR systems
This paper is organized as follows. Section 2 describes the SAR signal measurement model and the conventional CS-SAR imaging model. Section 3 presents the solution to the imaging model and elaborates on the details of the proposed SAR imaging network. Experimental results for both simulated and measured data in different bands are given in Section 4. Finally, conclusions are drawn in Section 5.

2. Sparsity-Driven SAR Measurement and Imaging Modeling

2.1. SAR Signal Measurement Model

In this paper, we focus on strip SAR imaging, whose model is shown in Figure 1. Assume the radar platform is moving at a velocity v along the azimuth direction, with the range direction perpendicular to the azimuth direction and the height axis denoted as z . The closest slant range between the radar and the target is R 0 , and the angle between the target and the zero-Doppler plane is β . To realize a large bandwidth signal and acquire a high-resolution image, a linear frequency modulation (LFM) signal is transmitted [39]. Then the corresponding baseband echo signal can be expressed as [41]
s η , t = i = 1 I g i ω a ω r exp j 4 π f c R i η c exp j π K r t 2 R i η c 2
where I is the total number of scattering points in the observed scene, g i is the reflection coefficient of the i’th scattering point, η and t are the slow time and fast time, respectively, and ω a and ω r are the window functions along the slow time and fast time, respectively. f c is the carrier frequency, c is the speed of light, R i ( η ) is the instantaneous slant range of the i’th scattering point, and K r is the chirp rate.
In addition, the echo data can be expressed in a matrix-vector form as follows [41]:
y T = W x + n
where y T C M N × 1 is the vectorized form of the fully sampled echo matrix Y T C M × N , M and N stand for the number of samples in the slow time and the fast time, respectively. W C M N × P Q is the exact observation matrix, and P and Q represent the number of scattering points in the range and azimuth directions, respectively. x C P Q × 1 is the vectorized form of the backscatter coefficient matrix X C P × Q of the imaging scene. n C M N × 1 is the Gaussian noise matrix.
When the airborne SAR system is in a complex environment, part of the SAR echo data will be lost, and there will be phase errors in the SAR echo data, resulting in performance degradation of the traditional imaging method. To simplify the model, only azimuthal sparse sampling and phase error are considered in this paper, and the SAR echo signal model in Equation (2) can be rewritten as
y = E S W x + n
where y C L N × 1 is the vectorized form of the presence of phase error and the sparse sampling echo matrix Y C L × N , L stands for the number of slow-time samplings, and L < M . S C L N × M N is a block downsampling matrix, S = I N D , and I N is a unit matrix of size N × N . D C L × M is a downsampling matrix, and denotes the Kronecker product. E C L N × L N is a block phase error matrix, E = I N Φ , and Φ is a phase error matrix, Φ = d i a g e x p j ϕ ; ϕ C L × 1 is an error vector.
Based on the aforementioned SAR signal model, the sparse sampling problem can be addressed through CS theory [27]. CS theory demonstrates that a high-dimensional signal can be recovered from a small number of samples, provided that the signal exhibits sparsity in a certain transform domain. In the following section, the CS-SAR imaging model is introduced.

2.2. CS-SAR Imaging Model

The sparse SAR imaging model is to recover the observation scene x from the sparse echo signal y, i.e., it is necessary to solve the SAR echo signal model based on Equation (3).
x = argmin x y E S W x 2 2 + λ T x p
In the second term, λ is a sparse regularization coefficient; T is a feature matrix representing the internal features of the observed scene, such as edges, contours, and continuity; and p is a sparse regularization corresponding to l p ( 0 p 1 ) .
In conventional CS-SAR imaging, the signal sparsity is represented by the l 0 -norm, which indicates the number of nonzero elements of the signal [42]. The optimization problem under the sparsity constraints based on the l 0 -norm is NP-hard. Among them, the l 1 -norm is a good approximation of the non-convex l 0 -norm, which can be used to solve the optimization problem of x based on the l 1 -norm of the ADMM. In the ADMM, an auxiliary variable z is introduced, at which point the above optimization problem is equivalent [36]:
min x , z y E S W x 2 2 + λ T z 1 s . t . x z = 0
Equation (5)’s augmented Lagrangian function is:
L ρ x , z , α = y E S W x 2 2 + λ T z 1 + α H x z + ρ 2 x x 2 2
where ρ denotes the penalty parameter, and α denotes the Lagrange multiplier. The ADMM divides the problem solution into the following three parts for iterative solution:
x k = argmin x L ρ x k 1 , z k 1 , α k 1 z k = argmin z L ρ x k , z k 1 , α k 1 α k = α k 1 + η x k z k
where k is the number of iterations of the algorithm, and η is the Lagrangian update step. The first two optimization problems can be solved by calculating the first-order derivative of the augmented Lagrangian function L ρ x , z , α for x and z , respectively. Moreover, by equating the first-order derivatives to zero, Equation (7) can obtain its solution as follows [36]:
x k = W H S H E H E S W + ρ I M N 1     W H S H E H y + ρ z k 1 α k 1 z k = S x k + α k ; λ ρ α k = α k 1 + η x k z k
In Equation (8), the subscript H denotes the conjugate transpose, and S is a soft threshold function corresponding to the sparse regularization of the l 1 -norm:   S x , t = x x m a x ( x t , 0 ) . After performing k iterations on Equation (8) until the output threshold is satisfied, the image x is obtained.

3. Proposed Methodology

3.1. SAR Imaging and Autofocusing

At present, the sparse imaging network of synthetic aperture radar (SAR) usually adopts the end-to-end design, which has limited adaptability to different band radar systems [32]. In this paper, the imaging operator part of the network model is designed separately, and its parameters do not participate in the optimization of the network model. After the completion of the network optimization, the corresponding approximate observation operator can be selected in the multi-band radar imaging task according to the actual radar system, which improves the universality of the network model. For radar data processing in different bands, different radar systems are used to construct the corresponding approximate observation operators. In order to achieve better single-band radar imaging, the imaging operator can also be trained as a network parameter to optimize the single-band radar imaging task of the network model.
As stated in the above section, the solution of the CS-SAR imaging problem involves the inversion and multiple iterations of the exact measurement matrix W . Due to the large scale of the matrix W , the computing and storage capabilities required are high, which greatly reduces the computational efficiency of SAR imaging algorithms. To solve this problem, this paper adopts the matched filtering (MF) SAR imaging method [43], such as the Chirp Scaling algorithm, which constructs the approximate measurement operator by decoupling the azimuth and range processing. The exact observation matrix W is replaced by the approximate observation operator, i.e., W G · , whose corresponding inverse transformation is I · , which represent the inverse imaging operator and the imaging operator, respectively, in the following procedure [43]:
G · = F a 1 F r 1 F r F a · H 3 * H 2 * H 1 * I · = F a 1 F r 1 F r F a · H 1 H 2 H 3
where   is the Hadarmard product; F r , F a , F r 1 and F a 1 are the Fourier transform and inverse Fourier transform operators along the range direction and azimuth direction, respectively; and H 1 , H 2 , and H 3 are phase terms for RCMC, range compression, secondary range compression and consistent RCMC, and azimuthal compression and additional phase correction, respectively. H 1 * , H 2 * , and H 3 * are the conjugate terms of the corresponding phase terms [43]. Then, the solution model in Equation (4) is rewritten accordingly as
x = argmin x y E S G x 2 2 + λ T x 1
The solution of the sparse SAR imaging model based on the approximate observation operator reduces the scale of the SAR echo measurement matrix and the computational amount. However, the matrix vectorization solution of the traditional CS-SAR imaging still generates an order of magnitude higher computation than that of the MF method, which is difficult to adapt to the real-time processing of SAR images of large scenes. In order to reduce the computational and storage burden of imaging processing, based on the nature of the Kronecker product [44], the first sub-equation of Equation (8) is rewritten as
x k = G H I N D + ρ I M N 1   S H E H y + ρ G z k 1 G α k 1
Its vector expression is written in the corresponding matrix form for ease of computation:
X k = G H m a s k + ρ I M × N 1   D H Φ H Y + ρ G Z k 1 U k 1
where X is the matrix form of x ; Y is the matrix form of y ; Z is the matrix form of z ; U is the matrix form of α ; G H { · } = I { · } , m a s k is a sparse sampling matrix, whose elements have the value of either 1 or 0, which represents whether the corresponding sample is available or not; and I is an all-one matrix. At this point, the iterative matrix of the image is obtained, serving as an important step in the X module.
In addition, the phase error term Φ should also be estimated and compensated for during the imaging process for focused SAR imaging. This is an optimization problem, i.e., [41]:
Φ k = argmin Φ Y Φ D G X 2 2 + λ T x 1
Therefore, after the iterative optimization of the SAR image scene X based on the reconstruction algorithm, this paper utilizes the solution method in the literature [45] for the estimation of phase error compensation.
Φ k = d i a g e x p j a n g l e n = 1 N Φ k 1 D G X k * Y
where Φ k 1 is the phase error estimated from the previous iteration, here a diagonal matrix. A related module is introduced to estimate and compensate for the phase error, and the phase error is eliminated by k iterations of Equation (14).
As mentioned above, by combining sparse SAR image reconstruction with an autofocus iterative solution, SAR autofocusing imaging is realized. However, there are still some problems with this approach. The soft threshold function S · based on l 1 -norm corresponds to the sparse regularization of the “1” regularizer. This method often relies on the sparsity of the scene, that is, the threshold function is closely related to the sparsity of the output image. In general, the rarer the image, the higher the threshold, making it difficult to adapt to complex scenes.
In order to better adapt to the reconstruction of complex scenes, this paper uses a nonlinear operator P as a non-fixed L p regularization to replace the soft threshold function S · , which is used to constrain the uncertain transform domain of the target scene. The traditional L 1 regularization is to process the data of the model by determining the norm and the hyperparameter of the fixed value in the preset mathematical form. The P operator is introduced as a data-driven adaptive module, which can dynamically adjust the regularization mode and parameters. Through the end-to-end learning of nonlinear sparse transform and dynamic parameters, the algorithm can flexibly adapt to the sparse characteristics of different scenes. This design breaks through the dependence of traditional compressed sensing methods on the hypothesis of scene sparsity and provides a new idea for the practical application of multi-scene SAR. The corresponding solution of Equation (4) can be rewritten as
X k = I m a s k + ρ I M × N 1     D H Φ H Y + ρ G Z k 1 U k 1 Z k = P X k + U k U k = U k 1 + η X k Z k
For the second sub-equation of Equation (15), the gradient descent algorithm is utilized to approximate the solution, viz:
Z k ,   i = 1 l r ρ Z k ,   i 1 + l r ρ X k + U k l = 1 L l r T l H H l T l ( Z k ,   i 1 )
where l r is the step size of the gradient descent algorithm, and T l and T l H stand for convolution layer 1 and convolution layer 2, respectively. H l   denotes the corresponding nonlinear transformation in a regularized function, and the iterations are indexed by i { 1,2 , , N t } . Equation (16) shows the iterative calculation steps of the Z module.
Therefore, by jointly solving the sparse SAR reconstruction and autofocus problems, SAR autofocusing imaging under the condition of sparse sampling and phase error can be realized. However, the parameters ρ , η of the algorithm solution need to be pre-set, and it is difficult to choose an appropriate parameter to adapt to the reconstruction needs of complex scenes.

3.2. Construction of the Proposed Network

Based on the advantages of data-driven deep networks, we design the proposed algorithmic model as a corresponding network.
Firstly, the parameters in the network are set to be learnable. Through training, the corresponding gradient updates are carried out using backpropagation in deep networks to optimize the parameters. Secondly, the data-driven characterization of different sparse transformations overcomes the lack of reconstruction accuracy of traditional methods based on L 1 regularization [33] for complex scene reconstruction. It should be noted that the parameters in the algorithmic network do not need to be set manually based on experience but can be learned from data, which avoids the difficulty of parameter setting and improves the adaptability and robustness of the algorithm to different scenes.
In addition, the traditional network is usually end to end to realize the reconstruction of the echo to the image, mapping the input echo to the reconstruction result as a corresponding function through training. The corresponding network design is complex and often less adapted to radar systems of different bands. In this paper, there is a separable network design for different radar systems, which separates the design of the imaging operator from the network design. That is, we utilize a network design that combines the sparse reconstruction of raw echoes with conventional MF algorithms, and the reconstruction of the echoes is separable from the imaging process, so as to select the appropriate imaging operators for different radar systems.
The detailed network structure is shown in Figure 2. After the input, the echo signal enters the initial stage X ( 0 ) and interacts with the subsequent module through the inverse operator connection of the approximate operator in Equation (9). The network proposed in this paper consists of four modules and is structured into K stages. At each stage, all four modules perform their corresponding computations. Taking the one stage as an example, the X-module is responsible for azimuth feature extraction, including approximation operators; the Z-module performs range-direction optimization and sparse constraints; the U-module handles dynamic parameter updating; and the Φ -module is dedicated to phase error estimation and self-focusing. The final image is obtained after the four modules complete K stages iterations. The operation diagram corresponding to the computational details of each module is shown in Figure 3, utilizing the deep network to realize the solution of the iterative scheme derived in Section 2. The components and roles of these four modules are described in detail below.
  • X-module: It mainly contains four inputs, the input echo y and the upper layer of module outputs Φ , Z, and U. The forward propagation process of the data refers to the first sub-equation in Equation (15) above. The parameter ρ in the network reconstruction layer is a learnable parameter that can be optimized and updated by backpropagation during the training process. It should be noted that during the network design process, the approximate observation module in the X-module is designed separately for better adaptation to different radar systems. The phase parameters of the approximate observation operators do not participate in the training learning. The X module primarily completes the initial imaging of the region, providing a data reference for subsequent optimization;
  • Z-module: After completing the iterations in the X-module, the results are input into the Z-module. It mainly contains two inputs, which are the output of the X-module in this iteration, and the output of the U-module in the previous iteration. The module mainly consists of two convolutional layers ( C 1 and C 2 ) and a nonlinear layer activation layer H , which represent convolutional feature extraction and nonlinear activation operations, respectively. The specific calculation process refers to Equation (16). In this paper, the nonlinear activation layer H is implemented by a piecewise linear function [32], which contains a series of control points p i k , q i k i = 1 J . p i k i = 1 J is uniformly set in the corresponding position located in [−1, 1], and q i k i = 1 J represents the value of the corresponding position of the p i k of the kth iteration. Its corresponding parameters can be updated to learn from the training data. By combining the X-module of this iteration and the U-module of the last iteration, the range-oriented optimization and sparse constraint of the image are completed, so as to effectively suppress range-oriented noise and interference and improve the robustness of the image;
  • U-module: It is mainly used to update parameters, and the specific calculation process refers to the third sub-equation of Equation (15), where parameter η is also a learnable parameter, and its forward propagation and input–output relation are shown in Figure 3. The U-module dynamically updates the differences between azimuth-oriented feature X ( k ) and range-oriented feature Z ( k ) by learning parameters η , balances sparse constraints and data fidelity, prevents overfitting or underfitting, and feeds back the update results to the next stage, supports multi-stage iterative optimization, and improves imaging accuracy and network stability;
  • Φ -module: It is primarily used to estimate phase errors, which are determined based on the optimized reconstructed image. This process evaluates whether the input echo contains phase errors. The specific calculation method is detailed in Equation (14). The phase error matrix Φ ( k ) is estimated by minimizing the difference between the observed data Y and the reconstructed scene G { X ( k ) } . Specifically, Φ ( k ) is a diagonal matrix whose diagonal elements represent the phase error correction factors for each sampling point. The forward propagation process and the input–output relationship are illustrated in Figure 3. The corrected phase parameter Φ ( k ) is then fed back to the X -module, enabling multi-stage iterative optimization. This feedback mechanism estimates the phase error in this stage and feeds it back to the iteration in the next stage, and it accurately removes its influence through multiple estimates to improve the imaging quality. However, it is worth noting that the module has a limited ability to estimate large-range errors. In the future, the ability to correct large-range errors can be improved by introducing more complex error models (such as polynomial models) or joint optimization strategies.
Figure 2. Network structure.
Figure 2. Network structure.
Remotesensing 17 01279 g002
Figure 3. Structures of the modules in the one stage of the network.
Figure 3. Structures of the modules in the one stage of the network.
Remotesensing 17 01279 g003

3.3. Network and Experimental Background

  • Loss Functions: In this paper, the loss function of the proposed network is defined as
    L o s s N = 1 N n = 1 N X r e X l a F X l a F
    where X r e is the reconstructed scene output by the network, X l a is the labeled image generated by the fully sampled echo using the traditional MF algorithm, and N represents the number of echoes included in the dataset. · F is the Frobenius norm. The loss function uses the normalized Frobenius norm to measure the difference between the reconstructed image and the traditional matched filter (MF) label globally, ensuring that the target and the background scattering intensity match accurately. Normalization eliminates amplitude differences and improves training stability. MF results are labeled to inherit physical model interpretability and optimize MF defects. Loss-guided network modules (such as sparse optimization of the X/Z-module and phase correction of the Φ -module) cooperate to improve the imaging quality and system adaptability to complex scenes.
  • Structural Parameters: The parameters in the network mainly include two categories; one is the parameters of ρ , η , and the other is the parameters of the convolutional layer and nonlinear layer of the network. In the algorithm for solving Equation (15), there are unknown parameters: the penalty parameter ρ and the Lagrangian update step size η . ρ and η balance data fidelity and sparse constraints, control multiplier update speed, and affect reconstruction accuracy, convergence speed, and image detail. Reasonable selection can improve imaging quality and algorithm robustness. In this paper, the parameter is optimized by the model-training situation and is not a fixed value. Compared with the traditional fixed parameterization method, the proposed method can optimize the parameters during network training, so as to better adapt to different imaging tasks. Although it can find optimal solutions based on network learning, their initial values must be set before the algorithm iteration is performed. The penalty function ρ and update step are set according to the default of the classical ADMM algorithm, and the network convergence is the fastest and the training is stable [33]. These parameters are designed to be learnable, and the optimized model parameters do not need to be reset for different scenes.
  • Backpropagation and gradient calculation: Since both SAR echo data and SAR image data are complex forms, a back propagation (BP) algorithm is used to train the network through the loss function and complex domain. For any complex matrix O with a real-valued function f ( O ) in the corresponding SAR complex data, the derivative of f ( O ) with respect to O can be calculated by the complex derivation formula:
    f O O = f O R e O + j f O I m O
    where R e O and I m { O } denote the real and imaginary parts of the complex matrix O , respectively. Using the derivation formula can then be used to solve for the gradient in the complex domain. The parameters can be updated according to the computed loss function. In this paper the adam optimizer is used for the gradient update.
  • Network Configuration: The total number of iterations K is set to 4. The regularization term N t is set to 1. The convolution channel L is set to 16. The control point J is set to 10. The size of the convolution kernel is 5 × 5. The learning rate is initially set to 0.001, adjusted using the adaptive learning rate, and updated every 50 epochs.
  • Experimental Data: Due to the small amount of measured radar data, it is difficult to train the network with measured echoes. Therefore, the network model is trained through using SAR simulation echoes of random point targets and complex surface targets. The specific radar parameter settings used in the training dataset are shown in Table 1:
6.
Evaluation Index: For a quantitative evaluation of the reconstruction results, this paper adopts different evaluation criteria according to different scenarios of point targets and surface targets. For the evaluation of the quality of point target reconstruction, some basic point target analysis indexes [38], such as impulse response width (IRW) and peak sidelobe ratio (PSLR), are selected to conduct a quantitative evaluation of the reconstructed point target. In addition, normalized mean square error (NMSE), peak signal-to-noise ratio (PSNR), structural similarity (SSIM), and image entropy (En) are used to evaluate the reconstruction quality of surface targets. Reconstructed image indicators corresponding to surface targets are defined as follows [41]:
N M S E = X r e X l a F 2 X l a F 2
where X r e and X l a are, as described in the previous section, the reconstructed and labeled scenes, respectively. The definition of PSNR is as follows [41]:
P S N R = 10 l o g 10 1 M S E
where M S E = X r e X l a F 2 N X , and N X is the total number of pixel points in the scene. The definition of SSIM is as follows [41]:
S S I M = 2 μ x r e μ x l a + C 1 2 σ x r e x l a + C 2 μ x r e 2 + μ x l a 2 + C 1 σ x r e 2 + σ x l a 2 + C 2
where μ x r e , μ x l a , σ x r e , σ x l a , σ x r e x l a are the local means, standard deviations, and cross-covariance for images X r e and X l a . C 1 and C 2 are set as C 1 = ( K 1 L ) 2 , C 2 = ( K 2 L ) 2 , where K 1 = 0.01 , K 2 = 0.03 , and L is the range of pixel values of the image, which is set to 0~1.
E n = i , j I , J X i , j 2 log X i , j 2 i , j I , J X i , j 2 + log i , j I , J X i , j 2
where i , j I , J X i , j 2 represents the total energy of the reconstruction result.
For the evaluation of point target image quality, a numerical comparison can be made directly with the theoretical value. For the evaluation of surface targets, this can be known from the definition of index parameters, in which NMSE and PSNR can evaluate the quality of image reconstruction. SSIM represents the advantages and disadvantages of feature reconstruction, and En is used to evaluate the quality of image focusing. The smaller the NMSE and En, the better the image quality; and the larger PSNR and SSIM, the better the image quality.

4. Results

To verify the superiority of the proposed method, we conduct experiments on both simulated and real SAR echo data of different bands. In this article, the traditional MF method (i.e., the Chirp Scaling algorithm), the modified l 1 -ADMM algorithm, and ISTA-NET are compared with the proposed method. The traditional l 1 -ADMM brings huge computational costs, which consequently makes the large-scale reconstruction task computationally impossible. And ISAT-NET still needs more iterations when dealing with high-dimensional data or complex scenes. And the performance of the ISAT-NET method depends on the quality and distribution of the training data, which makes it less adaptable to the scene. Therefore, a modified l 1 -ADMM imaging method combined with a approximate observation operator is used as a comparison method.

4.1. Experiment with Simulation Data

In the simulation experiment, random azimuthal phase errors π / 2 , π / 2 are added to the echoes, and the reconstruction results under different sampling rates are tested.
In the simulation point target test, we perform echo simulation on some random point target scenes (i.e., these scenes only contain random scattering points, and do not involve the reconstruction of complex surface targets) and use the generated echoes to reconstruct the target scene by different methods, and the reconstruction results are shown in Figure 4.
The specific experimental results are shown in Figure 4. It can be seen from the reconstructed images that the conventional MF method has a defocusing problem under the condition of a sparse azimuth sampling, which leads to the degradation of azimuth resolution, especially under the condition of a low sampling rate. In addition, the image quality is also affected by the error of the surrounding scattering points, resulting in the appearance of false scattering points near the target scattering point. The modified l 1 -ADMM method can reduce the defocusing problem caused by sparse sampling, but the method is affected by the imaging scene and the algorithm threshold setting. In order to achieve a better processing effect, you need to set different thresholds for different scenarios. Although the ISTA-NET method can handle sparse reconstruction problems, its convergence speed is slow. In sparse sampling scenarios, ISTA is more sensitive to noise and phase errors, and the reconstruction results are prone to artifacts and resolution loss. In contrast, the proposed method can solve the defocusing problems caused by sparse sampling and phase errors. And it does not need to manually set the corresponding threshold.
In order to further analyze the reconstruction effect of different methods, the evaluation indexes, such as PSLR and IRW, are used to analyze the point target in the center right of Figure 4 under different sampling rates. The point target profiles are given in Figure 5, and the specific point target analysis values are shown in Table 2. It can be seen that in the azimuth direction, the image quality of the MF method is seriously degraded due to the phase error and sparse sampling. In particular, with a sampling rate of 20%, its IRW is only 5.4103 m, and the PLSR is −7.2782 dB, which are much different from the theoretical values. As the sampling rate increases, its IRW and PLSR are not up to the theoretical values, although they are improved. Compared with the traditional method, the l 1 -ADMM method shows some improvement, but there are still defocus issues and even spurious points. Specifically, its IRW improves with the sampling rate, and it meets the requirement when the sampling rate is greater than 60%. In addition, its PLSR is the best in all methods, which shows that l 1 -ADMM has a better suppression effect on the sidelobes. ISTA-NET’s method gives a better IRW and PSLR at low sampling rates, while when the sampling rate increases, both its IRW and PSLR deteriorate, which may be due to the occurrence of overfitting. For the proposed method, at a sampling rate of 20%, its IRW is closest to the theoretical value of 2.6401 m in all methods, although it does not meet the theoretical resolution requirements, and when the sampling rate is increased to 40%, the IRW meets the requirements. In addition, the PLSR of the proposed method is close to the theoretical value and is basically unaffected by the change in sampling rate. In a word, the proposed method has the best performance.

4.2. Results with Simulation Surface Data

In order to test the adaptability of the network to complex scenarios, the MSAR dataset in [46] is selected to simulate the surface target echo data test. The reconstruction results of different methods are shown in Figure 6, and the image reconstruction is evaluated based on evaluation metrics, such as loss, PSNR, SSIM, and En.
From the results of the surface target simulation data processing, the MF method is affected by phase error and sparse sampling, and the image after processing is seriously defocused, which makes it difficult to distinguish the target. The l 1 -ADMM method is less affected by the sparse sampling defocus, but the image quality degrades seriously in the presence of phase error. ISTA-NET image quality is better compared to the previous two, but the presence of bokeh can lead to a loss of detail. In addition, the scene reconstructed by the l 1 -ADMM method is darker, and some weaker target features may be lost in the reconstruction process due to different threshold settings. For example, for the scene at the yellow box in Figure 6, whose partially enlarged image is displayed in Figure 6d, it is difficult to identify its scene targets. In contrast, the proposed method can overcome the influence of phase error in echo well, can reconstruct target features including weak scattering points, and is less affected by a sparse sampling rate.
In order to further analyze the quality of image reconstruction, the imaging results obtained by different reconstruction methods are quantitatively analyzed, and the results are shown in Table 3. In terms of loss, PSNR, and SSIM, the proposed method has the best performance in image feature reconstruction compared with the other three methods. For example, at a sampling rate of 20%, the loss, PSNR, and SSIM are 0.3017, 39.1558 dB, and 0.9564, respectively, which are better than the image reconstruction values of MF, l 1 -ADMM, and ISTA-NET. Due to the lack of sparse representation ability, the l 1 -ADMM method ignores background information in the reconstruction process. This may cause the weak scattering targets in scenes not to be reconstructed, which results in the loss of target scene information, causing a decrease in the entropy value of the image. Therefore, the l 1 -ADMM method has the lowest entropy value in terms of En. In Table 3, the main reason why the image quality of ISTA-Net does not deteriorate with the increase in sampling rate is due to the difference in image sparsity. When the network processes highly sparse images, it can already meet the sparse reconstruction conditions of compressed sensing under low sampling rates, but high sampling will introduce redundant information, resulting in network confusion or overfitting noise, which will degrade the quality. For analysis accuracy, the phase estimation module is analyzed below. The estimated results of the actual phase error and error estimation module are shown in Figure 7. It can be seen from Figure 7a that the phase error estimated by the Φ -module is basically equal to the actual error. Figure 7b shows the estimated error between the two. From the figure, we can see the relative error result after the Φ -module estimates and eliminates the original error. In this case, the residual is only ( π / 4 , π / 4 ) . The proposed method can effectively suppress the phase error under the limited number of iterations, and the estimated error is within the acceptable range of the actual SAR system, and the imaging effect is obviously better than the existing methods. The residuals can be further optimized by introducing adaptive convergence criteria or higher-order phase models.
In contrast, the proposed method can not only eliminate the defocusing caused by the phase error but also retain scene detail information in the reconstruction process. Therefore, the proposed method has the best reconstruction performance under the comprehensive index.

5. Discussion

5.1. Experiment with Measured Data

In order to verify the effectiveness of the proposed method in this paper, the measured data of different bands are processed. When testing radar-measured data of different bands, it is only necessary to adjust and match the corresponding phase parameter data, rather than train the radar data of different bands one by one. It should be noted that the data size of the training network is fixed, so for data of different specifications, some methods can be used to unify its data size, such as cropping or complementary zero. Specifically, this sampling rate is set to 20%, with some phase error in the azimuth direction.
The SAR data of the stationary ground scene were acquired in Mianyang City, Sichuan Province, China, on June 2011 by the X-band airborne dual-antenna SAR system, which was developed by the Institute of Electronics, Chinese Academy of Sciences. The system parameters for the measured radar data are shown in Table 4.
For the surface target experimental data, three scene echoes are selected for testing, and their experimental results are shown in Figure 8. As can be seen from Figure 8, the image reconstruction quality of the traditional MF method for the three scenes is not good due to the sampling sparsity and phase error. For example, the objects in scenario 1 are indistinguishable, and the objects in scenario 2 appear defocused, reducing the image resolution. Among them, the target outline in scenario 3 is indistinguishable. Compared with the traditional MF method, the l 1 -ADMM method can reconstruct the target scene to a certain extent, but its reconstruction performance is greatly affected by the empirical threshold. In order not to lose the target feature information, the threshold value for reconstructing the surface target scene should be smaller than that for reconstructing the point target scene, that is, the threshold needs to be set according to the scene type. The ISTA-NET method provides a large improvement in the clarity of imaging over the previous two. The reconstruction effect is even better than the method proposed in this paper in the imaging of scenario 1, but its scene adaptation is poor and the imaging quality in scenario 2 and scenario 3 is degraded. In contrast, the reconstruction performance of the proposed method is best, and one does not need to set the threshold according to the scene type.
In order to quantitatively analyze the reconstruction results of different methods, reconstruction indexes such as PSNR, SSIM and En are given in Table 5. Among them, the proposed method performs better than the other three methods in scenarios 2–3. For example, scenario 3 has the smallest loss and the highest PSNR and SSIM, which are 0.1649, 35.3073, and 0.9733, respectively. And due to the characteristics of ISTA-NET in terms of strong recovery of sparsity, the image loss obtained by ISTA-NET is lower in scenario 1. However, the method proposed in this paper focuses on the robustness of a wide range of scenes and performs more comprehensively in complex scenes. In addition, it can be seen from the entropy value that the entropy of the proposed method is the lowest, which seems to be different from the simulation results. Different from sparse scenarios, in order to obtain richer background information when dealing with complex scenarios, the threshold of l 1 -ADMM is set lower than the simulation data in the previous section. This also shows that the l 1 -ADMM method needs parameters to be reset according to different scenes. In contrast, the proposed method does not need the threshold value to be reset but optimizes the parameters during the training process and has a better scene adaptation and robustness compared to ISTA-NET.

5.2. Results with Measured Data in THz Band

To verify the validity of the proposed method, real airborne THz-SAR-measured data were gathered. The data were acquired by the 0.22-THz airborne stripmap SAR system developed by the Beijing Institute of Radio Measurement. The system parameters for the measured radar data are shown in Table 6. The data scene was a GFTHz-shaped real scene consisting of strong scattering point targets.
The experimental results of different methods are shown in Figure 9. As can be seen from Figure 9, the image reconstructed by the traditional MF method has serious defocus in the azimuth, resulting in a decline in image resolution. Compared with the traditional MF method, the l 1 -ADMM method can improve the imaging effect, but because of the existence of phase error, the image information is destroyed, and some point targets cannot be reconstructed, such as the letters F and T. While the ISTA-NET method is able to reconstruct the image, the target edges are not sharp enough and produce more severe noise, making it difficult to ensure details in the imaging. In contrast, the proposed method has the best effect in image reconstruction. However, it should be noted that the training of the network model in this paper is optimized for a certain range of azimuth phase error. Therefore, the reconstruction effect of the proposed method is not ideal for some point targets with large azimuth phase errors, while the reconstruction effect of the proposed method is good for some point targets with small errors. Therefore, the proposed method is best combined with the error compensation method based on measurement data, so that the imaging effect will be better.
In order to further analyze the imaging quality of the target, the point targets in the red circle in Figure 9. are selected for quantitative analysis, and their profiles and quantitative indexes are shown in Figure 10 and Table 7, respectively. It can be seen that for point target 2, the IRW and PSLR of the MF method are 0.2104 and −8.6605 dB, respectively, which are significantly different from the theoretical values of 0.0998 and 13.2600. In addition, the IRW and PLSR of the l 1 -ADMM method are improved compared with the MF method, being 0.1610 and −11.1654 dB, respectively, but they do not reach the theoretical value. The IRW and PSLR of ISTA-NET show deterioration compared to the l 1 -ADMM algorithm, being 0.1891 and −2.2068 dB, respectively, which can lead to severe detail loss. In contrast, the IRW and PSLR of the proposed method agree well with the theoretical values, which proves the effectiveness of the proposed method in sparse imaging and error compensation. This is because the design idea of modular separation in this paper significantly reduces the computational complexity and improves the generalization ability for radar systems in different bands, while retaining the efficiency of traditional imaging algorithms (such as CS). By introducing L p regularization, the algorithm can automatically optimize parameters through end-to-end training, thus reducing the dependence on manual parameter tuning. Therefore, compared with traditional methods, the algorithm proposed in this paper uses data-driven self-optimization parameters to improve the robustness and adaptability of imaging in different scenes and different bands. The suboptimal parameters brought by manual parameter adjustment are avoided. Despite the above advantages, this method still has some defects. When processing images with large phase errors, the current phase error compensation module will have a large error. At the same time, there is a high demand for data and computing power in the training stage. If there is any problem with the training data, the imaging results will be greatly affected.

6. Conclusions

To solve the problem that existing sparse SAR imaging networks have a poor adaptive ability in different bands, a novel deep unfolding network for multi-band SAR sparse imaging and autofocusing based on approximate observation is proposed. Firstly, the approximate observation operator module is designed separately from the sparse imaging optimization network to adapt to radar systems of different bands. Compared with the fixed l 1 regularization, the proposed method adopts network training to achieve an adaptive optimization based on l p parameters to characterize the uncertain transformation domain of the target scene, which has a better reconstruction effect for complex scenes. In view of the phase errors that may be introduced in the SAR echo acquisition process, an error compensation module is added in the imaging optimization process, and the imaging and autofocusing of the target scene are realized at the same time. Finally, airborne SAR simulation and real imaging experiments in the 0.22 THz and 9.6 GHz bands are performed to verify the effectiveness of the proposed method in multi-band scenarios. In the experiment, four sparse levels of 20%, 40%, 60%, and 80% are used to compare and analyze the imaging performance of different methods in complex scenes. The results show that the proposed method is superior to the existing methods in key evaluation indexes such as loss, PSNR, SSIM, and En. In summary, the deep spread network proposed in this paper has significant advantages in the sparse imaging and self-focusing of multi-band SAR. The introduction of approximate observation operators and P operators realizes a dynamic adaptation to different band radar systems and different scenes and successfully reduces the dependence on manual parameter tuning. The introduction of a phase error compensation module further enhances the robustness of the algorithm in low sampling rate and strong noise environments, and it provides a more efficient and reliable solution for the practical application of multi-band and multi-scene SAR.
The proposed method has high training requirements for imaging and incurs a significant computational cost. In addition, the shortcoming of the proposed method is that the current phase error compensation module can only estimate and compensate for a small range of phase errors, and it cannot well estimate and compensate for a large range of errors. In the future, it is necessary to add a larger range of more complex errors in the network-training process, so as to realize the estimation of and compensation for arbitrary errors. For example, the incorporation of an adaptive convergence criterion or a higher-order phase model can further optimize the phase error.

Author Contributions

X.L., M.Z., Y.L. (Yiheng Liang), Y.L. (Yinwei Li), G.X. and B.W. contributed to the methodology proposal, experimental execution, and analysis of the results, and they were jointly involved in writing and revising the manuscript. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by the Artificial Intelligence Promotes Scientific Research Paradigm Reform and Empowers Discipline Advancement Plan under Grant Z-2024-302-061. It was also supported in part by the National Natural Science Foundation of China (Grant Nos. 61988102 and 12105177), in part by the Natural Science Foundation of Shanghai (Grant No. 21ZR1444300), and in part by the Shanghai Social Development Science and Technology Research Project (Grant No. 22dz1200302).

Data Availability Statement

The raw data supporting the conclusions of this article will be made available by the authors on request.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Curlander, J.C.; Mcdonough, R.N. Synthetic Aperture Radar: Systems and Signal Processing, 1st ed.; Wiley: Hoboken, NJ, USA, 1991. [Google Scholar]
  2. Lin, C.; Huang, P.; Wang, W.; Li, Y.; Xu, J. Unambiguous signal reconstruction approach for SAR imaging using frequency diverse array. IEEE Geosci. Remote Sens. Lett. 2017, 14, 1628–1632. [Google Scholar] [CrossRef]
  3. Xu, G.; Chen, Y.; Ji, A.; Zhang, B.; Yu, C.; Hong, W. 3-D high-resolution imaging and array calibration of ground-based millimeter-wave MIMO radar. IEEE Trans. Microw. Theory Tech. 2024, 72, 4919–4931. [Google Scholar] [CrossRef]
  4. Makarov, P.A. Two-dimensional autofocus technique based on spatial frequency domain fragmentation. IEEE Trans. Image Process. 2020, 29, 6006–6016. [Google Scholar] [CrossRef]
  5. Fornaro, G.; Franceschetti, G.; Perna, S. Motion compensation errors: Effects on the accuracy of airborne SAR images. IEEE Trans. Aerosp. Electron. Syst. 2005, 41, 1338–1352. [Google Scholar] [CrossRef]
  6. Prats, P.; Macedo, K.A.C.D.; Reigber, A.; Scheiber, R.; Mallorqui, J.J. Comparison of topography- and aperture-dependent motion compensation algorithms for airborne SAR. IEEE Geosci. Remote Sens. Lett. 2007, 4, 349–353. [Google Scholar] [CrossRef]
  7. Chen, J.; Li, M.; Yu, H.; Xing, M. Full-aperture processing of airborne microwave photonic SAR raw data. IEEE Trans. Geosci. Remote Sens. 2023, 61, 5218812. [Google Scholar] [CrossRef]
  8. Linnehan, R.; Miller, J.; Asadi, A. Map-drift autofocus and scene stabilization for video-SAR. In Proceedings of the 2018 IEEE Radar Conference (RadarConf18), Oklahoma City, OK, USA, 23–27 April 2018; pp. 1401–1405. [Google Scholar]
  9. Li, Y.; Wu, J.; Mao, Q.; Xiao, H.; Meng, F.; Gao, W.; Zhu, Y. A Novel 2-D Autofocusing Algorithm for Real Airborne Stripmap Terahertz Synthetic Aperture Radar Imaging. IEEE Geosci. Remote Sens. Lett. 2023, 20, 1–5. [Google Scholar] [CrossRef]
  10. Mao, X.; He, X.; Li, D. Knowledge-aided 2-D autofocus for spotlight SAR range migration algorithm imagery. IEEE Trans. Geosci. Remote Sens. 2018, 56, 5458–5470. [Google Scholar] [CrossRef]
  11. Kragh, J.T.; Kharbouch, A.A. Monotonic iterative algorithm for minimum-entropy autofocus. In Proceedings of the 2006 IEEE International Conference on Image Processing, Atlanta, GA, USA, 8–11 October 2006; pp. 645–648. [Google Scholar]
  12. Chen, Y.-C.; Li, G.; Zhang, Q.; Zhang, Q.-J.; Xia, X.-G. Motion compensation for airborne SAR via parametric sparse representation. IEEE Trans. Geosci. Remote Sens. 2017, 55, 551–562. [Google Scholar] [CrossRef]
  13. Ye, W.; Yeo, T.S.; Bao, Z. Weighted least-squares estimation of phase errors for SAR/ISAR autofocus. IEEE Trans. Geosci. Remote Sens. 1999, 37, 2487–2494. [Google Scholar] [CrossRef]
  14. Zhu, J.; Xie, Z.; Jiang, N.; Song, Y.; Han, S.; Liu, W.; Huang, X. Delay-Doppler Map Shaping through Oversampled Complementary Sets for High-Speed Target Detection. Remote Sens. 2024, 16, 2898. [Google Scholar] [CrossRef]
  15. Zhang, B.; Hong, W.; Wu, Y. Sparse microwave imaging: Principles and applications. Sci. China Inf. Sci. 2012, 55, 1722–1754. [Google Scholar] [CrossRef]
  16. Camlica, S.; Gurbuz, A.C.; Arikan, O. Autofocused spotlight SAR image reconstruction of off-grid sparse scenes. IEEE Trans. Aerosp. Electron. Syst. 2017, 53, 1880–1892. [Google Scholar] [CrossRef]
  17. Pu, W.; Wu, J. OSRanP: A novel way for radar imaging utilizing joint sparsity and low-rankness. IEEE Trans. Comput. Imag. 2020, 6, 868–882. [Google Scholar] [CrossRef]
  18. Crespo Marques, E.; Maciel, N.; Naviner, L.; Cai, H.; Yang, J. A review of sparse recovery algorithms. IEEE Access 2019, 7, 1300–1322. [Google Scholar] [CrossRef]
  19. Tropp, J.A.; Gilbert, A.C. Signal recovery from random measurements via orthogonal matching pursuit. IEEE Trans. Inf. Theory 2007, 53, 4655–4666. [Google Scholar] [CrossRef]
  20. Cai, T.T.; Wang, L. Orthogonal Matching Pursuit for Sparse Signal Recovery with Noise. IEEE Trans. Inf. Theory 2011, 57, 4680–4688. [Google Scholar] [CrossRef]
  21. Wei, W.; Min, J.; Qing, G. A compressive sensing recovery algorithm based on sparse Bayesian learning for block sparse signal. In Proceedings of the 2014 International Symposium on Wireless Personal Multimedia Communications (WPMC), Sydney, NSW, Australia, 7–10 September 2014; pp. 547–551. [Google Scholar]
  22. Yang, L.; Zhao, L.; Bi, G.; Zhang, L. SAR Ground Moving Target Imaging Algorithm Based on Parametric and Dynamic Sparse Bayesian Learning. IEEE Trans. Geosci. Remote Sens. 2016, 54, 2254–2267. [Google Scholar] [CrossRef]
  23. Beck, A.; Teboulle, M. A fast iterative shrinkage-thresholding algorithm for linear inverse problems. SIAM J. Imag. Sci. 2009, 2, 183–202. [Google Scholar] [CrossRef]
  24. Donoho, D.L.; Maleki, A.; Montanari, A. Message passing algorithms for compressed sensing: I. motivation and construction. In Proceedings of the 2010 IEEE Information Theory Workshop (ITW 2010), Cairo, Egypt, 6–8 January 2010; pp. 1–5. [Google Scholar]
  25. Parikh, N.; Boyd, S. Proximal algorithms. Found. Trends Optim. 2014, 1, 127–239. [Google Scholar] [CrossRef]
  26. Wu, C.; Zhang, Z.; Chen, L.; Yu, W. The Same Range Line Cells Based Fast Two-Dimensional Compressive Sensing for Airborne MIMO Array SAR 3-D Imaging. In Proceedings of the IGARSS 2019—2019 IEEE International Geoscience and Remote Sensing Symposium, Yokohama, Japan, 28 July–2 August 2019; pp. 3653–3656. [Google Scholar] [CrossRef]
  27. Fang, J.; Xu, Z.; Zhang, B.; Hong, W.; Wu, Y. Fast compressed sensing SAR imaging based on approximated observation. IEEE J. Sel. Top. Appl. Earth Observ. Remote Sens. 2014, 7, 352–363. [Google Scholar] [CrossRef]
  28. Bi, H.; Zhu, D.; Bi, G.; Zhang, B.; Hong, W.; Wu, Y. FMCW SAR sparse imaging based on approximated observation: An overview on current technologies. IEEE J. Sel. Top. Appl. Earth Observ. Remote Sens. 2020, 13, 4825–4835. [Google Scholar] [CrossRef]
  29. Zhang, J.; Lu, X.; Song, Y.; Yu, D.; Bi, H. Sparse Stripmap SAR Autofocusing Imaging Combining Phase Error Estimation and L₁-Norm Regularization Reconstruction. IEEE Trans. Geosci. Remote Sens. 2023, 61, 5205212. [Google Scholar] [CrossRef]
  30. Zhang, J.; Ghanem, B. ISTA-net: Interpretable optimization-inspired deep network for image compressive sensing. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Salt Lake City, UT, USA, 28–22 June 2018. [Google Scholar]
  31. Zhang, Z.; Liu, Y.; Liu, J.; Wen, F.; Zhu, C. AMP-Net: Denoising based deep unfolding for compressive image sensing. IEEE Trans. Image Process. 2021, 30, 1487–1500. [Google Scholar] [CrossRef]
  32. Yang, Y.; Sun, J.; Li, H.; Xu, Z. ADMM-CSNet: A deep learning approach for image compressive sensing. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 42, 521–538. [Google Scholar] [CrossRef]
  33. Li, R.; Zhang, S.; Zhang, C.; Liu, Y.; Li, X. Deep learning approach for sparse aperture ISAR imaging and autofocusing based on complex valued ADMM-Net. IEEE Sens. J. 2021, 21, 3437–3451. [Google Scholar] [CrossRef]
  34. Wei, S.; Liang, J.; Wang, M.; Shi, J.; Zhang, X.; Ran, J. AF-AMPNet: A deep learning approach for sparse aperture ISAR imaging and autofocusing. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5206514. [Google Scholar] [CrossRef]
  35. Liu, Z.; Yang, S.; Gao, Q.; Feng, Z.; Wang, M.; Jiao, L. AFnet and PAFnet: Fast and accurate SAR autofocus based on deep learning. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5238113. [Google Scholar] [CrossRef]
  36. Pu, W. SAE-Net: A deep neural network for SAR autofocus. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5220714. [Google Scholar] [CrossRef]
  37. Kang, L.; Sun, T.; Luo, Y.; Ni, J.; Zhang, Q. SAR imaging based on deep unfolded network with approximated observation. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5228514. [Google Scholar] [CrossRef]
  38. Wang, M.; Wei, S.; Liang, J.; Zeng, X.; Wang, C.; Shi, J. RMIST-Net: Joint range migration and sparse reconstruction network for 3-D mm W imaging. IEEE Trans. Geosci. Remote Sens. 2021, 60, 5205117. [Google Scholar] [CrossRef]
  39. Yin, T.; Guo, W.; Zhu, J.; Wu, Y.; Zhang, B.; Zhou, Z. Underwater Broadband Target Detection by Filtering Scanning Azimuths Based on Features of Sub-band Peaks. IEEE Sens. J. 2015, 1, 1. [Google Scholar] [CrossRef]
  40. Zhu, J.; Yin, T.; Guo, W.; Zhang, B.; Zhou, Z. An Underwater Target Azimuth Trajectory Enhancement Approach in BTR. Appl. Acoust. 2025, 230, 110373. [Google Scholar] [CrossRef]
  41. Li, M.; Wu, J.; Huo, W.; Li, Z.; Yang, J.; Li, H. STLS-LADMM-Net: A Deep Network for SAR Autofocus Imaging. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5226914. [Google Scholar] [CrossRef]
  42. Hanif, A.; Mansoor, A.B.; Imran, A.S. Sub-Nyquist sampling and detection in Costas coded pulse compression radars. EURASIP J. Adv. Signal Process. 2016, 2016, 125. [Google Scholar] [CrossRef]
  43. Cumming, I.G.; Wong, F.H. Digital Signal Processing of Synthetic Aperture Radar Data: Algorithms and Implementation, 1st ed.; Artech House: Norwood, MA, USA, 2004; p. 9787121169779. [Google Scholar]
  44. Zhang, S.; Liu, Y.; Li, X. Computationally Efficient Sparse Aperture ISAR Autofocusing and Imaging Based on Fast ADMM. IEEE Trans. Geosci. Remote Sens. 2020, 58, 8751–8765. [Google Scholar] [CrossRef]
  45. Onhon, N.Ö.; Cetin, M. A sparsity-driven approach for joint SAR imaging and phase error correction. IEEE Trans. Image Process. 2012, 21, 2075–2088. [Google Scholar] [CrossRef]
  46. Xia, R.; Chen, J.; Huang, Z.; Wan, H.; Wu, B.; Sun, L.; Yao, B.; Xiang, H.; Xing, M. CRTransSar: A Visual Transformer Based on Contextual Joint Representation Learning for SAR Ship Detection. Remote Sens. 2022, 14, 1488. [Google Scholar] [CrossRef]
Figure 1. Model of SAR imaging system.
Figure 1. Model of SAR imaging system.
Remotesensing 17 01279 g001
Figure 4. Point target imaging results. The first to fourth rows represent 20%, 40%, 60%, and 80% sampling rates, respectively. (a) MF, (b) l 1 -ADMM, (c) ISTA-NET, (d) ours.
Figure 4. Point target imaging results. The first to fourth rows represent 20%, 40%, 60%, and 80% sampling rates, respectively. (a) MF, (b) l 1 -ADMM, (c) ISTA-NET, (d) ours.
Remotesensing 17 01279 g004
Figure 5. Performance evaluation of point target in center right of Figure 4. (ad) Range and azimuthal point target profiles at 20%, 40%, 60%, and 80% sampling rates, respectively.
Figure 5. Performance evaluation of point target in center right of Figure 4. (ad) Range and azimuthal point target profiles at 20%, 40%, 60%, and 80% sampling rates, respectively.
Remotesensing 17 01279 g005aRemotesensing 17 01279 g005b
Figure 6. The first to fourth rows represent 20%, 40%, 60%, and 80% sampling rates, respectively. (a) MF, (b) l 1 -ADMM, (c) ISTA-NET, (d) ours.
Figure 6. The first to fourth rows represent 20%, 40%, 60%, and 80% sampling rates, respectively. (a) MF, (b) l 1 -ADMM, (c) ISTA-NET, (d) ours.
Remotesensing 17 01279 g006aRemotesensing 17 01279 g006b
Figure 7. Phase error: (a) actual phase and estimate phase (b) estimated error.
Figure 7. Phase error: (a) actual phase and estimate phase (b) estimated error.
Remotesensing 17 01279 g007
Figure 8. Imaging results of 20% sampling rate for different surface targets. (a) MF, (b) l 1 -ADMM, (c) ISTA-NET, (d) ours.
Figure 8. Imaging results of 20% sampling rate for different surface targets. (a) MF, (b) l 1 -ADMM, (c) ISTA-NET, (d) ours.
Remotesensing 17 01279 g008aRemotesensing 17 01279 g008b
Figure 9. THz-SAR imaging results of multiple corner reflectors. (a) MF, (b) l 1 -ADMM, (c) ISTA-NET, (d) ours.
Figure 9. THz-SAR imaging results of multiple corner reflectors. (a) MF, (b) l 1 -ADMM, (c) ISTA-NET, (d) ours.
Remotesensing 17 01279 g009
Figure 10. Performance evaluation of point targets of Figure 9: (a) target 1, (b) target 2.
Figure 10. Performance evaluation of point targets of Figure 9: (a) target 1, (b) target 2.
Remotesensing 17 01279 g010
Table 1. SAR system parameters for training data.
Table 1. SAR system parameters for training data.
ParameterValue
Carrier frequency9.6 GHz
Signal bandwidth60 MHz
Pulse width4 μs
Fast-time sampling80 MHz
Synthetic aperture time1.6 s
Radar platform velocity80 m/s
Pulse repetition frequency40 Hz
Table 2. Point target performance values comparison under different sampling rates.
Table 2. Point target performance values comparison under different sampling rates.
SampleMethodIRW/m [Azimuth]PSLR/dB [Azimuth]
100%label2.6401−13.2603
20%MF5.2509−5.6680
l 1 -ADMM4.3346−18.4429
ISTA-Net3.0911−13.5845
Ours3.2715−13.3593
40%MF3.6671−5.6551
l 1 -ADMM3.0002−23.6377
ISTA-Net3.2071−9.0388
Ours2.6147−13.7778
60%MF2.8678−8.9854
l 1 -ADMM2.5160−15.6035
ISTA-Net3.9686−6.6601
Ours2.6205−13.1205
80%MF2.7052−10.2679
l 1 -ADMM2.5646−13.9712
ISTA-Net4.3519−5.9550
Ours2.5999−13.8977
The bold text represents the optimal value under the current sampling rate.
Table 3. Surface target performance values comparison under different sampling rates.
Table 3. Surface target performance values comparison under different sampling rates.
SampleMethodLossPSNRSSIMEn
20%MF1.589528.30580.691912.1035
l 1 -ADMM0.733634.33970.68968.7402
ISTA-Net0.267531.24460.940010.8770
Ours0.210339.15580.956410.6137
40%MF1.207130.22540.754912.0185
l 1 -ADMM0.664434.68750.73019.1305
ISTA-Net0.231731.33680.953010.9314
Ours0.215240.89720.971610.5412
60%MF1.144130.67560.764112.0557
l 1 -ADMM0.606134.57140.75339.5090
ISTA-Net0.250831.04990.927510.7917
Ours0.211741.94210.977910.5807
80%MF0.848432.42940.822711.9008
l 1 -ADMM0.560234.69280.76879.7558
ISTA-Net0.224631.21090.969510.6674
Ours0.164441.58400.982710.7998
The bold text represents the optimal value under the current sampling rate.
Table 4. SAR system parameters in X-band.
Table 4. SAR system parameters in X-band.
ParameterValue
Carrier frequency9.6 GHz
Signal bandwidth492 MHz
Pulse width13 μs
Fast-time sampling600 MHz
Synthetic aperture time0.6 s
Radar platform velocity105 m/s
Pulse repetition frequency3125 Hz
Table 5. X-band performance values comparison.
Table 5. X-band performance values comparison.
MethodLossPSNRSSIMEn
Scene 1MF0.674224.81960.787212.1657
l 1 -ADMM0.468528.22080.906812.1259
ISTA-Net0.240830.10580.911812.0311
Ours0.356430.03140.920111.9226
Scene 2MF0.330729.20210.876212.0678
l 1 -ADMM0.212831.82850.913211.9612
ISTA-Net0.231930.76370.915211.9130
Ours0.176533.62480.965611.8108
Scene 3MF0.473928.61830.866711.9115
l 1 -ADMM0.237032.23420.919111.6393
ISTA-Net0.172131.94170.927611.5030
Ours0.164935.30730.973311.3485
The bold text represents the optimal value under the current sampling rate.
Table 6. SAR system parameters in THZ band.
Table 6. SAR system parameters in THZ band.
ParameterValue
Carrier frequency214 GHz
Signal bandwidth900 MHz
Pulse width0.15 μs
Fast-time sampling1.12 GHz
Synthetic aperture time0.1 s
Radar platform velocity67 m/s
Pulse repetition frequency1250 Hz
Table 7. THz-band performance values comparison.
Table 7. THz-band performance values comparison.
MethodIRW/mPSLR/dB
Target 1Target 2Target 1Target 2
Theory0.0998−13.2600
MF0.16620.2104−9.7297−8.6605
l 1 -ADMM0.15060.1680−11.9509−11.1654
ISTA-NET0.08500.1891−2.8339−2.2068
Ours0.09990.1035−15.9870−11.3810
The bold text represents the optimal value under the current sampling rate.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Li, X.; Zhan, M.; Liang, Y.; Li, Y.; Xu, G.; Wang, B. A Novel Deep Unfolding Network for Multi-Band SAR Sparse Imaging and Autofocusing. Remote Sens. 2025, 17, 1279. https://doi.org/10.3390/rs17071279

AMA Style

Li X, Zhan M, Liang Y, Li Y, Xu G, Wang B. A Novel Deep Unfolding Network for Multi-Band SAR Sparse Imaging and Autofocusing. Remote Sensing. 2025; 17(7):1279. https://doi.org/10.3390/rs17071279

Chicago/Turabian Style

Li, Xiaopeng, Mengyang Zhan, Yiheng Liang, Yinwei Li, Gang Xu, and Bingnan Wang. 2025. "A Novel Deep Unfolding Network for Multi-Band SAR Sparse Imaging and Autofocusing" Remote Sensing 17, no. 7: 1279. https://doi.org/10.3390/rs17071279

APA Style

Li, X., Zhan, M., Liang, Y., Li, Y., Xu, G., & Wang, B. (2025). A Novel Deep Unfolding Network for Multi-Band SAR Sparse Imaging and Autofocusing. Remote Sensing, 17(7), 1279. https://doi.org/10.3390/rs17071279

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop