Next Article in Journal
Coupling Flank Collapse and Magma Dynamics on Stratovolcanoes: The Mt. Etna Example from InSAR and GNSS Observations
Next Article in Special Issue
Real Representation of the Polarimetric Scattering Matrix for Monostatic Radar
Previous Article in Journal
Mountain Tree Species Mapping Using Sentinel-2, PlanetScope, and Airborne HySpex Hyperspectral Imagery
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Very High Resolution Automotive SAR Imaging from Burst Data

by
Mattia Giovanni Polisano
1,*,
Marco Manzoni
1,
Stefano Tebaldini
1,
Andrea Monti-Guarnieri
1,
Claudio Maria Prati
1 and
Ivan Russo
2
1
Department of Electronics, Information and Bioengineering (DEIB), Politecnico di Milano, 20133 Milan, Italy
2
Huawei Technologies Italia S.r.l., 20129 Segrate, Italy
*
Author to whom correspondence should be addressed.
Remote Sens. 2023, 15(3), 845; https://doi.org/10.3390/rs15030845
Submission received: 18 January 2023 / Accepted: 30 January 2023 / Published: 2 February 2023

Abstract

:
This paper proposes a method for efficient and accurate removal of grating lobes in automotive Synthetic Aperture Radar (SAR) images. Grating lobes can indeed be mistaken as real targets, inducing in this way false alarms in the target detection procedure. Grating lobes are present whenever SAR focusing is performed using data acquired on a non-continuous basis. This kind of acquisition is typical in the automotive scenario, where regulations do not allow for a continuous operation of the radar. Radar pulses are thus transmitted and received in bursts, leading to a spectrum of the signal containing gaps. We start by deriving a suitable reference frame in which SAR images are focused. It will be shown that working in this coordinate system is particularly convenient since it allows for a signal spectrum that is space-invariant and with spectral gaps described by a simple one-dimensional function. After an inter-burst calibration step, we exploit these spectral characteristics of the signal by implementing a compressive sensing algorithm aimed at removing grating lobes. The proposed approach is validated using real data acquired by an eight-channel automotive radar operating in burst mode at 77 GHz. Results demonstrate the practical possibility to process a synthetic aperture length as long as up to 2 m reaching in this way extremely fine angular resolutions.

Graphical Abstract

1. Introduction

Advanced Driver Assistant Systems (ADAS) are increasingly being considered in the context of the research towards autonomous driving vehicles. By and large, ADAS require abundant and reliable data acquired by various sensors, such as cameras, LiDARs, and radars. Automotive radars, in particular, have proven to be excellent tools for sensing the surrounding environment thanks to the day and night imaging capability, and the ability to measure distances and velocities with great accuracy. Although radars are used for ADAS systems, the interest in the development of this technology in the automotive scenario has been growing widely in the last decades [1,2,3,4,5].
Current automotive radars generally provide coarser resolution than optical and LiDAR sensors, which is perceived as an intrinsic limitation of such technology. Indeed, vehicle-based radar imaging is typically implemented using Multiple-Input Multiple-Output (MIMO) devices, for which angular resolution is bounded by the number of Tx/Rx channels forming the MIMO array. For this reason, achieving fine angular resolution requires the implementation of more complex, large, and expensive devices. Nowadays, radars used for automotive purposes operate in the W-band [6] (above 76 GHz) and are mostly used for crash avoidance and object detection at short, medium or long range depending on their wavelength.
To overcome such limitations, recent researches have focused on the concept of synthetic aperture radar (SAR), where an arbitrarily long array is formed by exploiting the natural motion of the vehicle [1,2,3,4]. Still, a necessary condition for any conventional SAR imaging algorithm is that the radar pulses are acquired continuously, meaning that the device is constantly operating. Unluckily, this is not the case in the current automotive scenarios: due to radio-frequency regulations and hardware requirements, radar devices are typically designed to operate on a non-continuous basis by sending bursts of few hundreds of pulses alternated with periods of silence. This acquisition mode creates gaps in the spatial spectrum of SAR images, which manifest themselves through the appearance of grating lobes in SAR images, leading to false target detections. This problem is typical of SAR imaging and it has been widely covered in the literature [7,8].
The aim of this paper is to design, implement and test a routine able to remove grating lobes from SAR images generated by processing data acquired in bursts. The goal is achieved with several steps. First of all, a space-invariant reference system is derived in which the spectrum of the data depends uniquely on the position of the sensor and not on the position of the target. Unlike more common reference systems such as the Cartesian or the polar ones, this reference frame allows us to see the gaps in the spectrum of the data. The high resolution SAR image (still showing grating lobes) will be the coherent sum of the bursts. Each burst, however, must be focused properly using autofocusing techniques [3] in order to correct for possible trajectory errors. Then, each burst must be phase calibrated [9,10] in order to allow for a fully coherent summation. Both steps are described carefully in this paper. Finally, the removal of grating lobes is achieved with the CLEAN algorithm, developed in [7] for speckle and grating lobes suppression in astronomy and it is widely used also in telecommunications [11,12,13,14].
This paper is organized as follows: Section 2 covers the geometry and the signal model, and the basic notions of automotive SAR are dealt. Section 3 regards the wavenumber domain in SAR imaging, where the theory regarding the spatial spectral domain characteristics of a burst acquired signal are covered. In Section 4, the new mathematical framework which is the corner stone of this work and that allows a space-invariant spatial spectrum descriptor is developed, also the autofocus, phase calibration technique and the grating lobes suppression algorithm are described in this section. Section 5 shows the real data results used to validate this work and finally in Section 6, the conclusions are drawn.

2. Geometry, Signal Model and SAR Processing

In this paper we will consider a MIMO radar having N T X transmitting antennas and N R X receiving ones, forming an equivalent monostatic virtual array of N c h = N T X × N R X elements. The elements and their spatial locations are represented in Figure 1.
The radar is mounted with a arbitrary installation angle on the car. A forward looking geometry is indicated for target detection and navigation assistance; a side looking geometry is, instead, more indicated for parking detection, while a squinted geometry can be useful for collision avoidance. From now on, we will consider a forward looking geometry with the virtual elements displaced along the y direction.
It is well known that a system such as the one just described will provide a slant range resolution equal to [15]:
ρ r = c 2 B
where c is the speed of light and B is the bandwidth of the radar pulse. The angular resolution, instead, is related to the length of the virtual array as:
ρ ϕ M = λ 2 N ch Δ x cos ϕ = λ 2 L M cos ϕ
where Δ x is the distance between elements of the virtual array, L M is the total length of the virtual array, ϕ is the off boresight angle as depicted in Figure 1.
In order to avoid ambiguities, the distance between the virtual elements should be equal to λ / 4 [16]. With this spacing, if we take as an example a radar with eight virtual channels working at 77 GHz, the finest achievable resolution is roughly fifteen deg.
One way to improve angular resolution is by exploiting the well known SAR concept. While the MIMO array is transmitting pulses, the vehicle where it is mounted moves, generating the so-called synthetic aperture. While for the MIMO radar the resolution was bounded by the length of the virtual array, now the resolution is bounded just by the length of the aperture as
ρ ϕ S A R = λ 2 L A s s i n ( ϕ )
where L A s is the length of the synthetic aperture. Notice that now at the denominator there is a s i n function. While the MIMO array provides the best resolution at its boresight (in front of the car), the SAR image has its best resolution in the direction orthogonal to the motion (i.e., at the left or right of the car).
It is now useful to go deeper into the signal model and in SAR processing (also called SAR focusing). Without loss of generality, we assume that the vehicle travels in a rectilinear trajectory along x. We remark that this is not a necessary condition for SAR imaging to work, but it simplifies the following derivations. The car travels with a velocity equal to v x and the MIMO radar transmits a linearly frequency modulated chirp every Pulse Repetition Interval (PRI) second. Given the velocity, the radar system adapts its PRI in order to mantain a spatial sempling distance of [16]
P R I × v x = λ 4 × N c h
where N c h is the number of MIMO channels. A visual representation is given in Figure 2.
After the reception of the echo, the signal is range compressed. The signal model of the range compressed signal is [17]
s n ( r , τ ; p 0 ) = sin c r R n ( τ ; p 0 ) ρ r × exp j 4 π λ R n ( τ ; p 0 )
where
  • r is the slant range;
  • τ is the slow time sampled every PRI seconds;
  • n is the index of the virtual element;
  • p 0 = [ x 0 , y 0 ] T is the geometrical position of the target;
  • R n ( τ ; p 0 ) is the distance from the n th virtual element to the target at time τ .
Notice that in Equation (5), we have neglected an amplitude factor and all the bistatic effects (i.e., we are considering an ideal monostatic equivalent of the MIMO TX/RX scheme).
We can now exploit a set of samples to form a SAR image by using a simple but straightforward time domain back projection (TDBP) algorithm [16]:
I ( p ) = τ n = 1 N c h s n [ R n ( τ , p ) , τ ; p 0 ] × exp + j 4 π λ R n ( τ , p )
where I ( p ) is the SAR image at pixel p = [ x , y ] T , the expression s n [ R n ( τ , p ) , τ ; p 0 ] indicates the range compressed signal evaluated at a range R n ( τ , p ) , which is the distance between the n th virtual phase center and the generic pixel p at time τ .
We highlight that Equation (6) can be split into two separate summations. First, we form a set of low-resolution snapshots by coherently summing over the channels:
I M ( τ , p ) = n = 1 N c h s n [ R n ( τ , p ) , τ ; p 0 ] × exp + j 4 π λ R n ( τ , p )
where I M ( τ , p ) is the low-resolution image generated by the MIMO radar at time τ . We now sum over all the slow times, obtaining the final SAR image:
I ( p ) = τ I M ( τ , p )
The two-step summation opens the possibility for advanced and efficient focusing schemes [4] and autofocusing algorithms to refine the vehicle trajectory [3,18].

3. Wavenumbers in SAR Images

In this section, we will recall the concept of wavenumber which is widely used in geophysical, radar and SAR imaging. Any SAR experiment can be described in two conjugate domains: the direct domain referred to usually as time or space domain, and the frequency or wavenumber domain. This representation comes useful since any SAR experiment can be considered as the enlightenment of the spatial frequency components of a target, as extensively explained in [19].
If we focus for a moment on a simple Cartesian 3D reference system, the transformation from the spatial domain to the wavenumber domain can be performed by a simple 3D Fourier transform, that is by projecting the spatial signal on the basis function:
e j k · r = e j ( k x x + k y y + k z z ) = e j ψ
where k = [ k x , k y , k z ] T is the so-called wavevector and its components are called wavenumbers. The wavenumbers can be computed by simply taking the partial derivative of the phase ψ w.r.t. of each spatial variable. Taking advantage of the gradient notation:
k = ψ
where is the gradient operator and ψ is the phase in Equation (9). The transmitted signal can be represented in the spatial domain (in the far field approximation) as a plane wave illuminating the target. The direction of propagation of such a wave is defined by the wavevector k and it is parallel to the line joining the sensor and the target. It follows that a movement of the sensor corresponds to a change in the illuminated wavenumber, or, in other words, to an expansion of the bandwidth of the signal in the wavenumber domain.
To clarify the need of a custom reference frame in place of a Cartesian or polar one, we must highlight a few characteristics of the signal in the wavenumber domain, such as the dependence of the illuminated wavenumber by both the position of the sensor and the target in the scene. This issue is discussed more in detail in the next section.

4. Imaging with Bursted Data

This section encloses the heart of this work. First, a description of the burst acquisition mode is provided as well as the spectral coverage of the most common reference systems and the derivation of a proper reference system where the spectral space invariance property holds. Second, the procedure for burst phase calibration is provided and finally, the CLEAN algorithm is described. A visual representation of our workflow is given in Figure 3.

4.1. Burst Data Acquisitions

As described in Section 3, the reciprocal position of radar and target determines the illuminated wavenumbers and thus the SAR angular resolution. Usually, the processed synthetic aperture is uniformly sampled; therefore, the coverage in the wavenumber domain is uniform (without “holes”).
In some cases, and in particular in the automotive radar, SAR acquisitions may be performed on a non-continuous basis. The reason behind this lies on the regulations establishing the coexistence between different telecommunication devices [20] and for electronic battery power saving. In practice, this acquisition mode, called burst mode, consists in turning on and off, periodically, the radar along the traveled trajectory. Figure 4 explains the concept clearly.
The result of this non-continuous acquisition is that, in the spatial frequency domain, some wavenumbers are not covered, generating some gaps in the spectrum. In the focused image, these gaps manifest themselves as artifacts commonly known as grating lobes. We recall once again that the positions of the covered areas (from now on called tiles) and the positions of the gaps in a Cartesian reference frame are a function of both the radar position and the target position. For an artificial problem composed by a single target, the gaps will be visible in the spectrum, but this is not the case if the scene is composed by a variety of different targets distributed in any position of the field of view.
The need arises for suitable data representation that allows a common spectral description for all the targets in the image in order to allow a standard signal processing. In the following section, a mathematical framework will be developed in order to overcome this problem.

4.2. 2D Tessellation

In this subsection, we are going to analyze the spectral behavior of the most common reference systems: we start from the simplest one which is the Cartesian; then, we will proceed with a frequency description of the signal in the polar coordinates and finally on a custom (and appropriate for our needs) reference frame.

4.2.1. Cartesian Reference System

Given a 2D Cartesian reference system as in Figure 5 and a target located in p , the range equation for a given time instant τ is
R ( τ ; p ) = ( x ( τ ) x p ) 2 + ( y ( τ ) y p ) 2
where
  • x ( τ ) is the position of the radar along the x axis at time τ ;
  • y ( τ ) is the position of the radar along the y axis at time τ ;
  • ( x p , y p ) are the coordinates of the target.
The spectral contributes ( k x , k y ) can be computed as the partial derivative of the range equation w.r.t. for each variable:
k x = 4 π λ R ( τ ; p ) x = 4 π λ x ( τ ) x p R ( τ ; p ) = 4 π λ cos θ ( τ )
k y = 4 π λ R ( τ ; p ) y = 4 π λ y ( τ ) y p R ( τ ; p ) = 4 π λ sin θ ( τ )
where θ ( τ ) is the look angle from the boresight of the radar to the target. The above equations are obviously space variant, i.e., they depend on both the position of the target p and the position of the radar.

4.2.2. The Polar ( r , θ ) Reference System

Another common reference system is the polar one., depicted in Figure 6. It is possible to fix a relationship between the polar and the Cartesian system as
{ x = r cos θ y = r sin θ
where r is the range and θ the angle.
It is rather easy, thanks to the derivative chain rule, to compute the spectral contributes under this reference system ( k r , k θ ) , as
{ k r = 4 π λ ( R ( τ ; p ) x x r + R ( τ ; p ) y y r ) = 4 π λ [ cos ( θ ( τ ) θ p ) ] k θ = 4 π λ ( R ( τ ; p ) x x θ + R ( τ ; p ) y y θ ) = 4 π λ r [ sin ( θ ( τ ) θ p ) ]
where θ ( τ ) is the angle between the sensor and the target and θ p is the angular position of the target in the polar reference frame. If we place the origin of the polar grid at the center of the aperture and we assume a conveniently small aperture where θ ( τ ) θ p , we can expand Equation (15) in Taylor series around the point θ ( τ ) θ p = 0 , obtaining
k r ( τ ) = 4 π λ 1 1 2 d θ ( τ ) 2
which can be considered almost constant for short enough apertures. For the angular wavenumber, instead, we have:
k θ ( τ ) = 4 π λ [ r · d θ ( τ ) ]
which, instead, is not constant. Since the dependency on the position of the radar is not perfectly highlighted, let us use the relation (14) to explicit the differential d θ as a function of ( x , y ) :
d θ = 1 r d x · s i n ( θ ) + d y · c o s ( θ )
in order to refer Equation (17) explicitly to the radar position, it is possible to look at the differentials d x and d y as the displacement of the sensor position along the trajectory, namely
k θ ( τ ) = 4 π λ [ Δ x ( τ ) · s i n ( θ p ) + Δ y ( τ ) · c o s ( θ p ) ]
which depends on both the sensor position and the position of the target. Again, the polar reference frame is space variant and therefore not suitable for our purpose.

4.2.3. The (r,e) Reference System

In this section, we develop a reference system such that the illuminated wavenumbers are uniquely dependent on the position of the sensor and not on the position of the targets in the scene. We call this system ( r , e ) , since one variables remains the range (which is space invariant) and the other is instead a transformation of the angle θ , such that the resulting wavenumber is independent upon the target position:
k e ( τ ) = k θ ( τ ) d θ d e = 4 π λ [ Δ x ( τ ) · s i n ( θ ) + Δ y ( τ ) · c o s ( θ ) ] · d θ d e
There is no particular solution d θ / d e allowing a total independence of k e ( τ ) from the target position. However, if we assume that the target trajectory is along only one direction, let us say x, we can discard the other term by forcing Δ y ( τ ) = 0 . Therefore (20) becomes
k e ( τ ) 4 π λ ( Δ x ( τ ) · s i n ( θ ) ) · d θ d e
It comes straightforward that in order to have space invariance the ratio d θ / d e must be equal to
d θ d e = 1 s i n ( θ )
and therefore:
d e d θ = s i n ( θ )
From which we can derive the change of variable needed:
e ( θ ) = π 2 θ s i n ( θ ) · d θ = c o s ( θ )
The reference system built upon this assumption is ambiguous at bore-sight, since the function c o s ( θ ) is even. One possible solution is to process separately two sides of the scene, left and right. However, this solution can cause some problems in merging the two images. Another possible solution is to change Equation (22), taking into consideration the ambiguities. We then consider
d e d θ = | s i n ( θ ) |
which leads to
e ( θ ) = 1 cos θ if θ 0 cos θ 1 if θ 0
Thanks to Equation (26), it is possible compute the spectral coverage in the k e domain, using the derivative chain rule:
k e = 4 π λ R ( τ , p ) x x θ θ e + R ( τ , p ) y y θ θ e
Plugging Equation (25) in Equation (27) one gets
k e ( τ ) = 4 π λ r x ( τ ) R if θ 0 4 π λ r x ( τ ) R if θ 0
The spectral descriptor has sudden change of sign around θ = 0 . This can become an issue for the space invariance property that we require from our custom reference system. In particular, due to the dependency on θ in Equation (28), the space spectral descriptor will lose it’s space invariance property. For instance, let us assume to have a target at a positive angle θ and to have only positive values of the radar sensor x. The covered tile in the spatial spectral domain will be purely positive. On the other hand, if we have a target with a negative angle θ , only negative wavenumbers will be covered, which is not the spatial invariance property we would like to have. Anyway, in order to make the spectral contributes of all the targets in the scene sum up in the same spectral portion, the synthetic aperture must be centered around x ( τ ) = 0 ; in addition, an odd number of bursts must be processed. This must be done in order to exploit the odd symmetry of the spatial spectral estimator, in order to let the targets’ spatial spectral component sum up in the same spectral portion.

4.3. Constant Spectrum Approximation and Validity Limits

In order to achieve a space spectral invariance property in Equation (28), we would like to consider as a first approximation
r R
therefore, we define a validity region for this approximation, where the value of r is comparable with the total synthetic aperture length. In addition, in order to take into account the burst acquisition mode, we will introduce an indicator function I ( x ( τ ) ) , which takes the value 1 if the radar is acquiring the data and 0 otherwise.
By taking into account the above considerations, Equation (28) becomes
k ¯ e = 4 π λ I ( x ( τ ) ) x ( τ )
By using Equation (30), it is possible to compute the spatial spectral bandwidth and the central wavenumber in a straightforward manner. The spectral bandwidth can be computed by considering the displacement of the vehicle within one burst:
B = 4 π λ Δ x = 4 π λ L s y n t h
where Δ x is the displacement along the direction of motion, and L s y n t h refers to the total synthetic aperture. In addition, the previous equation can be used to state the resolution in the ( r , e ) reference system, indeed:
ρ e = 2 π B = λ 2 L s y n t h
where ρ e is the resolution along the e-variable. Similarly, it is possible to define the central wavenumber by using the midpoint of the synthetic aperture x 0 , as
k e 0 = 4 π λ x 0
The above equations can be referred to a constant spectrum approximation (CSA). Actually the CSA presents some limits, in particular outside the validity region defined thanks to Equation (29), i.e., in the near range zone, where the range becomes comparable to the synthetic aperture length. However, thanks to the CSA, it is possible to predict exactly the image spectral component according to the two laws expressed in Equation (31) and in Equation (33).

4.4. Residual Motion Compensation and Phase Calibration

This subsection describes at high level the autofocus and calibration procedure performed on each single burst after their focusing. We remark that these procedures are mandatory to remove the residual trajectory error and to be able to jointly exploit all the bursts, achieving in this way extremely fine resolutions.

4.4.1. Residual Motion Compensation

An important aspect of accurate SAR imaging is the knowledge of the platform trajectory, which should be known with an accuracy of a fraction of a wavelength [10]. A threshold can be set by simply requiring that the velocity error must be smaller than the velocity resolution of the system:
Δ v e g o = λ 2 T c
where T c is the synthetic aperture time, i.e., the time taken by the vehicle to move along the synthetic aperture. An error larger than Δ v e g o will result in a mis-localization of the target of more than a resolution cell. Small wavelengths and high integration times call for a very high accuracy in the knowledge of the trajectory, which is typically an unmet condition in standard navigation units. This condition calls for a dedicated procedure to refine the platform motion directly from radar data. The residual motion compensation procedure used during this work has been fully developed and explained in [18]. In this paper, we will briefly recall the main concepts.
The availability of a multichannel device greatly simplifies the task of residual velocity estimation, which can be carried out to within sufficient accuracy by analysis of the Doppler frequency at different off-boresight angles, see for example [3]. The most common model in literature is the constant velocity error. This model is generally valid over short synthetic apertures [3,21,22,23]. This hypothesis leads to an easy expression of the residual Doppler frequency over each target in the scene:
f d ( θ ) = 2 λ Δ v e g o , r ( θ ) = 2 λ ( Δ v e g o , x cos θ + Δ v e g o , y sin θ )
where Δ v e g o , r ( θ ) is the residual radial velocity as seen by a target at angular position θ , Δ v e g o , x is the residual velocity along x, while Δ v e g o , y is the one along y. The estimation of the parameters of interest ( Δ v e g o , x and Δ v e g o , y ) can be carried out by detecting a set of N fixed Ground Control Points (GCPs) in the scene and by retrieving the residual Doppler frequency over each of them. In this way we obtain a linear system of equations:
f d ( θ 0 ) f d ( θ 1 ) f d ( θ 2 ) f d ( θ N ) = 2 λ cos θ 0 sin θ 0 cos θ 1 sin θ 1 cos θ 2 sin θ 2 cos θ N sin θ N Δ v e g o , x Δ v e g o , y
Notice that the measurement vector containing all the residual Doppler frequencies can be obtained by a simple Fourier Transform of the slow-time over each detected GCP in the scene. This is the case thanks to the assumption of constant velocity error that in turn is responsible for a constant residual Doppler frequency. The design matrix is built upon the knowledge of the off-boresight position of each GCP. In this setting, a multichannel device is essential to provide an a priori knowledge of the angular positions of the GCPs. The system of Equation (36) can be inverted using a simple least squares (LS) approach obtaining an estimate of the residual velocities.

4.4.2. Phase Calibration

Once the residual motion compensation procedure has been completed, each burst is properly focused. A further calibration step is now necessary to compensate for the displacement between single bursts apertures. Let us assume that we have a set of N b u r s t SAR images and that they are not aligned with each other. The displacement can be modeled as a phase shift between bursts as
ϕ ( x ) k ( x ) T Δ x
where k ( x ) is the wavevector associated with the spatial position x and Δ x is the residual displacement between two bursts. In order to explain the phase calibration procedure, let us consider N b u r s t = 2 , i.e., a couple of SAR images, one as the master image and the other as the slave one. Let us also consider a set of GCP shared between the two; more precisely, we will refer to a i M as the complex amplitude of the i th GCP of the master image, and to a i S as the complex amplitude of the i th GCP of the slave one. The procedure consists in maximizing the following expression w.r.t. Δ x :
max Δ x i N G C P a i M + a i S × exp { j ( k i ( x i ) T Δ x ) }
where k i ( x i ) is the wavevector associated with the spatial position x i of the i th GCP.
In practice, this procedure consists in an exhaustive search that takes place for each dimension of the images. Once the optimum value of Δ x has been found, a phase field is applied to the slave image in order to compensate for the residual phase shift. In the end, the final image is computed as the sum of the two images.

4.5. The CLEAN Algorithm

The CLEAN algorithm was first introduced for removing grating lobe induced artifacts in astronomy [7] and also is widely adopted in radar imaging [11,12,13,14]. It consists in successively selecting bright targets and removing their side-lobe responses by subtracting the impulse response of each target individually in an iterative manner. Our use-case lies in a class of techniques called compressive sensing, which aim at reconstructing a signal, given an under-determined system to be solved; those problems are commonly used to tackle unevenly sampled, periodically recurring, gapped, arbitrarily sparse data problems, which are common in SAR data processing and in astronomy [7,8,11,12,13,14,24,25]. We chose CLEAN as a target estimation algorithm for our work due to its low computational cost. Because of this, CLEAN is suitable for a real time implementation as required by the automotive application. This algorithm works under the assumption of impulsive sources to be estimated. Generally speaking, given a generic discrete signal s ( n ) , CLEAN performs a peak estimation, i.e., it selects the maximum value of s ( n ) as the signal source and then it gives as result a new signal s ˜ ( n ) , made up by the highest peaks of s ( n ) .
More accurately, let us define a generic discrete signal s ( n ) , of N t elements and its sparse spectrum S ( ω n ) of N ω element, and assume that there exists a impulse response filter h ( n ) known a priori such that it perfectly describes the unitary impulse response of the system. The signal s ( n ) can be described as the sum of scaled impulses, namely:
s ( n ) = k = 1 N t α k δ ( n k )
where α k is a complex value describing the amplitude of the signal. The CLEAN algorithm performs a peak estimation by taking the position of the maximum of s ( n ) and its value, as
k = arg max n s ( n )
α = s ( k )
where k is the position of the maximum and α its value. A new impulsive discrete signal is now formed by considering a single peak in the position of the maximum (k) with a value defined as before:
s ^ ( n ) = α δ ( n k )
Then, the impulse response of the estimated source s ¯ ( n ) is computed thanks to a convolution as
s ¯ ( n ) = k = 0 n s ^ ( k ) h ( n k )
where h ( n ) is the filter describing the unitary system impulse response. The next step of the CLEAN algorithm involves the computation of an error sequence e ( n ) , as
e ( n ) = s ( n ) s ¯ ( n )
After that, the CLEAN procedure is iteratively applied to e ( n ) , until stopping criteria are met. In our work, we have used a threshold under which the signal is considered pure noise and a maximum number of peaks are to be found; as a rule of thumb, 20 targets were selected. The advantage of CLEAN is that it both performs the side-lobes removing procedure and a denoising procedure, since as said before, s ˜ ( n ) is indeed an impulsive signal made up by the estimated peaks of s ( n ) .

5. Results

In this section, the results of our work are shown. The data have been acquired during an acquisition campaign performed with an eight-channel radar called ScanBrick ® developed by Aresys ® and working at 77 GHz. The ScanBrick ® used a bandwidth B = 1 GHz, which provides a range resolution of about 15 cm. In Table 1, the system parameters are summed up. For a more detailed explanation of the hardware set up of the experiment, see [16].
The Radar data are accompanied by the IMU data in order to know the location of the vehicle with respect to a Cartesian external reference system. The vehicle is also equipped with a camera mounted on the car rooftop in order to provide the optical ground truth of the scene.
The radar is mounted in a forward looking configuration; therefore, the radar boresight is pointing in the direction of motion. The eight-channel radar is implemented with two transmitting antennas N T x = 2 and with four receiving ones N r x = 4 , leading to eight virtual monostatic antennas spaced by λ / 4 . The angular resolution provided by an eight-channels radar is about 14 deg. Each burst is a portion of the trajectory made of M = 256 slow time samples. In each burst, the car travels for as little as 30 cm which would lead to a maximum resolution of 0.37 deg (at the right or the left of the car).
On the other hand, the whole aperture comprehensive of both on and off times, as in Figure 4, is about 1.66 m, which leads to an extremely fine angular resolution of 6.58 × 10 2 deg: five times finer than the angular resolution provided by one burst.
The radar uses a multiplexing scheme known as time division multiplexing (TDM). It consists in delaying the signal transmission from the second Tx-antenna in order to not create interference. The delay is given by the PRI. For this reason, there is a spatial shift between the two virtual antenna clusters. Indeed, since the vehicle is moving at a given velocity and the transmission is delayed, the second antenna cluster is shifted in space. This shift can be computed since both the velocity and the PRI are known by the system. This was not an issue, since the TDBP algoritm we used for focusing the images takes into consideration the position of the antennas phase centers and compensates, exactly, the antenna shift. Thus, no artifacts due to the TDM arose.
The SAR image after the phase calibration procedure, considering the whole aperture, is depicted in Figure 7a. The image is dense of peaks as an effect of the aberrations due to the burst acquisition mode. After the grating lobes suppressing procedure using N m a x = 20 as the maximum number of targets per-row, the image shown in Figure 7b has preserved the brightest targets. Let us notice that, thanks to the CLEAN algorithm, a point target reaches one-pixel resolution.
The spectrum of the grating lobes corrupted image is depicted in Figure 7c. The spatial spectrum is consistent with the theory developed in the previous sections. The spectral estimation, as said before, clearly shows a validity zone where the gaps are observable. The size of each tile is proportional to the length of the synthetic aperture exploited by each burst, as expected from Equation (30). In the near range, instead, the CSA is no longer valid and we see that the three tiles in the spectrum merge around r = 0 .
Both the Figure 7a and Figure 7b have been projected into the Cartesian domain, respectively, in Figure 8a,b. In the former, it is difficult to distinguish the shapes of the objects due to the grating lobes presence and details not being sharp; in the latter, instead, the details are sharper and the resolution is increased. A rough comparison can be performed with the optical image in order to identify the shapes of the objects in the scene. The optical image is presented in Figure 9.
We would like to draw attention to two particular objects in the images. The gate in Figure 10c has been zoomed in the SAR images and can be found in Figure 10a,b. In the former, the grating lobes can be seen in a clear manner since the effect is very similar to a blurring effect. In Figure 10b, instead, the bright back-scattering points have a shape that is sharper.
The other notable detail is depicted in Figure 11c, whose SAR equivalents are Figure 11a,b. In the latter, the shape of the parked car is much more clear, since details are recovered and grating lobes are suppressed.
In Figure 12, we propose a comparison between different processing with one or multiple bursts. Figure 12a represents the focused car by using a single burst. The resolution is rather poor, but at least no grating lobes are visible. Figure 12b shows the incoherent sum of the three bursts. In this case, noise is abated by the incoherent average, but there is no significant resolution improvement. An even better imaging capability can be found in Figure 12c which is the coherent sum of the three SAR images. Some details start to emerge, but, on the other hand, the grating lobes artifacts are clearly visible. The last image in Figure 12d shows our processing; details are reconstructed and the shape of the car is much more defined. Sidelobes from strong reflecting surfaces are also suppressed (see the bottom left of the image).
The previous comments can be written also for Figure 13. In Figure 13d, the single poles can be spotted in the images as point scattering targets after the grating lobes removing procedure. The information retrieved by the obtained images can be merged with that coming from cameras and LiDAR for a future step ahead into the autonomous driving.

6. Conclusions

In this paper, we have presented an algorithm to remove grating lobes from vehicle-based SAR images. Automotive radars are required by the law to operate in a non-continuous manner. While this condition does not represent an issue in classical automotive radar imaging, this is not the case for automotive SAR. Indeed, SAR processing is performed by exploiting a set of pulses acquired continuously from different spatial positions. If the transmission is not continuous, however, there are some gaps in the illuminated spatial spectrum of the scene; therefore, the reconstructed image will contain some aberrations called grating lobes.
This paper proposes a complete mathematical description of the spatial spectrum of a SAR image focused using burst data. A workflow to correct grating lobes is then presented. We start by detailing why a Cartesian or polar reference system cannot be used to solve this problem. We then continue by deriving a suitable reference system, which is a transformation of the polar one allowing for a space-invariant spectral representation. In other words, the spectrum of the image will depend only upon the position of the sensor and not of the target itself. The workflow continues by exploiting a compressive sensing algorithm called CLEAN, which allows the recovery of an extremely fine-resolution image without grating lobes.
The whole process has been validated through a real data set acquired with an automotive radar operating in W-band, mounted on a vehicle in forward-looking configuration. The procedure proposed in this paper has proven to remove the aberration in the image, leading to a clear recognition of the targets and their details.

Author Contributions

Conceptualization, M.G.P., M.M. and S.T.; methodology, S.T.; software, M.G.P.; validation, M.G.P. and M.M.; formal analysis, M.G.P. and S.T.; investigation, M.G.P., S.T., M.M., A.M.-G., C.M.P. and I.R.; resources, S.T., A.M.-G. and C.M.P.; data curation, M.G.P.; writing—original draft preparation, M.G.P. and M.M.; writing—review and editing, All; visualization, M.G.P. and M.M.; supervision, S.T. and C.M.P.; project administration, S.T. and I.R.; funding acquisition, S.T. and I.R. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded within under the program of the Joint Research Lab with participants Politecnico di Milano and Huawei Technologies Italia.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Acknowledgments

We would like to thank Paolo Falcone from Aresys for the great support provided with all hardware-related matters during acquisition campaigns.

Conflicts of Interest

Dr. Ivan Russo is currently a Huawei Technologies employee. The company approved the publication of the paper with him as a co-author.

References

  1. Rizzi, M.; Manzoni, M.; Tebaldini, S.; Monti–Guarnieri, A.V.; Prati, C.M.; Tagliaferri, D.; Nicoli, M.; Russo, I.; Mazzucco, C.; Alfageme, S.T.; et al. Multi-Beam Automotive SAR Imaging in Urban Scenarios. In Proceedings of the 2022 IEEE Radar Conference (RadarConf22), New York, NY, USA, 21–25 March 2022. [Google Scholar]
  2. Tagliaferri, D.; Rizzi, M.; Nicoli, M.; Tebaldini, S.; Russo, I.; Monti-Guarnieri, A.V.; Prati, C.M.; Spagnolini, U. Navigation-aided automotive SAR for high-resolution imaging of driving environments. IEEE Access 2021, 9, 35599–35615. [Google Scholar] [CrossRef]
  3. Manzoni, M.; Rizzi, M.; Tebaldini, S.; Monti-Guarnieri, A.V.; Prati, C.M.; Tagliaferri, D.; Nicoli, M.; Russo, I.; Mazzucco, C.; Duque, S.; et al. Residual Motion Compensation in Automotive MIMO SAR Imaging. In Proceedings of the 2022 IEEE Radar Conference (RadarConf22), New York, NY, USA, 21–25 March 2022. [Google Scholar]
  4. Tebaldini, S.; Rizzi, M.; Manzoni, M.; Guarnieri, A.M.; Prati, C.; Tagliaferri, D.; Nicoli, M.; Spagnolini, U.; Russo, I.; Mazzucco, C.; et al. A Quick and Dirty processor for automotive forward SAR imaging. In Proceedings of the 2022 IEEE Radar Conference (RadarConf22), New York, NY, USA, 21–25 March 2022. [Google Scholar]
  5. Tagliaferri, D.; Rizzi, M.; Tebaldini, S.; Nicoli, M.; Russo, I.; Mazzucco, C.; Monti-Guarnieri, A.V.; Prati, C.M.; Spagnolini, U. Cooperative synthetic aperture radar in an urban connected car scenario. In Proceedings of the 2021 1st IEEE International Online Symposium on Joint Communications & Sensing (JC&S), Dresden, Germany, 23–24 February 2021. [Google Scholar]
  6. Hasch, J.; Topak, E.; Schnabel, R.; Zwick, T.; Weigel, R.; Waldschmidt, C. Millimeter-wave technology for automotive radar sensors in the 77 GHz frequency band. IEEE Trans. Microw. Theory Tech. 2012, 60, 845–860. [Google Scholar] [CrossRef]
  7. Högbom, J.A. Aperture synthesis with a non-regular distribution of interferometer baselines. Astron. Astrophys. Suppl. Ser. 1974, 15, 417. [Google Scholar]
  8. Li, J.; Stoica, P. Efficient mixed-spectrum estimation with applications to target feature extraction. IEEE Trans. Signal Process. 1996, 44, 281–295. [Google Scholar] [CrossRef]
  9. Wu, H.; Zwick, T. A novel motion compensation algorithm for automotive SAR: Simulations and experiments. In Proceedings of the German Microwave Conference Digest of Papers, Berlin, Germany, 15–17 March 2010; pp. 222–226. [Google Scholar]
  10. Tebaldini, S.; Rocca, F.; d’Alessandro, M.M.; Ferro-Famil, L. Phase calibration of airborne tomographic SAR data via phase center double localization. IEEE Trans. Geosci. Remote Sens. 2015, 54, 1775–1792. [Google Scholar] [CrossRef]
  11. Tsao, J.; Steinberg, B.D. Reduction of sidelobe and speckle artifacts in microwave imaging: The CLEAN technique. IEEE Trans. Antennas Propag. 1988, 36, 543–556. [Google Scholar] [CrossRef]
  12. Bose, R. Lean CLEAN: Deconvolution algorithm for radar imaging of contiguous targets. IEEE Trans. Aerosp. Electron. Syst. 2011, 47, 2190–2199. [Google Scholar] [CrossRef]
  13. Wang, Q.; Xing, M.; Lu, G.; Bao, Z. High-resolution three-dimensional radar imaging for rapidly spinning targets. IEEE Trans. Geosci. Remote Sens. 2007, 46, 22–30. [Google Scholar] [CrossRef]
  14. Lanterman, A.D.; Munson, D.C., Jr. Deconvolution techniques for passive radar imaging. In Algorithms for Synthetic Aperture Radar Imagery IX; SPIE: Bellingham, WA, USA, 2002; Volume 4727. [Google Scholar]
  15. Cumming, I.G.; Wong, F.H.C. Digital Processing of Synthetic Aperture Radar Data: Algorithms and Implementation; OCLC:ocm56805179; Artech House Remote Sensing Library, Artech House: Boston, MA, USA, 2005. [Google Scholar]
  16. Tebaldini, S.; Manzoni, M.; Tagliaferri, D.; Rizzi, M.; Monti-Guarnieri, A.V.; Prati, C.M.; Spagnolini, U.; Nicoli, M.; Russo, I.; Mazzucco, C. Sensing the Urban Environment by Automotive SAR Imaging: Potentials and Challenges. Remote Sens. 2022, 14, 3602. [Google Scholar] [CrossRef]
  17. Curlander, J.C.; McDonough, R.N. Synthetic Aperture Radar: Systems and Signal Processing; Wiley Series in Remote Sensing; Wiley: New York, NY, USA, 1991. [Google Scholar]
  18. Manzoni, M.; Tagliaferri, D.; Rizzi, M.; Tebaldini, S.; Monti-Guarnieri, A.V.; Prati, C.M.; Nicoli, M.; Russo, I.; Duque, S.; Mazzucco, C.; et al. Motion Estimation and Compensation in Automotive MIMO SAR. arXiv 2022, arXiv:2201.10504. [Google Scholar] [CrossRef]
  19. Wu, R.S.; Toksoz, M.N. Diffraction tomography and multisource holography applied to seismic imaging. Geophysics 1987, 52, 11–25. [Google Scholar] [CrossRef]
  20. Zheng, L.; Lops, M.; Eldar, Y.C.; Wang, X. Radar and Communication Coexistence: An Overview: A Review of Recent Methods. IEEE Signal Process. Mag. 2019, 36, 85–99. [Google Scholar] [CrossRef]
  21. Iqbal, H.; Loffler, A.; Mejdoub, M.N.; Gruson, F. Realistic SAR Implementation for Automotive Applications. In Proceedings of the 2020 17th European Radar Conference (EuRAD), Utrecht, The Netherlands, 13–15 January 2021; pp. 306–309. [Google Scholar] [CrossRef]
  22. Gao, X.; Roy, S.; Xing, G. MIMO-SAR: A Hierarchical High-Resolution Imaging Algorithm for mmWave FMCW Radar in Autonomous Driving. IEEE Trans. Veh. Technol. 2021, 70, 7322–7334. [Google Scholar] [CrossRef]
  23. Gisder, T.; Meinecke, M.M.; Biebl, E. Algorithmic Steps for SAR Backprojection on Radar Based Motion Estimation. In Proceedings of the 2020 21st International Radar Symposium (IRS), Warsaw, Poland, 5–8 October 2020; pp. 385–390. [Google Scholar] [CrossRef]
  24. Larsson, E.G.; Liu, G.; Stoica, P.; Li, J. High-resolution SAR imaging with angular diversity. IEEE Trans. Aerosp. Electron. Syst. 2001, 37, 1359–1372. [Google Scholar] [CrossRef]
  25. Larsson, E.G.; Stoica, P.; Li, J. Amplitude spectrum estimation for two-dimensional gapped data. IEEE Trans. Signal Process. 2002, 50, 1343–1354. [Google Scholar] [CrossRef]
Figure 1. (a): Real array configuration for an 8-channel automotive radar made up by two transmitting antennas (in blue) and four receiving antennas (in red) in forward-looking configuration. (b): Virtual array for an 8-channel automotive radar spaced by λ / 4 in forward-looking configuration.
Figure 1. (a): Real array configuration for an 8-channel automotive radar made up by two transmitting antennas (in blue) and four receiving antennas (in red) in forward-looking configuration. (b): Virtual array for an 8-channel automotive radar spaced by λ / 4 in forward-looking configuration.
Remotesensing 15 00845 g001
Figure 2. Uniform Linear array with spacing distance d = λ / 4 in forward SAR looking configuration, with a reference point target p 0 moving at a speed v with a Pulse Repetition Interval P R I .
Figure 2. Uniform Linear array with spacing distance d = λ / 4 in forward SAR looking configuration, with a reference point target p 0 moving at a speed v with a Pulse Repetition Interval P R I .
Remotesensing 15 00845 g002
Figure 3. Grating lobes suppression workflow block diagram.
Figure 3. Grating lobes suppression workflow block diagram.
Remotesensing 15 00845 g003
Figure 4. Burst acquisition mode representation: while the vehicle is moving the radar works for a period of time T o n = N × P R I , where N is the number of slow time samples for each burst, and it is silent for a period of time T o f f .
Figure 4. Burst acquisition mode representation: while the vehicle is moving the radar works for a period of time T o n = N × P R I , where N is the number of slow time samples for each burst, and it is silent for a period of time T o f f .
Remotesensing 15 00845 g004
Figure 5. Back projection grid representation in a Cartesian reference system for a forward looking Automotive SAR.
Figure 5. Back projection grid representation in a Cartesian reference system for a forward looking Automotive SAR.
Remotesensing 15 00845 g005
Figure 6. Back projection grid representation in a Polar reference system for a forward looking Automotive SAR.
Figure 6. Back projection grid representation in a Polar reference system for a forward looking Automotive SAR.
Remotesensing 15 00845 g006
Figure 7. (a): Scene focused in the (r,e) reference system. (b): CLEAN processed image, the image is much less dense since the grating lobes have been removed. (c): Spatial spectrum image, for each burst there is a column of data which is space invariant in the validity region of the CSA.
Figure 7. (a): Scene focused in the (r,e) reference system. (b): CLEAN processed image, the image is much less dense since the grating lobes have been removed. (c): Spatial spectrum image, for each burst there is a column of data which is space invariant in the validity region of the CSA.
Remotesensing 15 00845 g007
Figure 8. (a): Image geocoded into the Cartesian reference system. (b): Image geocoded into the Cartesian reference system without the grating lobes.
Figure 8. (a): Image geocoded into the Cartesian reference system. (b): Image geocoded into the Cartesian reference system without the grating lobes.
Remotesensing 15 00845 g008
Figure 9. Optical reference taken from a camera placed on top of the vehicle.
Figure 9. Optical reference taken from a camera placed on top of the vehicle.
Remotesensing 15 00845 g009
Figure 10. (a): SAR image of a gate detail before the grating lobes suppressing algorithm. (b): SAR image of a gate detail after the grating lobes suppressing algorithm. (c): Optical reference of the gate detail.
Figure 10. (a): SAR image of a gate detail before the grating lobes suppressing algorithm. (b): SAR image of a gate detail after the grating lobes suppressing algorithm. (c): Optical reference of the gate detail.
Remotesensing 15 00845 g010
Figure 11. (a): SAR image of a car detail before the grating lobes suppressing algorithm. (b): SAR image of a car detail after the grating lobes suppressing procedure. (c): Optical reference of the car detail.
Figure 11. (a): SAR image of a car detail before the grating lobes suppressing algorithm. (b): SAR image of a car detail after the grating lobes suppressing procedure. (c): Optical reference of the car detail.
Remotesensing 15 00845 g011
Figure 12. (a): Car detail of a SAR image generated using the synthetic aperture of a single burst length. (b): Car detail of a SAR image generated using the incoherent sum of the three bursts. (c): Car detail of a SAR image generated using the coherent sum of the three bursts. (d): Car detail of a SAR image generated using the grating lobes removing procedure.
Figure 12. (a): Car detail of a SAR image generated using the synthetic aperture of a single burst length. (b): Car detail of a SAR image generated using the incoherent sum of the three bursts. (c): Car detail of a SAR image generated using the coherent sum of the three bursts. (d): Car detail of a SAR image generated using the grating lobes removing procedure.
Remotesensing 15 00845 g012
Figure 13. (a): Gate detail of a SAR image generated with the synthetic aperture of a single burst length. (b): Gate detail of a SAR image generated using the incoherent sum of the three bursts. (c): Gate detail of a SAR image generated using the coherent sum of the three bursts. (d): Gate detail of a SAR image generated using the grating lobes removing procedure.
Figure 13. (a): Gate detail of a SAR image generated with the synthetic aperture of a single burst length. (b): Gate detail of a SAR image generated using the incoherent sum of the three bursts. (c): Gate detail of a SAR image generated using the coherent sum of the three bursts. (d): Gate detail of a SAR image generated using the grating lobes removing procedure.
Remotesensing 15 00845 g013
Table 1. Data Parameters.
Table 1. Data Parameters.
ParametersValue
Carrier frequency f c 77 GHz
Bandwidth B1 MHz
Active Tx channels2
Active Rx channels4
Geometry (mode)forward looking
Range Resolution15 cm
Synthetic aperture length1.66 m
Angular Resolution6.58 × 10−2 deg
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Polisano, M.G.; Manzoni, M.; Tebaldini, S.; Monti-Guarnieri, A.; Prati, C.M.; Russo, I. Very High Resolution Automotive SAR Imaging from Burst Data. Remote Sens. 2023, 15, 845. https://doi.org/10.3390/rs15030845

AMA Style

Polisano MG, Manzoni M, Tebaldini S, Monti-Guarnieri A, Prati CM, Russo I. Very High Resolution Automotive SAR Imaging from Burst Data. Remote Sensing. 2023; 15(3):845. https://doi.org/10.3390/rs15030845

Chicago/Turabian Style

Polisano, Mattia Giovanni, Marco Manzoni, Stefano Tebaldini, Andrea Monti-Guarnieri, Claudio Maria Prati, and Ivan Russo. 2023. "Very High Resolution Automotive SAR Imaging from Burst Data" Remote Sensing 15, no. 3: 845. https://doi.org/10.3390/rs15030845

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop