Next Article in Journal
Line-Structured Light-Based Three-Dimensional Reconstruction Measurement System with an Improved Scanning-Direction Calibration Method
Previous Article in Journal
Toward Efficient UAV-Based Small Object Detection: A Lightweight Network with Enhanced Feature Fusion
Previous Article in Special Issue
Analyzing the Adversarial Robustness and Interpretability of Deep SAR Classification Models: A Comprehensive Examination of Their Reliability
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Improved Real-Time SPGA Algorithm and Hardware Processing Architecture for Small UAVs

1
National Key Laboratory of Microwave Imaging Technology, Aerospace Information Research Institute, Chinese Academy of Sciences, Beijing 100190, China
2
School of Electronic, Electrical and Communication Engineering, University of Chinese Academy of Sciences, Beijing 100094, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2025, 17(13), 2232; https://doi.org/10.3390/rs17132232
Submission received: 12 May 2025 / Revised: 19 June 2025 / Accepted: 27 June 2025 / Published: 29 June 2025

Abstract

Real-time Synthetic Aperture Radar (SAR) imaging for small Unmanned Aerial Vehicles (UAVs) has become a significant research focus. However, limitations in Size, Weight, and Power (SwaP) restrict the imaging quality and timeliness of small UAV-borne SAR, limiting its practical application. This paper presents a non-iterative real-time Feature Sub-image Based Stripmap Phase Gradient Autofocus (FSI-SPGA) algorithm. The FSI-SPGA algorithm combines 2D Constant False Alarm Rate (CFAR) for coarse point selection and spatial decorrelation for refined point selection. This approach enables the accurate extraction of high-quality scattering points. Using these points, the algorithm constructs a feature sub-image containing comprehensive phase error information and performs a non-iterative phase error estimation based on this sub-image. To address the multifunctional, low-power, and real-time requirements of small UAV SAR, we designed a highly efficient hybrid architecture. This architecture integrates dataflow reconfigurability and dynamic partial reconfiguration and is based on an ARM + FPGA platform. It is specifically tailored to the computational characteristics of the FSI-SPGA algorithm. The proposed scheme was assessed using data from a 6 kg small SAR system equipped with centimeter-level INS/GPS. For SAR images of size 4096 × 12,288, the FSI-SPGA algorithm demonstrated a 6 times improvement in processing efficiency compared to traditional methods while maintaining the same level of precision. The high-efficiency reconfigurable ARM + FPGA architecture processed the algorithm in 6.02 s, achieving 12 times the processing speed and three times the energy efficiency of a single low-power ARM platform. These results confirm the effectiveness of the proposed solution for enabling high-quality real-time SAR imaging under stringent SwaP constraints.

1. Introduction

SAR imaging represents a significant advancement in remote sensing, offering day-and-night, all-weather imaging capabilities that surpass those of traditional optical imaging technologies. This technology has extensive applications in military and civilian domains, such as vessel tracking, disaster assessment, and terrain surveillance [1]. Unmanned aerial vehicles (UAVs) serve as a pivotal platform for SAR applications, leveraging their flexibility, portability, and adaptability for real-time imaging, target detection, and tracking efforts.
However, due to their small and light nature, UAVs are particularly susceptible to atmospheric turbulence, which can lead to deviations from the ideal flight trajectory and result in motion errors [2,3]. In SAR systems, motion errors result in range migration and azimuth phase errors (APE) upon echo reception, significantly impairing the image resolution [4,5].
Accurate measurement and compensation of motion errors are crucial for acquiring high-precision and high-quality SAR imagery. One effective strategy is the integration of a high-precision and heavy-duty INS/GPS to concurrently capture the platform’s motion data [6,7], enabling effective motion compensation (MoCo) techniques to correct these errors. Another option is to estimate the motion error based on the echo data using the autofocus method, including the residual range cell migration (RCM) and APE, and then compensate for the error, which usually requires substantial computing power [8]. However, due to the limitations of SwaP, small UAV platforms with payloads under 7 kg cannot be equipped with high-precision INS/GPS. They cannot carry large processors with high computing power. Therefore, as shown in Figure 1, from the perspective of the system, it is necessary to comprehensively consider the INS/GPS localization accuracy, processor computing power, and algorithm complexity, and propose a real-time SAR imaging scheme that enables the completion of MoCo, imaging, and autofocus processes on a small UAV platform.
Under strict SwaP constraints, small UAV platforms, including multi-rotor configurations, typically rely on integrated light INS/GPS navigation systems that employ MEMS-IMUs to determine positional data for motion compensation in SAR imaging [9,10]. As illustrated in Figure 1, compared to the case without MoCo processing, two-step MoCo based on GPS/INS data can eliminate residual RCM errors, significantly enhancing the image quality [6,7]. However, the achievable measurement precision and sampling rate of these systems may not be sufficient to meet the demands of high-resolution SAR imaging at the Ku, Ka, and W bands, resulting in residual azimuth phase errors that can adversely impact image quality [11]. A widely adopted approach for achieving final focused SAR images is to iteratively refine defocused images through autofocus processing [12,13].
In recent years, researchers have employed data-driven autofocusing algorithms for high-precision SAR imaging. These algorithms are primarily categorized into two types: parametric and non-parametric [8].
Parametric methods assume a priori parameter models for the APE function and estimate the APE function through parameter estimation. Among parametric methods, the Map Drift (MD) algorithm is notable for its efficiency in estimating quadratic APE by analyzing sub-aperture image offsets [14,15,16]. However, it cannot estimate the higher-order APE. Another category of algorithms employs polynomial APE models based on the image quality metrics. These algorithms optimize image entropy [17,18], contrast [19], and other evaluation criteria to search for polynomial coefficients, thereby estimating the higher-order APE. Although these algorithms can accurately estimate high-order APE, they require numerous iterations, resulting in a significant computational burden and longer processing times.
The non-parametric method addresses phase errors of arbitrary forms using numerical methods, with Phase Gradient Autofocus (PGA) being the most representative technique [20]. The standard PGA algorithm encompasses four key steps: center circular shifting, windowing, phase-gradient estimation, and iterative correction. Initially proposed for spotlight-mode SAR imaging, it extracts phase errors from specifically selected points’ phase histories to enhance the focusing quality of the imagery. Enhanced PGA algorithms have been developed to accelerate the algorithm convergence. The QPGA selects high-contrast feature points to reduce the number of iterations [21], and the Weighted PGA (WPGA) enhances the phase gradient accuracy through maximum likelihood estimation [22]. To extend the applicability of the PGA to stripmap SAR, the method has been advanced by incorporating sub-aperture segmentation and phase curvature estimation, termed Stripmap PGA (SPGA) or PCA [23,24,25,26]. To enhance the performance and efficiency of the SPGA algorithm, industry scholars have adopted a range of improvement strategies. These include the use of Kalman filtering to increase the accuracy of phase error estimation [27], the employment of semidefinite relaxation for phase estimation [28], the integration PGA with the MD algorithm [29], the utilization of signal reconstruction within sub-apertures [30], and the removal of linear phase errors based on prior models [31]. Despite the satisfactory outcomes of these improvements, data dependencies and computational complexity resulting from iterations persist, making it challenging for these algorithms to meet stringent complexity requirements for real-time processing on small UAVs [8].
From an alternative perspective, advances in integrated circuit technology have significantly enhanced the performance of digital components, thereby enabling real-time Synthetic Aperture Radar (SAR) imaging and autofocus processing on-board systems. Common processors utilized in these applications include Central Processing Units (CPU), Digital Signal Processors (DSP), Graphics Processing Units (GPU), Field-Programmable Gate Arrays (FPGA), and Application-Specific Integrated Circuits (ASIC). While CPUs are designed for general computing, they exhibit limited efficiency in handling high-throughput radar signals due to the substantial instruction overhead. Conversely, DSPs improve floating-point operations and data throughput; however, they are hindered by inefficiencies in instruction parsing [32]. GPUs increase computational power with more cores, but this leads to higher power consumption [33]. Typically, the power consumption hovers around 15 watts. Implementing specialized hardware on FPGAs or ASICs provides a viable solution by optimizing the hardware for specific tasks and enhancing energy efficiency [34]. Despite exhibiting lower performance and higher power consumption than an ASIC, an FPGA’s flexibility and reconfigurability provide a superior cost-performance ratio when both cost and performance are taken into account [35]. Due to their compact size, high performance, low power consumption, and reconfigurability, FPGAs are highly appealing for on-board embedded high-performance computing. However, deploying tasks on FPGAs necessitates complex hardware circuit design, making the integration of autofocus algorithms into energy-efficient architectures challenging.
Considering the above problems, this paper proposes a lightweight real-time stripmap autofocus algorithm for kilogram-class small UAVs and an efficient hardware architecture design implementation on an ARM + FPGA platform. The primary contributions of this paper are summarized as follows:
(1) After reviewing the existing literature on PGA algorithms, we summarize the processing flow of the QW-SPGA algorithm. The FSI-SPGA was proposed in response to the iterative issues associated with the QW-SPGA algorithm. By enhancing the quality of point selection, extracting phase error feature sub-images, and performing global phase error recovery based on these feature sub-images, the algorithm significantly reduces the complexity while maintaining accuracy.
(2) Based on the characteristics of the FSI-SPGA algorithm, an efficient hardware computing architecture was designed for implementation on an FPGA. By utilizing multi-level pipelining, optimizing memory access, and employing dynamic partial reconfigurability methods, the design significantly enhances the utilization of hardware resources and energy efficiency. This architecture was implemented and validated on a Multi-Processor System on Chip (MPSoC) platform.
The remainder of this paper is organized as follows: Section 2 introduces the SAR motion error model, current autofocus methods, and associated challenges. Section 3 discusses the principles and implementation of the FSI-SPGA algorithm. Section 4 highlights the enhancements made to the FSI-SPGA algorithm and presents its hardware architecture design on an FPGA. Finally, Section 5 validates the effectiveness of the proposed scheme using real-world data from an MPSoC platform.

2. Theory and Problem Analysis

2.1. Residual Motion Error Model for SAR

The geometry of the UAV SAR data acquisition is illustrated in Figure 2, where the x-axis aligns with the course direction and the y-axis aligns with the cross-course direction. The dotted line represents the ideal trajectory of the UAV, and the yellow solid line indicates the actual trajectory.
Ideally, the aircraft moves in a straight line at a constant velocity v , but the realized flight trajectory often deviates from the ideal flight trajectory. It is assumed that the repetition period of the pulse is T a and the ideal flight altitude is H . Ideally, the antenna phase center (APC) is at [ v t a , 0 , H ] at the t a slow time, but due to the flight trajectory deviation of Δ x t a ,   Δ y t a ,   Δ z t a , the actual APC position is [ v t a + Δ x ( t a ) ,   Δ y ( t a ) ,   H + Δ z ( t a ) ] . For the point target Q located at coordinates ( x , y , z ) , its instantaneous slope R ( t a ) can be expressed as [36]:
R t a = v t a Δ x t a x 2 + Δ y t a y 2 + H + Δ z t a z 2 = R 0 t a + Δ R t a ,
where R 0 ( t a ) represents the instantaneous slant distance between the point target and the ideal trajectory, and Δ R ( t a ) represents the motion error. In general, by mounting a high-precision INS/GPS on the motion platform to measure the motion error, the residual RCM and APE can be well corrected after motion compensation of the original data [11]. However, for small UAVs, due to the constraints of SwaP, only lightweight INS/GPS with low accuracy can be mounted. In this scenario, the measured position cannot be approximated to an exact value. In this paper, after range pulse compression, two-step MoCo [6,7], and residual distance migration correction based on the ω K algorithm [13], the SAR signal in the range-azimuth frequency domain can be represented as:
s r c m , n = A 0 w a n Δ T a x v s i n c B m Δ T r 2 r c 2 Δ R ~ n Δ T a c    exp j 4 π f 0 R 0 n Δ T a c exp j 4 π f 0 Δ R ~ n Δ T a c
Δ R ~ n Δ T a = Δ R n Δ T a Δ R ^ n Δ T a ,
where Δ T a is the pulse repetition interval, Δ T r is the distance sampling interval, and Δ R ^ n Δ T a is the actual measured value of the motion error. The model indicates the relationship between the phase and distance errors and determines the form in which the phase error exists.
For high-resolution SAR imaging, the analysis in Equation (2) shows that the residual RCM error is mainly caused by Δ R ~ n Δ T a . If the flight geometry is designed and an appropriately accurate GPS/INS is selected, such that the error Δ R ~ n Δ T a < δ r / 4 , where δ r is the range resolution cell is, then the residual RCM will be transformed into APE. In the strip SAR image, the distance-compressed phase historical domain data at the k-strong point can be expressed as [27]:
s k t a = A k r e c t t a η k T a exp j π K a t a 2 + j 2 π f k t a + j ϕ A k + ϕ e t a
where A k is the amplitude of the target’s scattering coefficient, η k is the slow time corresponding to the target’s azimuth position, 2 π f k t a is the linear phase corresponding to the azimuth position, the φ A k is the phase of the target’s scattering coefficient, φ e ( t a ) is the azimuth phase error, K a is the azimuth frequency modulation slope, and r e c t ( ) represents the rectangular window function.

2.2. Real-Time Autofocus for UAV SAR Problem Analysis

In response to the APE in question, the industry widely adopts the PGA method for its estimation and compensation. Due to the slow convergence speed of the PGA algorithm, it must undergo many iterations to achieve the ideal focusing effect, and there are many improved methods to improve the estimation accuracy and convergence speed of the PGA. The QPGA algorithm enhances the convergence speed by selecting higher-quality strong scattering points [21]. The WPGA algorithm [22,37], on the one hand, reduces the impact of weak scattering points on phase estimation through weighting. However, it precisely estimates phase errors even under low signal-to-noise ratios by leveraging the least squares method, thus improving the convergence speed of the algorithm. However, these algorithms are improved based on the traditional PGA algorithm and are mainly used for spotlight SAR. Spotlight SAR and stripmap SAR differ in that the synthetic aperture time for different scattering points within the stripmap SAR imaging scene varies, implying that the APE does not completely overlap. In order to perform a consistent estimation of phase errors for several specifically selected points within a sub-aperture, it is necessary to estimate the phase errors by taking the second derivative of the phase and then synthesizing the full-aperture phase error history through a process of double integration. Hence, the PGA technique, when applied to stripmap SAR, generally refers to the PCA algorithm [23,24,25,26,27]. For various application scenarios, many adaptive improvements have been made to the SPGA, leading to the derivation of numerous algorithms [28,29,30,31]. A non-iterative SPGA algorithm based on the removal of the linear phase was proposed in [31]. This is suitable for image defocus caused by multiple linear phase segments. While it ensures fast convergence during phase error estimation, it still follows an iterative process: point selection, azimuthal inverse compression, phase error estimation, azimuthal compression, and back to point selection. Currently, no publicly available research integrates high-quality point selection, weighting, and other general techniques to enhance the performance of the SPGA.
The QW-SPGA algorithm is summarized in this paper by combining the QPGA, WPGA, and SPGA techniques. Figure 3a shows the overall processing flow of QW-SPGA. In the process of processing the measured data, the author found that the focusing performance of the QW-SPGA was better than that of the traditional SPGA method. This paper establishes a quantitative estimation of the computational complexity of the QW-SPGA algorithm:
t Q W S P G A = t c p + t i n a c + t e s + t c o × N i t e r = O N i t e r M N ( 4 log 2 M + 3 + N r e f ) + K M 4 log 2 M + 5 ,  
where t c p = O ( M N N r e f ) represents the time taken to select the scattering point set, t e s = O K M 4 log 2 M + 5 represents the time taken to estimate the phase error, t c o = O M N 2 + 2 log 2 M represents the time taken to compensate for the image error, and t i n a c = O M N 1 + 2 log 2 M represents the time taken to inversely compress the image azimuth. N i t e r is the number of iterations. N r e f is the number of the reference unit in 2D CFAR. K is the number of selected scatterers. M and N are the number of azimuth points and range points, respectively. As can be seen from Equation (5), the algorithm runtime is proportional to the iteration number N i t e r .
For lightweight drones weighing around 6 kg, real-time SAR imaging with high resolution and wide-swath coverage poses significant challenges in terms of weight, power consumption, and processing efficiency. Specifically, achieving high-resolution SAR imaging with a swath width of 3.6 km and a resolution of 0.3 m while maintaining a flight speed of 12 m/s requires efficient data processing and power-management capabilities.
Given these constraints, accumulating data for two synthetic apertures of 150 m each takes 25 s. For real-time imaging, the system must simultaneously process the current data while completing the imaging, autofocus, and interpreting the previous data. The range-Doppler (RD) algorithm, accelerated by FPGA, is currently the most efficient method for SAR imaging, requiring 6 s for processing [38]. However, incorporating motion error compensation based on INS/GPS data adds an additional 3 s. Therefore, autofocus processing must be completed within a similar timeframe.
To achieve real-time autofocus, the iterative process must be minimized, with the number of iterations N i t e r approaching 1, and the error estimation time t e s minimized. This requirement is further complicated by the power constraints of the lightweight platforms. Typically, a 6 kg drone has a battery capacity of no more than 180 Wh, with the majority of the energy allocated to flight propulsion [39]. Thus, the data processing system must operate within a power limit of 15 W.
Under these constraints, FPGA-based hardware acceleration is superior to general-purpose processors, such as CPUs and GPUs [38]. However, algorithms like QW-SPGA, which involve high computational complexity due to uncertain point selection and multiple iterations, are impractical for deployment on resource-limited FPGAs. Therefore, optimizing the hardware architecture to meet these stringent requirements is essential for the successful implementation of real-time SAR imaging on lightweight drones.

3. Lightweight Autofocusing Algorithm Design

Compared with the conventional SPGA algorithm, the QW-SPGA algorithm integrates two techniques: contrast-based high-quality point selection and weighted maximum likelihood (WML) estimation. These enhancements can improve the estimation accuracy of the phase errors during each iteration, thereby accelerating the convergence of the algorithm. In actual experiments, when the number of iterations N i t e r of the algorithm is between 6 and 8, a satisfactory focusing effect can be achieved. In the context of real-time SAR imaging using small UAVs, strict time and computational power constraints necessitate a detailed examination of the key steps involved in the QW-SPGA algorithm. This scrutiny reveals several persistent issues that warrant attention:
(1) Scattering points exhibiting high contrast along the azimuth can be selected using the maximum contrast criterion. However, this approach does not ensure the isolation of scattering points along the range, which leads to the sidelobes of the same scattering point contributing to the phase estimation multiple times.
(2) The quadratic integration of the phase curvature introduces additional noise, resulting in a loss of accuracy.
(3) In theory, the set of high-quality selected points should encompass all the information required to estimate phase errors. However, the QW-SPGA method initiates point selection at each iteration, which may not be optimal.
To address these issues, this subsection improves the QW-SPGA algorithm and proposes a non-iterative and high-precision autofocus algorithm called FSI-SPGA. The overall process of FSI-SPGA is shown in Figure 3b, which mainly includes three main steps. The specific principles and processing flow of the FSI-SPGA algorithm are presented in the following subsections.

3.1. Selection of High-Quality Scattering Points

To capture the entire azimuth phase error information, the selected scatterers must be high-quality, isolated, and complete. High quality means that the scatterers must have a high signal-to-noise ratio to yield reliable phase error data. Isolated implies that they must be sufficiently distant from other scatterers to prevent error coupling. Complete requires that the synthetic aperture periods of the scatterers fully cover the azimuth time. Considering these factors, the FSI-SPGA algorithm comprises two stages in the point selection process: coarse and fine selections, which are detailed as follows.
(1) Coarsely select scattering points
In traditional spotlight-mode SAR imaging, the PGA algorithm identifies scatterers by selecting peaks in the data of the corresponding range gates [21]. However, in the stripmap mode, the scatterers are not all in the same synthetic aperture period. To extract isolated strong scatterers across the entire image, the classic 2D CFAR algorithm from radar target detection can be used for the preliminary scatterer selection. CFAR algorithms [40] are categorized into four types based on noise estimation: Cell Averaging (CA), Ordered Statistics (OS), Greater of (GO), and Smaller of (SO). The CA-CFAR is suitable for uniform-clutter scenarios, the OS-CFAR is ideal for multi-target situations, and the GO-CFAR and SO-CFAR are designed for clutter-edge cases. However, in this study, as we only apply 2D CFAR for preliminary scatterer selection in defocused SAR images, clutter-edge scenarios are not our focus. In terms of computational complexity, for the scene depicted in Figure 4, 2D OS-CFAR has a time complexity of O ( N a N r N r e f log 2 ( N r e f ) ) , whereas 2D CA-CFAR only has a time complexity of O ( N a N r N r e f ) , where N a is the number of pixels in the azimuth direction, N r is the number of pixels in the range direction, and N r e f is the number of reference cells. Moreover, practical tests indicate that when extracting phase error profiles for image focusing, the scatterer set selected by the 2D CA-CFAR is more robust than that selected by the 2D OS-CFAR. CA-CFAR is less effective in multi-target environments, but is advantageous when selecting scattering points with strong isolation and a high signal-to-noise ratio. The computational structure of the 2D CA-CFAR is optimized in Section 4.3.2 to enhance memory access and computational efficiency using FPGAs. Compared with CA-CFAR and OS-CFAR, tests on real-world data (as shown in Figure 5) reveal that, despite some differences in the selected target points between the two methods, most scatterers overlap and could cover the entire synthetic aperture time in the azimuth direction. Therefore, in this paper, a 2D CA-CFAR approach is employed for the preliminary selection of points within the image domain.
The processing procedure for the 2D CA-CFAR is illustrated in Figure 4a. The detection core traverses the power spectrum image and consists of a central, protection, and reference unit. The protection unit is specifically designed to estimate the noise energy, thereby mitigating the risk of target energy leakage. The power spectral image is formed from the raw SAR image data through a square-law geophone, i.e., P ( m , n ) = I m , n 2 , I is the SAR complex image matrix. When performing point-target detection, the noise energy around the central unit is first estimated as follows:
A m , n = 1 N r e f n r e f 1 P n r e f ,
where ( m , n ) represents the coordinates of the central unit, N r e f = C a C r G a G r is the number of reference units, and P ( n r e f ) is any one of the reference units. The detection threshold can be expressed as T ( m , n ) = α A ( m , n ) , where the threshold factor α = N r e f P f a 1 / N r e f 1 , and the value is 10 4 ~ 10 3 according to P f a the empirical false alarm rate. Finally, by comparing the energy of the central unit with the noise energy, it is judged whether the central unit is a scattering point.
(2) Finely select scattering points
After the coarse selection of points is completed, a collection of scattering points randomly distributed over the image can be formed as follows:
Ψ = m k , n k , P m k , n k , k = 1,2 , , K .
To ensure the accuracy of the autofocus, the point set requires further filtering to ensure that the distance between any two points is sufficiently large, and each point is the optimal choice within its local region. For any point m c , n c , P m c , n c Ψ , as shown in Figure 4b, we define its adjacent space:
m m c + n n c d ,
Scattering points in the adjacent space are divided into two categories: false points on the side lobes and real points competing with the candidate point. To eliminate the false point target and extract the high-energy target, the scattering point with the highest energy within the region is selected through an energy comparison. After filtering all points in Ψ , the final high-quality scattering point set is obtained. Figure 5a,b illustrate the outcomes of coarse and refined point selection, respectively. It is evident that the refined point selection process successfully eliminates a significant number of redundant scattering points, resulting in a uniform and high-quality set of scattering points.

3.2. Feature Sub-Image Construction

When estimating the phase error in the traditional PGA, each iteration involves multiple transformations between the image and phase history domains. Notably, estimating the phase error based on high-quality scatterers does not require the entire SAR complex image. In a defocused image, azimuth phase errors cause the main peak energy of the scatterer’s spread function to disperse. However, the extent of energy dissipation is typically within a certain range.
As shown in Figure 6c, for a standard linear frequency-modulated (LFM) signal, after pulse compression, most of the sidelobe energy is distributed within a range of 10 times the mainlobe width on either side of the mainlobe [41]. The energy of the lowest integration sidelobe is used as the threshold. Using the energy level of the lowest integrated sidelobe as a threshold, signals exceeding this threshold are extracted as the scatterer feature signal. Based on Equation (2), we introduced the quadratic and higher-order phase errors depicted in Figure 6a,b, and assessed the range of the sidelobe energy spread caused by phase errors up to 35 radians. The simulation results shown in Figure 6c indicate that the quadratic phase errors cause the most severe mainlobe widening. Consequently, the escaped sidelobe energy region is the largest, approximately six times the width of the integration sidelobe. However, pure quadratic phase errors are rare in practice. For higher-order phase errors, which are typical in practical scenarios, the sidelobe energy region is roughly four times the width of the integrated sidelobe. To address this, this paper adopts an intermediate value, using five times the width of the integrated sidelobe as the ambiguity region for independent scatterers. The ambiguity length L can be approximated using the formula:
L = 20 ρ a x a   α β ,
where ρ a represents the theoretical azimuth resolution, x a denotes the actual distance between two points in the azimuth direction of the SAR image, α is between 2 and 2.5 [41], and β is set to 5 in this study’s scenario.
Subsequently, the fuzzy areas of all scattering points are extracted. As shown in Figure 7, the blurred areas of each scattering point are stitched together according to their azimuth coordinates to form a K × L feature sub-image, which contains the phase error information of the entire SAR image.

3.3. Phase Error Estimation

After completing the above steps, a high-quality scattering point set and the corresponding feature sub-images of size K × L are obtained. These contain sufficient information to estimate the APE when the synthetic aperture time of each scattering point coincides and encompasses the entire azimuth sampling time. This process significantly reduces the data dimension when filtering out high signal-to-noise ratio data. For APE estimation using the feature sub-images, this paper updates the scatterer positions and removes linear phase errors. These enhancements improve the accuracy of phase error estimation and reduce the image azimuth offset. Figure 8 illustrates the APE estimation process.
Firstly, the azimuth zeroing process is carried out on the feature sub-image:
I 0 = O K × M     I s u b     O K × M ,
where the number of square azimuth zeros is M = ( 0.8 L p L ) / 2 and L p is the number of synthetic aperture points. The azimuth inverse compression is carried out on the I 0 of the zero-filled feature sub-image to obtain the distance-compressed azimuth time domain data of the feature sub-image:
S f e a = I F F T a F F T a I 0 exp j ϕ .
The phase of the azimuth matching filter corresponding to each scattering point f m ,   R k can be expressed as:
ϕ f m , R k = π λ R k f m 2 2 v 2 .
The energy that may be mixed with the escape from other scattering points in the time domain data S f e a , derived from the compressed azimuth time domain, can be windowed in the equivalent image signal domain to obtain the chirp signal with an error at a single scattering point:
S f e a = I F F T a r e c t w m F F T a S f e a exp j ϕ f m , R k ,
where
r e c t w m = 1 , L w 2 1 m < L w 2 0 ,                                       o t h e r w i s e
is a rectangular window function with a window length of L w , and the L w decreases proportionally in feature sub-image iterations.
After the above processing is completed, the distance compression phase history domain signal is:
S f e a k , m = A k exp j 2 π f k Δ T a m + ϕ e k , m .
In Equation (15), the phase function consists of a linear term due to the offset of the scattering point position and an APE term due to motion error. First, it is estimated as a whole, and in order to avoid loss of accuracy, a phase gradient is used here to estimate the phase error:
ϕ e ˙ ^ k , m = A n g l e S f e a k , m C o n j S f e a k 1 , m ,
where A n g l e ( ) means to find the phase, and   C o n j ( ) means to take the conjugation. φ e ˙ ( k , )   represents the phase gradient of the sub-aperture where the k-th scattering point is located. It is necessary to splice the phase gradient of each sub-aperture to form a complete azimuth synthesis aperture phase gradient. The specific splicing process is shown in Figure 9, where the phase gradient constant interval of each sub-aperture is first estimated:
Δ ϕ e ˙ ^ k 1 , k = 1 N p i = 1 N p ϕ e ˙ ^ k , i ϕ e ˙ ^ k 1 , L p N p + i ,
where N p = m k 1 m k + L p is the number of overlapping points of two adjacent sub-apertures. Then, based on the gradient interval, each sub-aperture is spliced to form a phase gradient of the synthetic aperture:
ϕ e ˙ ^ k , m = ϕ e ˙ ^ k , m Δ ϕ e ˙ ^ k 1 , k ,
where k = 2,3 , , K , and m = 1 , , L p . It is equivalent to completing the removal of the relative phase linearity term between the scattering points and unifying the linear phase error. Finally, the phase gradient of the overlapping parts of each sub-aperture is averaged to obtain the full aperture phase gradient estimation φ e ˙ ^   along the entire azimuth direction, and the gradient constant term is removed, that is, the uniform linear phase error is removed:
ϕ e ˙ ^ m = ϕ e ˙ ^ m 1 M a i = 1 M a ϕ e ˙ ^ i .
The actual estimate is not smooth, and glitches will occur due to the low signal-to-noise ratio, so smooth filtering is required to filter out burrs. Finally, the phase error of the entire azimuth direction is obtained by phase integration:
ϕ e ^ m = i = 1 m ϕ e ˙ i .
Finally, the sub-image needs to be compensated to obtain the feature sub-image after this iteration:
I 0 = I F F T a F F T a S f e a ϕ e ^ m exp j ϕ f m , R k .
Since both the linear phase term and the APE are corrected to a certain extent, the focus of each scattering point will increase and the position will change, so the scattering point set needs to be updated. As a rule of thumb, the scattering point position changes within a few tens of pixels, so the method of updating the scattering point set in this paper is as follows:
Δ m k = a r g m a x I 0 k , m r e c t w m L p 2 ,
m k = m k + Δ m k .
After updating the azimuth positions of all scattering points sequentially, the feature sub-image undergoes cyclic shifting to align the brightest spot with the azimuth center, updating the feature sub-image. In general, after 6–8 iterations, the phase error estimation results converge.
Figure 10 illustrates the extracted feature sub-images, phase error gradients, and aggregation effects from measured data, preliminarily validating the effectiveness of the method proposed in this paper. Detailed experiments and analyses are presented in Chapter 5.

4. Architecture Design and Implementation

In Section 3, we propose the FSI-SPGA algorithm, which ensures the precision and robustness of the QW-SPGA algorithm while reducing the computational complexity to the order of magnitude of a single QW-SPGA iteration. This makes the FSI-SPGA algorithm suitable for deployment on an FPGA. However, in real-time imaging applications for small UAV SAR, autofocusing is only one aspect of the processing workflow. The onboard system must complete multiple tasks, including MoCo, SAR imaging, autofocusing, and target recognition. Under strict SWaP constraints, this paper presents a real-time imaging and autofocusing architecture for small UAV SAR based on an MPSoC. This subsection provides an overview of the hardware architecture of the digital processing system and details the deployment strategy for the FSI-SPGA algorithm.

4.1. System Hardware Architecture

As shown in Figure 11, the proposed architecture employs dynamic partial reconfiguration (DPR) to partition the FPGA into static and dynamic regions. The static region, which remains unchanged after power-up, includes data interfaces and control logic for the DAC/ADC and peripherals, SAR data transmission and preprocessing, ARM-FPGA data interaction, and reconfigurable drivers. The static region transmits bitstream files to the dynamic region for logic reconstruction and interacts with the dynamic region modules via the AXI high-speed bus. The dynamic region can be reconfigured within milliseconds, allowing hardware resources to be repurposed for different processing tasks. Time-multiplexing of resources enhances utilization and accelerates multiple non-simultaneous functions, thereby improving overall processing efficiency.
The architecture features two DDR memory sets: DDR-PS on the ARM side and DDR-PL on the FPGA side. DDR-PL is memory-mapped as a FIFO for data buffering, which is controlled by the Memory Interface Generator (MIG) within the FPGA. It caches the transmitted waveforms and preprocessed data. DDR-PS is divided into host memory for ARM data caching and device-shared memory for dynamic FPGA data caching and ARM-FPGA data exchange. The ARM can transfer data between the host memory and device shared memory, while the FPGA accesses the device shared memory via the AXI-HP bus for high-speed data transfer.
Several key issues must be addressed for the efficient implementation of the FSI-SPGA algorithm on this hardware architecture: (1) decomposition of the algorithm; (2) mapping of the decomposed processing steps onto the ARM/FPGA framework; and (3) design of hardware acceleration cores within the FPGA’s dynamic region. The subsequent subsections provide a detailed exposition of these issues.

4.2. Algorithm Decomposition and Operator Mapping

Section 3 details the FSI-SPGA algorithm to explain the principles of the algorithm. To concretely describe the algorithm’s computational processes and characteristics and to guide the design of its deployment on ARM/FPGA, this paper presents a computation graph primarily composed of matrix/vector compute blocks, as shown in Figure 11. The graph is created based on the following rules:
Each compute block represents a single operation on the data matrix or vector.
A common matrix/vector-level computational IP is treated as a compute block.
The data dependencies between the blocks can be explicitly indicated.
To elucidate the characteristics of the computational data flow and the attributes of the intermediate data, the graph includes data blocks, parameter blocks, and scatterer sets alongside the computational process blocks. The entire computational flowchart of the FSI-SPGA is divided into two parts: APE estimation and image APE correction.
The computational graph illustrates the entire process and data flow of the FSI-SPGA algorithm and serves as an input for mapping the algorithm onto the ARM + FPGA heterogeneous architecture. When mapping the various computational blocks from the graph to the processors, it is essential to first identify their computational characteristics. Computational processes can be categorized into three types based on their characteristics: compute-intensive, data-intensive, and control-intensive. Compute-intensive processes have performance bottlenecks due to computation; data-intensive processes have performance bottlenecks due to IO access; and control-intensive processes have performance bottlenecks due to complex control. For compute-intensive blocks, the performance can be enhanced by designing specific hardware acceleration cores on the FPGA. For data-intensive blocks, IO operations can be reduced by implementing multi-level hardware pipelines on the FPGA or utilizing ARM’s cache for memory access optimization. Control-intensive processes are challenging to accelerate in parallel and provide limited acceleration advantages. However, they are essential to the overall computational framework and are therefore more appropriately executed on an ARM. Based on these principles, the final mapping results of the computational blocks in the FSI-SPGA algorithm for the processors are shown in Figure 12.

4.3. Hardware Computing Unit Design Based on FPGA

4.3.1. FPGA Hardware Accelerator Model

In accordance with the computational graph and operator mapping scheme, the ARM end handles the parameter computation, data scheduling, and data flow configuration. Meanwhile, the FPGA harnesses its abundant and flexible hardware resources to function as an accelerator that is specifically designed for high-throughput data processing. Based on the overall hardware architecture of the system, as shown in Figure 13a, this paper employs a dataflow-driven hardware acceleration model. The FPGA retrieves data from the device memory through Direct Memory Access (DMA) and converts this data into a data stream. This stream is processed via dedicated multi-stage pipelined accelerator cores, achieving high-throughput parallel processing. The processed data stream is then written back to the device memory. For the aforementioned acceleration model, the processing latency of a single accelerator is determined by the ratio of the data stream to the data-stream bandwidth. This latency is influenced by both the memory access bandwidth and the processing bandwidth of the accelerator core. Consequently, the processing latency τ s can be assessed using the following formula:
τ s = N m min B m η ,     f k e r b k e r D ,
where N m represents the amount of data to be processed, B m denotes the theoretical peak bandwidth of memory, η is the memory access efficiency, f k e r is the operating frequency of the accelerator core, and b k e r is the bus width of the accelerator core. The pipeline interval D refers to the interval between stages in the accelerator’s pipeline. As shown in Figure 13b, accelerators with the same operating frequency can have different hardware structures, resulting in varying pipeline intervals. The pipeline interval is inversely proportional to the processing bandwidth of the accelerator core.
For the entire task, as shown in Figure 14, there are additional considerations beyond the processing time of multiple accelerator cores. These include the data transfer time between the host memory and device memory and the switching time between different accelerators. The data transfer time t m in memory is primarily dependent on the volume of data and the effective read/write bandwidth of the memory. The switching delay t r e c between accelerator cores requires a discussion of two switching modes. Mode 1 involves updating the Bitstream in the dynamic region of the FPGA, a process that requires additional time. Mode 2 refers to reconfiguring the data flow relationships between kernels within the dynamic region, which requires the presence of identical kernels across different accelerators. Therefore, when accelerating the entire processing task with this accelerator model, the delay T can be calculated as follows:
T = 2 τ m + i = 1 K τ s i + i = 1 K τ r e c i p i ,
where τ s i represents the data stream delay for the ii-th subtask, τ r e c i denotes the delay for updating the bitstream for the i-th subtask, and p i indicates the switching mode for the i-th subtask, where p i = 1 signifies switching mode 1, and p i = 0 signifies switching mode 2.
To reduce the overall task delay, this paper designs an architecture with high memory access efficiency, small pipeline intervals, and high operating frequency from the perspective of a single accelerator to enhance the effective processing bandwidth. From the perspective of multiple accelerators, a reconfigurable accelerator is designed to minimize the additional time consumed by bitstream updates. By combining the computational graph of the FSI-SPGA algorithm and the design principles of hardware accelerators, this paper divides the functionality of the FPGA hardware accelerator into two parts: one part involves 2D-CFAR detection, which pertains to two-dimensional memory access operations, and the other part includes a series of computational processes such as azimuth de-skewing, azimuth inverse compression, azimuth compression, and phase compensation.

4.3.2. D-CFAR Hardware Accelerator

2D-CFAR detection is commonly used for target detection within range-doppler matrices. However, in these applications, the matrices to be detected are typically small and can be efficiently stored in a high-speed cache for processing on a frame-by-frame basis. In the FSI-SPGA algorithm, 2D-CFAR detection is primarily employed for the coarse selection of scattering points, which often requires the processing of large matrices stored in the DDR. Traditional 2D-CFAR detection necessitates two-dimensional matrix summation operations, which involve a significant amount of non-contiguous memory access. The efficiency of non-contiguous memory access in DDR is typically an order of magnitude lower than that of contiguous memory access, thereby limiting its data processing performance. To address this issue, this paper decouples the two-dimensional matrix summation into two one-dimensional summations:
S i , j = k = i N k = i + N l = j N l = j + N I k , l L k , j = l = j N l = j + N I k , l ,   S i , j = k = i N k = i + N L k , j ,
where I represents the matrix to be processed. Based on this decoupling result, the two one-dimensional sliding summations are converted into two one-dimensional add-subtract sliding windows:
L k , j = L k , j 1 I k , j N 1 + I k , j + N ,
S i , j = S i 1 , j L i N 1 , j + L i + N , j .
As depicted in Figure 15, after the ARM side completes the computation of the initial values, the FPGA can process the horizontal and vertical one-dimensional add-subtract sliding windows in parallel through a Ping-Pong interleaved structure. During the execution of the Ping operation, data from the DDR is continuously input into Buffer 1 by rows and subjected to horizontal add-subtract sliding window summation, while Buffer 2 undergoes vertical add-subtract sliding window summation. By utilizing multiple add-subtract sliding window units for parallel processing, the memory access efficiency can be approximated as B m η f k e r b k e r D . Conversely, during the Pong operation, the roles of Buffer 1 and Buffer 2 are exchanged. In this ping-pong structure, to ensure full pipeline operation of the ping-pong mechanism, it is necessary to guarantee that the delays of the Ping and Pong operations are identical, that is, T p i n g = T p o n g .
The add-subtract sliding summation pipeline unit in Figure 15a uses fixed-point numbers instead of floating-point numbers because the latter does not meet the timing requirements. To effectively use the wide dynamic range of SAR images, data mean normalization is performed first, followed by quantization with 24-bit fixed-point numbers. This provides a dynamic range of about 144 dB, which is sufficient for practical test accuracy. By using fixed-point arithmetic, the clock of the accelerator can reach 300 MHz. Utilizing this accelerator enables the rapid computation of the computation unit matrix and the protection unit matrix, thereby enhancing the data processing efficiency of the entire 2D-CFAR detection process.

4.3.3. Reconfigurable Matched Filtering (RMF) Hardware Accelerator

In the FSI-SPGA algorithm, there exists another category of computations that can be reconstructed from various combinations of vector-level computation blocks, such as FFT/IFFT, Vector Multiple (VcMul), and Phase Gradient Calculation (PGC). These computations are limited to operations in the azimuth direction, meaning that most data accesses are contiguous and involve repetitive computational blocks. For ARM, redundant instructions and limited computational resources restrict the performance of these computations. This paper leverages the high throughput and high parallelism of FPGAs by assembling data processing procedures into data streams using pre-built computational kernels. This reduces the number of times the data are read from the DDR, processed, and written back to the DDR, thereby enhancing the processing performance. For multiple processing procedures corresponding to multiple data streams, to eliminate additional reconfiguration delays, as shown in Figure 16, data output selectors (DS) and data input multiplexers (MUX) are added. By controlling the data flow structure through multiple data stream nodes, a high resource reuse rate is achieved, enabling the high-performance implementation of multiple computational functions. This paper refers to this accelerator as a reconfigurable matched-filtering hardware accelerator, which can be configured into four modes, each corresponding to one of the four computational processes.
In the RMF accelerator, the phase calculation in Equation (12) does not involve extremely large numbers; therefore, the fractional part of the floating-point numbers provides sufficient precision. The dynamic range of floating-point numbers, from 10 45 to 10 38 , is more than adequate for the amplitude dynamic range of SAR images. Therefore, all the data in this accelerator is in floating-point format. Although using lower-bit fixed-point numbers could enhance precision and reduce resource usage, the significant engineering effort required means that this was not pursued in this paper.

5. Experiments and Results

To validate the effectiveness of the proposed strip SAR real-time autofocus software and hardware solution in this paper, experimental data were collected based on a Ku-band UAV-mounted SAR system, and autofocus processing was performed. Additionally, the proposed embedded hardware processing framework was deployed and tested on the ZCU102 platform. This section presents the SAR system, the results of the algorithm processing experiments, and the performance evaluation of the FPGA hardware accelerator from three aspects.

5.1. SAR System

To verify the proposed algorithm, an SAR imaging experiment was conducted on a multirotor UAV based on the Ku-band SAR system developed by the Aerospace Information Innovation Institute of the Chinese Academy of Sciences, as shown in Figure 17. The SAR system includes a common antenna for the transceiver and receiver, RF front-end, digital processing machine, lightweight INS/GPS, and other key components, with a total weight of 6 kg. It uses a chirp pulse system for frontal and side-view strip SAR imaging, achieving a range resolution of 0.3 m and an azimuth resolution of 0.3 m. The specific parameters are listed in Table 1. For such small multi-rotor UAVs, high-frequency vibrations and airflow interference during flight can introduce high-frequency motion errors, causing image defocus, as shown in Figure 18a. The following subsections highlight the advantages of the proposed algorithm from the perspectives of algorithm accuracy and algorithm efficiency.

5.2. Comparison of Algorithm Accuracy

After obtaining the raw SAR echo data and the corresponding GPS/INS data, initial MoCo is applied to the raw data, and then the defocused SAR image is obtained by ωK, as shown in Figure 18a. The number of azimuth points in the image is 4096, the number of distance points is 12,288, the number of azimuth synthetic aperture points is 512, and the theoretical resolution is 0.25 m. However, motion errors cause significant resolution loss and blurring of the image. The image entropy of the entire image is 5.48. Subsequent processing using the SPGA, QW-SPGA, and FSI-SPGA methods reduces the image entropy to 5.13, 4.99, and 4.91, respectively, as shown in Figure 18b–d. In terms of the overall image focusing quality, FSI-SPGA and QW-SPGA perform similarly, and both surpassed the traditional SPGA algorithm.
As shown in Figure 19 and Figure 20, regions 1 and 2 in Figure 18 are enlarged. Figure 19a shows the image after preliminary MoCo processing. Figure 19b shows the results after applying the SPGA algorithm to SAR images that undergo MoCo processing for eight iterations. Although there is some improvement in defocusing, roads and buildings still exhibit defocusing. Figure 19c,d show the results after processing with the QW-SPGA algorithm (eight iterations) and the FSI-SPGA algorithm, respectively. Both achieve high-quality focusing on scenes such as houses, roads, and fields, significantly improving the image texture clarity.
To evaluate the image quality more intuitively and quantitatively, some targets from area 2 were selected for azimuth slice analysis, and the four targets to be analyzed were circled in red in Figure 20d. The azimuth slices corresponding to the three-point targets are shown in Figure 20e–h, and the calculation results of the resolution, peak sidelobe ratio, and integral sidelobe ratio are listed in Table 2. As shown in Figure 20 and Table 2, when comparing the target slice index, image entropy, and contrast—indicators for assessing the overall image quality—the performance of the FSI-SPGA algorithm is close to that of the QW-SPGA algorithm, and both outperform the traditional SPGA algorithm. Notably, the processing time required for FSI-SPGA is equivalent to only one iteration of the SPGA and QW-SPGA algorithms.

5.3. Algorithm Robustness Verification

To further test the robustness of the algorithm, multiple datasets are used for validation. Section 5.2 shows that the accuracy of the FSI-SPGA algorithm is close to that of the QW-SPGA algorithm, and both are better than that of the traditional SPGA algorithm. Therefore, this subsection only compares the image entropy, contrast, and processing time of the FSI-SPGA and QW-SPGA algorithms. The experimental results are presented in Figure 21 and Table 3. In the three images D1, D2, and D3, which have high-quality scatterers, the image-focusing quality of the QW-SPGA and FSI-SPGA algorithms is significantly improved. For the grassy area of image D1 and the lake-island area of image D4, where there are not enough high-quality scatterers, the two algorithms do not significantly improve the image focusing quality. As shown in Table 3, the FSI-SPGA algorithm is close to the QW-SPGA algorithm in terms of image entropy and contrast, but it is nearly four times faster. These experiments show that the proposed FSI-SPGA algorithm has good processing efficiency and robustness in handling defocused SAR images with high-quality scatterers.

5.4. FPGA Hardware Accelerator Verification Experiments

To further verify the proposed hardware architecture and FPGA accelerator model, as shown in Figure 22a, the proposed FPGA hardware accelerator is implemented based on the computing core circuit of the ZCU102 evaluation board. The layout and routing results of the two key accelerators in Vivado are shown in Figure 22b,c, respectively. Their corresponding hardware resource utilization is presented in Table 4.
The acceleration of the FSI-SPGA algorithm was implemented on the ZCU102 embedded platform. This paper used three architectures: ARM, CPU, and ARM + FPGA, to process data collected by the SAR system. As shown in Figure 23b–d, the processing results of these architectures are essentially the same. However, due to differences in the calculation accuracy of ARM/CPU instructions and FPGA-based IP calculations, minor discrepancies exist. Despite these, the image focusing quality is still significantly improved.
The performance testing results of the algorithm on hardware platforms are presented in Table 5, which primarily compares the processing time, power consumption, and Performance Power Ratio (PPR) of the algorithm across different architectures. In this paper, PPR is defined as the ratio of imaging points to the energy consumed. Compared with the test results on the ARM platform, the FSI-SPGA algorithm reduces the processing delay by nearly six times and increases the PPR by nearly nine times. When comparing the performance test results between the ARM + FPGA and ARM platforms, the use of the FPGA hardware acceleration model and accelerator proposed in this paper leads to a nearly nine-fold reduction in latency and a nearly two-fold increase in PPR at 4 K × 12 K imaging points, and a nearly twenty-five-fold reduction in latency and a nearly six-fold increase in PPR at 8 K × 12 K imaging points. Compared with CPUs at advanced process nodes and high performance, the proposed solution in this paper exhibits a relatively modest improvement in processing performance but achieves a significant reduction in power consumption, resulting in an order-of-magnitude increase in PPR.
These results indicate that the FSI-SPGA algorithm demonstrates significant performance improvements on both the ARM and ARM + FPGA platforms, particularly in terms of the processing delay and PPR. The use of an FPGA hardware acceleration model and accelerator can significantly enhance the algorithm’s processing speed and energy efficiency, which is crucial for applications requiring the rapid processing of large amounts of data, such as real-time image processing and machine vision. This performance enhancement is likely attributed to the parallel processing capabilities of FPGAs and the optimized design of hardware accelerators, which can utilize hardware resources more effectively to reduce the time and energy consumption required for data processing.

6. Discussion

6.1. Analysis of Algorithm Computational Complexity and Limitations

According to Figure 3b, the theoretical calculation time of the proposed FSI-SPGA algorithm can be estimated as
t F S I S P G A = t c p + t i n a c + t s u b e s N i t e r + t c o , = O M N ( 4 log 2 M + 3 + N r e f ) + N i t e r K L p 4 log 2 L p + 5 ,
where t s u b e s = O K L p 4 log 2 L p + 5 represents the time taken to estimate the error of the feature subgraph, L p represents the synthetic aperture points, and the meaning of the other variables is the same as that of Equation (5). Comparing Equations (5) and (29), it can be found that the algorithm proposed in this paper has no image iteration, but is replaced by feature sub-image iteration with t s u b e s t c o , which greatly reduces the complexity of the algorithm, and with the increase of image size, the computational efficiency is improved more obviously, and the computational complexity is on the same order of magnitude as the azimuth compression process of SAR imaging, which can meet the real-time processing requirements. In this paper, the processing time of the two algorithms on multiple SAR images of different sizes is tested, and the results are shown in Table 5, which verifies that the algorithms have high real-time performance and strong robustness.
The FSI-SPGA algorithm is a PGA-type algorithm. It works well in scenes with isolated strong scatterers of high quality. However, it is not suitable for areas like grasslands or sea rocks, where strong scatterers are absent. Moreover, this algorithm can only correct the phase errors. As pointed out in Section 2.2, when the GPS/INS system precision is low, causing range migration errors, other algorithms for range migration compensation [36] should be applied first. In our experiment, the equivalent backscattering coefficient of the system was maintained above 20 dB. The proposed algorithm exhibited robust performance on the test data of this system. However, it may fail to accurately estimate phase errors when the signal-to-noise ratio is too low.

6.2. Analysis of Computational Performance on Hardware Architectures

Compared with low-power general-purpose processors like ARM, FPGAs come with flexible, schedulable hardware resources that can support high levels of parallel processing, and they almost entirely eliminate redundant instruction overhead. Therefore, enhancing performance through the use of Single Instruction Multiple Data (SIMD) architectures is relatively easier for FPGAs than for ARM. On the one hand, for FPGAs, non-sequential memory access can increase the data access bandwidth, which limits the performance of the computational capabilities. On the other hand, complex computational components can also lead to resource redundancy and increased power consumption. To address these two issues, this paper first constructs a computational graph of the algorithm, enabling an assessment of the computational characteristics of each process within the algorithm. This allows for an accurate description of task computational features and the extraction of hotspot computations within the FSI-SPGA algorithm that require FPGA acceleration, thus avoiding the occupation of FPGA resources by computations with low usage frequencies. Furthermore, based on a pipelined FPGA acceleration model in which data flow from DDR through multiple acceleration cores and back to DDR, two accelerators were designed: a 2D-CFAR and a reconfigurable matching filter. These designs maximize memory access efficiency and minimize pipeline intervals, achieving the expression max min B m η ,   f k e r b k e r D . Subsequently, a reconfigurable data flow structure for repeated computational units is designed to realize high resource reuse. This ensures processing delay while reducing resource redundancy and processing power consumption, ultimately leading to a multi-fold enhancement in PPR, as shown in Table 5.

7. Conclusions

Owing to SWaP limitations, small UAVs cannot carry high-precision GPS/INS systems or large processors. The traditional SPGA algorithm struggles to ensure accuracy and requires multiple iterations, which fail to meet the demands of real-time autofocus processing. To address this issue, this paper proposes a non-iterative FSI-SPGA autofocusing algorithm, verified by measured data, which significantly reduces the computational cost while maintaining accuracy. Additionally, a high-efficiency hybrid architecture combining dataflow reconfigurability and dynamic partial reconfiguration was designed based on an ARM + FPGA platform tailored to the computational characteristics of the FSI-SPGA. For 4 K × 12 K SAR images, the FSI-SPGA algorithm demonstrated a 6× increase in processing efficiency compared with traditional methods. The high-efficiency reconfigurable ARM + FPGA architecture processed the algorithm in 6.02 s, achieving 12× the processing speed and 3× the energy efficiency of a single low-power ARM platform.
Nevertheless, the proposed solution in this paper has some limitations. First, the algorithm is not applicable to scenarios lacking isolated strong scatterers. Second, regarding hardware architecture, due to engineering complexities, there is no discussion or optimization of the data precision requirements for each processing step within the algorithm, nor is there an exploration of the effects of new devices on signal processing.
With the rapid development of microelectronics technology, innovative solutions for real-time radar signal processing have been developed. The author hopes that this article will attract the attention of the industry to the design of radar signal hardware processing architectures. In the future, new solutions that integrate software and hardware could promote the advancement of radar technologies.

Author Contributions

Conceptualization, H.W. and X.L.; methodology, H.W.; software, Y.L. (Yunlong Liu) and H.L.; validation, H.W., X.G., Y.L. (Yanlei Li), and Y.L. (Yunlong Liu); formal analysis, J.X.; data curation, X.G. and H.L.; writing—original draft preparation, H.W.; writing—review and editing, Y.L. (Yanlei Li) and J.X.; visualization, H.W.; supervision, X.L.; project administration, Y.L. (Yanlei Li) and X.L. All authors have read and agreed to the published version of this manuscript.

Funding

This research received no external funding.

Data Availability Statement

The data used to support the findings of this study are available from the corresponding author upon reasonable request. Due to privacy concerns, the data are not publicly available.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Trinder, J.C. Editorial for Special Issue “Applications of Synthetic Aperture Radar (SAR) for Land Cover Analysis”. Remote Sens. 2020, 12, 2428. [Google Scholar] [CrossRef]
  2. Agarwal, R. Recent Advances in Aircraft Technology; InTech: Rijeka, Croatia, 2012. [Google Scholar]
  3. Lort, M.; Aguasca, A.; López-Martínez, C.; Marín, T.M. Initial evaluation of SAR capabilities in UAV multicopter platforms. IEEE J. Sel. Top. Appl. Earth Observ. Remote Sens. 2018, 11, 127–140. [Google Scholar] [CrossRef]
  4. Fornado, G. Trajectory deviations in airborne SAR: Analysis and compensation. IEEE Trans. Aerosp. Electron. Syst. 1999, 35, 997–1009. [Google Scholar] [CrossRef]
  5. Fornaro, G.; Franceschetti, G.; Perna, S. Motion compensation errors: Effects on the accuracy of airborne SAR images. IEEE Trans. Aerosp. Electron. Syst. 2005, 41, 1338–1352. [Google Scholar] [CrossRef]
  6. Moreira, A.; Huang, Y. Airborne SAR processing of highly squinted data using a chirp scaling approach with integrated motion compensation. IEEE Trans. Geosci. Remote Sens. 1994, 32, 1029–1040. [Google Scholar] [CrossRef]
  7. Wang, R.; Loffeld, O.; Nies, H.; Knedlik, S.; Ender, J.H. Chirp-scaling algorithm for bistatic SAR data in the constant-offset configuration. IEEE Trans. Geosci. Remote Sens. 2009, 47, 952–964. [Google Scholar] [CrossRef]
  8. Chen, J.; Xing, M.; Yu, H.; Liang, B.; Peng, J.; Sun, G.-C. Motion compensation/autofocus in airborne synthetic aperture radar: A review. IEEE Geosci. Remote Sens. 2022, 10, 185–206. [Google Scholar] [CrossRef]
  9. Gryte, K.; Bryne, T.H.; Albrektsen, S.M.; Johansen, T.A. Field test results of GNSS-denied inertial navigation aided by phased-array radio systems for UAVs. In Proceedings of the 2019 International Conference on Unmanned Aircraft Systems (ICUAS), Atlanta, GA, USA, 11–14 June 2019; pp. 1398–1406. [Google Scholar]
  10. Zhang, J.; Fu, Y.; Zhang, W.; Yang, W. Signal characteristics analysis and intra-pulse motion compensation for FMCW CSAR. IEEE Sens. 2019, 19, 10461–10476. [Google Scholar] [CrossRef]
  11. Henke, D. Miranda35 experiments in preparation for small UAV-based SAR. In Proceedings of the GARSS 2019 IEEE International Geoscience and Remote Sensing Symposium, Yokohama, Japan, 28 July–2 August 2019. [Google Scholar]
  12. Xing, M.; Jiang, X.; Wu, R.; Zhou, F.; Bao, Z. Motion compensation for UAV SAR based on raw radar data. IEEE Trans. Geosci. Remote Sens. 2009, 47, 2870–2883. [Google Scholar] [CrossRef]
  13. Xu, G.; Xing, M.; Zhang, L.; Bao, Z. Robust autofocusing approach for highly squinted SAR imagery using the extended wavenumber algorithm. IEEE Trans. Geosci. Remote Sens. 2013, 51, 5031–5046. [Google Scholar] [CrossRef]
  14. Bezvesilniy, O.O.; Gorovyi, I.M.; Vavriv, D.M. Estimation of phase errors in SAR data by local-quadratic map-drift autofocus. In Proceedings of the 2012 13th International Radar Symposium, Warsaw, Poland, 23–25 May 2012. [Google Scholar]
  15. Calloway, T.M.; Donohoe, G.W. Subaperture autofocus for synthetic aperture radar. IEEE Trans. Aerosp. Electron. Syst. 1994, 30, 617–621. [Google Scholar] [CrossRef]
  16. Samczynski, P.; Kulpa, K.S. Coherent mapdrift technique. IEEE Trans. Geosci. Remote Sens. 2010, 48, 2010. [Google Scholar] [CrossRef]
  17. Wang, J.; Liu, X. SAR minimum-entropy autofocus using an adaptive-order polynomial model. IEEE Geosci. Remote Sens. Lett. 2006, 3, 512–516. [Google Scholar] [CrossRef]
  18. Xiong, T.; Xing, M.; Wang, Y.; Wang, S.; Sheng, J.; Guo, L. Minimum-entropy-based autofocus algorithm for SAR data using Chebyshev approximation and method of series reversion, and its implementation in a data processor. IEEE Trans. Geosci. Remote Sens. 2014, 52, 1719–1728. [Google Scholar] [CrossRef]
  19. Cacciamano, A.; Giusti, E.; Capria, A.; Martorella, M.; Berizzi, F. Contrast-optimization-based range-profile autofocus for polarimetric stepped-frequency radar. IEEE Trans. Geosci. Remote Sens. 2010, 48, 2049–2056. [Google Scholar] [CrossRef]
  20. Wahl, D.E.; Eichel, P.; Ghiglia, D.C.; Jakowatz, C.V., Jr. Phase gradient autofocus—A robust tool for high resolution SAR phase correction. IEEE Trans. Aerosp. Electron. Syst. 1994, 30, 827–835. [Google Scholar] [CrossRef]
  21. Chan, H.L.; Yeo, T.S. Noniterative quality phase-gradient autofocus (QPGA) algorithm for spotlight SAR imagery. IEEE Trans. Geosci. Remote Sens. 1998, 36, 1531–1539. [Google Scholar] [CrossRef]
  22. Ye, W.; Yeo, T.S.; Bao, Z. Weighted least-squares estimation of phase errors for SAR/ISAR autofocus. IEEE Trans. Geosci. Remote Sens. 1999, 37, 2487–2494. [Google Scholar] [CrossRef]
  23. Wahl, D.E.; Jakowatz, C.V.; Thompson, P.A.; Ghiglia, D.C. New approach to strip-map SAR autofocus. In Proceedings of the IEEE 6th Digital Signal Processing Workshop, Yosemite National Park, CA, USA, 2–5 October 1994. [Google Scholar]
  24. Thompson, D.G.; Bates, J.S.; Arnold, D.V.; Long, D.G. Extending the phase gradient autofocus algorithm for low-altitude stripmap mode SAR. In Proceedings of the 1999 IEEE Radar Conference. Radar into the Next Millennium (Cat. No.99CH36249), Waltham, MA, USA, 22 April 1999. [Google Scholar]
  25. Callow, H.J. Signal Processing for Synthetic Aperture Sonar Image Enhancement. Ph.D. Thesis, Department of Electrical and Electronic Engineering University Canterbury, Christchurch, New Zealand, 2003. [Google Scholar]
  26. Li, N. Extension and evaluation of PGA in ScanSAR mode using full-aperture approach. IEEE Geosci. Remote Sens. Lett. 2015, 12, 870–874. [Google Scholar]
  27. Li, Y.; O’Young, S. Kalman Filter Disciplined Phase Gradient Autofocus for Stripmap SAR. IEEE Trans. Geosci. Remote Sens. 2020, 58, 6298–6308. [Google Scholar] [CrossRef]
  28. Evers, A.; Jackson, J.A. A generalized phase gradient autofocus algorithm. IEEE Trans. Comput. Imag. 2019, 5, 606–619. [Google Scholar] [CrossRef]
  29. Jin, Y. Ultrahigh-Resolution Autofocusing for Squint Airborne SAR Based on Cascaded MD-PGA. IEEE Geosci. Remote Sens. Lett. 2022, 19, 4017305. [Google Scholar] [CrossRef]
  30. Deng, Y. A High-Resolution Airborne SAR Autofocusing Approach Based on SR-PGA and Subimage Resampling with Precise Hyperbolic Model. IEEE J. Sel. Topics Appl. Earth Observ. Remote Sens. 2023, 16, 2327–2338. [Google Scholar] [CrossRef]
  31. Xie, Y.; Luan, Y.; Chen, L.; Zhang, X. A Modified Iteration-Free SPGA Based on Removing the Linear Phase. Remote Sens. 2023, 15, 5535. [Google Scholar] [CrossRef]
  32. Wang, D.; Ali, M.; Blinka, E. Synthetic Aperture Radar (SAR) Implementation on a TMS320C6678 Multicore DSP. Texas Instruments 2015. Available online: https://www.ti.com.cn/cn/lit/wp/spry276/spry276.pdf (accessed on 1 January 2025).
  33. Wang, Y.; Li, W.; Liu, T.; Zhou, L.; Wang, B.; Fan, Z.; Ye, X.; Fan, D.; Ding, C. Characterization and Implementation of Radar System Applications on a Reconfigurable Dataflow Architecture. IEEE Comput. Archit. Lett. 2022, 21, 121–124. [Google Scholar] [CrossRef]
  34. Wielage, M.; Cholewa, F.; Fahnemann, C.; Pirsch, P.; Blume, H. High Performance and Low Power Architectures: GPU vs. FPGA for Fast Factorized Backprojection. In Proceedings of the 2017 Fifth International Symposium on Computing and Networking (CANDAR), Aomori, Japan, 19–22 November 2017. [Google Scholar]
  35. Mota, D.; Cruz, H.; Miranda, P.R.; Duarte, R.P.; de Sousa, J.T.; Neto, H.C.; Véstias, M.P. Onboard Processing of Synthetic Aperture Radar Backprojection Algorithm in FPGA. IEEE J. Sel. Top. Appl. Earth Observ. Remote Sens. 2022, 15, 3600–3611. [Google Scholar] [CrossRef]
  36. Li, N.; Wang, R.; Deng, Y.; Yu, W.; Zhang, Z.; Liu, Y. Autofocus Correction of Residual RCM for VHR SAR Sensors With Light-Small Aircraft. IEEE Trans. Geosci. Remote Sens. 2017, 55, 441–452. [Google Scholar] [CrossRef]
  37. Zeng, H.; Yang, W.; Wang, P.; Chen, J. A Modified PGA for Spaceborne SAR Scintillation Compensation Based on the Weighted Maximum Likelihood Estimator and Data Division. IEEE J. Sel. Topics Appl. Earth Observ. Remote Sens. 2022, 15, 3938–3947. [Google Scholar] [CrossRef]
  38. Cruz, H.; Véstias, M.; Monteiro, J.; Neto, H.; Duarte, R.P. A Review of Synthetic-Aperture Radar Image Formation Algorithms and Implementations. Remote Sens. 2022, 14, 1258. [Google Scholar] [CrossRef]
  39. RCbenchmark. Drone Building and Optimization: How to Increase Your Flight Time, Payload and Overall Efficiency. RCbenchmark. 2021. Available online: https://cdn.rcbenchmark.com/landing_pages/eBook/eBook_%20Drone%20Building%20and%20Optimization%20Chapters%201%20-%203.pdf (accessed on 14 January 2025).
  40. Diao, P.S.; Alves, T.; Poussot, B.; Azarian, S. A Review of Radar Detection Fundamentals. IEEE Trans. Aerosp. Electron. Syst. 2024, 39, 4–24. [Google Scholar] [CrossRef]
  41. Cumming, I.G.; Wong, F.H.-C. Digital Processing of Synthetic Aperture Radar Data: Algorithms and Implementation; Artech House: Boston, MA, USA, 2005. [Google Scholar]
Figure 1. Small UAV-borne SAR real-time imaging scheme.
Figure 1. Small UAV-borne SAR real-time imaging scheme.
Remotesensing 17 02232 g001
Figure 2. The trajectory of the frontal and lateral view strip pattern deviates from the imaging geometry.
Figure 2. The trajectory of the frontal and lateral view strip pattern deviates from the imaging geometry.
Remotesensing 17 02232 g002
Figure 3. Flowchart of autofocusing algorithm. (a) QW-SPGA algorithm. (b) FSI-SPGA algorithm.
Figure 3. Flowchart of autofocusing algorithm. (a) QW-SPGA algorithm. (b) FSI-SPGA algorithm.
Remotesensing 17 02232 g003
Figure 4. Schematic diagram of 2D points selection. (a) Coarse points selection. (b) Fine-grained filtering of points.
Figure 4. Schematic diagram of 2D points selection. (a) Coarse points selection. (b) Fine-grained filtering of points.
Remotesensing 17 02232 g004
Figure 5. High-Quality scattering point selection test for real SAR image Data. (a) Coarse point selection results. (b) Refined point selection results. (c) Azimuth continuity of scatterer sets.
Figure 5. High-Quality scattering point selection test for real SAR image Data. (a) Coarse point selection results. (b) Refined point selection results. (c) Azimuth continuity of scatterer sets.
Remotesensing 17 02232 g005
Figure 6. Assessment of defocus range for point targets. (a) Quadratic phase error curve. (b) Higher-order phase error curve. (c) Pulse compression results and energy sidelobe range assessment of ideal LFM signal vs. signal with phase errors.
Figure 6. Assessment of defocus range for point targets. (a) Quadratic phase error curve. (b) Higher-order phase error curve. (c) Pulse compression results and energy sidelobe range assessment of ideal LFM signal vs. signal with phase errors.
Remotesensing 17 02232 g006
Figure 7. High-quality scattering target feature map extraction.
Figure 7. High-quality scattering target feature map extraction.
Remotesensing 17 02232 g007
Figure 8. The process of estimating the phase error of the feature sub-image.
Figure 8. The process of estimating the phase error of the feature sub-image.
Remotesensing 17 02232 g008
Figure 9. Schematic diagram of scattering point sub-aperture phase gradient stitching. (a) Scattered point phase error gradient interval estimation. (b) Phase error gradient stitching and linear term removal.
Figure 9. Schematic diagram of scattering point sub-aperture phase gradient stitching. (a) Scattered point phase error gradient interval estimation. (b) Phase error gradient stitching and linear term removal.
Remotesensing 17 02232 g009
Figure 10. Test results of the FSI-SPGA algorithm. (a) Initial motion compensation of the defocused SAR image. (b) Extracted feature sub-images. (c) Phase error gradient curves. (d) SAR image processed by the FSI-SPGA algorithm.
Figure 10. Test results of the FSI-SPGA algorithm. (a) Initial motion compensation of the defocused SAR image. (b) Extracted feature sub-images. (c) Phase error gradient curves. (d) SAR image processed by the FSI-SPGA algorithm.
Remotesensing 17 02232 g010
Figure 11. SAR real-time autofocus hardware system architecture.
Figure 11. SAR real-time autofocus hardware system architecture.
Remotesensing 17 02232 g011
Figure 12. FSI-SPGA computational flowchart and ARM + FPGA processor task mapping scheme.
Figure 12. FSI-SPGA computational flowchart and ARM + FPGA processor task mapping scheme.
Remotesensing 17 02232 g012
Figure 13. FPGA Hardware Accelerator Model. (a) Dataflow Architecture Hardware Model; (b) Accelerator Timing Diagram with Different Pipeline Intervals.
Figure 13. FPGA Hardware Accelerator Model. (a) Dataflow Architecture Hardware Model; (b) Accelerator Timing Diagram with Different Pipeline Intervals.
Remotesensing 17 02232 g013
Figure 14. Total Task Reconfiguration Timeline.
Figure 14. Total Task Reconfiguration Timeline.
Remotesensing 17 02232 g014
Figure 15. 2D-CFAR Ping-Pong Pipeline Scheme. (a) Add-subtract sliding summation unit. (b) Operations on Buffer 1 during the Ping Phase. (c) Operations on Buffer 2 during the Ping Phase.
Figure 15. 2D-CFAR Ping-Pong Pipeline Scheme. (a) Add-subtract sliding summation unit. (b) Operations on Buffer 1 during the Ping Phase. (c) Operations on Buffer 2 during the Ping Phase.
Remotesensing 17 02232 g015
Figure 16. Reconfigurable RMF Dataflow Architecture Schematic Diagram. (a) Complete RMF accelerator architecture (b) Mode 1. (c) Mode 2. (d) Mode 3. (e) Mode 4.
Figure 16. Reconfigurable RMF Dataflow Architecture Schematic Diagram. (a) Complete RMF accelerator architecture (b) Mode 1. (c) Mode 2. (d) Mode 3. (e) Mode 4.
Remotesensing 17 02232 g016
Figure 17. Illustration of the Ku-band multi-rotor UAV-Borne SAR system. (a) UAV SAR system. (b) SAR data transmission and reception module.
Figure 17. Illustration of the Ku-band multi-rotor UAV-Borne SAR system. (a) UAV SAR system. (b) SAR data transmission and reception module.
Remotesensing 17 02232 g017
Figure 18. Ku-band SAR imaging results. (a) Initial MoCo; (b) SPGA method; (c) QW-SPGA method; (d) FSI-SPGA method.
Figure 18. Ku-band SAR imaging results. (a) Initial MoCo; (b) SPGA method; (c) QW-SPGA method; (d) FSI-SPGA method.
Remotesensing 17 02232 g018aRemotesensing 17 02232 g018b
Figure 19. Close-up views of area 1. (a) Initial MoCo; (b) SPGA method; (c) QW-SPGA method; (d) FSI-SPGA method.
Figure 19. Close-up views of area 1. (a) Initial MoCo; (b) SPGA method; (c) QW-SPGA method; (d) FSI-SPGA method.
Remotesensing 17 02232 g019
Figure 20. Close-up views of area 2. (a) Initial MoCo. (b) SPGA method. (c) QW-SPGA method. (d) FSI-SPGA method. Azimuth Point Target Slice Zoomed View. (e) P1. (f) P2. (g) P3. (h) P4.
Figure 20. Close-up views of area 2. (a) Initial MoCo. (b) SPGA method. (c) QW-SPGA method. (d) FSI-SPGA method. Azimuth Point Target Slice Zoomed View. (e) P1. (f) P2. (g) P3. (h) P4.
Remotesensing 17 02232 g020
Figure 21. Autofocus test results of four 4 K × 4 K defocused SAR images: (ad) Initial MoCo (D1–D4); (eh) QW-SPGA Method (D1–D4); (il) FSI-SPGA Method (D1–D4).
Figure 21. Autofocus test results of four 4 K × 4 K defocused SAR images: (ad) Initial MoCo (D1–D4); (eh) QW-SPGA Method (D1–D4); (il) FSI-SPGA Method (D1–D4).
Remotesensing 17 02232 g021aRemotesensing 17 02232 g021b
Figure 22. FPGA Accelerator Post-Place-and-Route Results. (a) ZCU102 platform. (b) 2D CFAR Hardware Accelerator. (c) RMF Hardware Accelerator.
Figure 22. FPGA Accelerator Post-Place-and-Route Results. (a) ZCU102 platform. (b) 2D CFAR Hardware Accelerator. (c) RMF Hardware Accelerator.
Remotesensing 17 02232 g022
Figure 23. FSI-SPGA algorithm hardware acceleration test result. (a) Initial MoCo. (b) FSI-SPGA method on ZU9EG’s ARM. (c) FSI-SPGA method on AMD 5800 H CPU. (d) FSI-SPGA method on ZU9EG’s ARM + FPGA.
Figure 23. FSI-SPGA algorithm hardware acceleration test result. (a) Initial MoCo. (b) FSI-SPGA method on ZU9EG’s ARM. (c) FSI-SPGA method on AMD 5800 H CPU. (d) FSI-SPGA method on ZU9EG’s ARM + FPGA.
Remotesensing 17 02232 g023
Table 1. Main System parameters for the experiment.
Table 1. Main System parameters for the experiment.
SymbolParametersValues (Units)
Imaging ModeStripmap
WaveformChirp pulses
BandKu
B Frequency Bandwidth480 MHz
θ a Azimuth Beam Width
θ i Incident Angle82°
F s Sampling Rate480 MHz
H Flying Height500 m
v Platform Velocity12 m/s
Table 2. Experimental results for quantitative analysis of focusing accuracy.
Table 2. Experimental results for quantitative analysis of focusing accuracy.
Autofocus AlgorithmStrong
Scatterer
Azimuth PSFFull Image
Res. (m)PSLR (dB)ISLR (dB)EntropyContrastRuntime (s)
Initial MoCoP10.33−1.006.265.484.336.21
P21.23−4.21−2.66
P30.40−1.684.93
P40.95−6.00−1.08
SPGA (8 iterations)P10.46−4.42−1.035.136.2595.6
P20.38−9.74−5.94
P30.44−7.88−7.18
P40.36−13.80−11.21
QW-SPGA (8 iterations)P10.2812.4410.384.917.1983.4
P20.2817.489.20
P30.31−10.91−9.10
P40.2816.91−11.41
FSI-SPGA (1 iteration)P10.30−10.82−7.824.987.166.12
P20.29−13.18−8.47
P30.2917.5110.68
P40.28−14.2111.61
Table 3. Quantitative analysis results of autofocus methods.
Table 3. Quantitative analysis results of autofocus methods.
Autofocus AlgorithmSAR ImageEntropyContrastRuntime (s)
Initial MoCoD16.8410.713.1
D25.4627.593.2
D36.2221.033.3
D46.1011.443.2
QW-SPGA (8 iterations)D16.6235.1514.5
D25.0761.1214.2
D36.1029.0115.1
D46.0613.5913.9
FSI-SPGA (1 iteration)D16.6233.993.3
D25.1054.933.4
D36.1027.703.6
D46.0517.613.1
Table 4. The resource usage of FPGA hardware.
Table 4. The resource usage of FPGA hardware.
AreaAccelClockLUTFFBRAMDSPs
Static Region/300 MHz259161096426
Dynamic Region2D-CFAR18,23326,286128.514
RMF82,982129,400472.5379
Table 5. The performance test results of the algorithm on a hardware platform.
Table 5. The performance test results of the algorithm on a hardware platform.
PlatformProcessor ModelAlgorithmImage Size
(pixel × pixel)
Runtime (s)Power (W)PPR (pixels/J)
ARMAMD ZU9EG (16 nm)QW-SPGA4 K × 12 K566.42.537,026
8 K × 12 K1305.232,135
FSI-SPGA4 K × 12 K83.4317,630
8 K × 12 K162.9325,230
CPUAMD
5800 H (7 nm)
QW-SPGA4 K × 12 K56.53029,694
8 K × 12 K130.325,751
FSI-SPGA4 K × 12 K7.3229,824
8 K × 12 K16.8199,728
MPSoC
(ARM + FPGA)
AMD ZU9EG (16 nm)FSI-SPGA4 K × 12 K6.126.91,191,902
8 K × 12 K6.362,293,849
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wang, H.; Liu, Y.; Li, Y.; Li, H.; Ge, X.; Xin, J.; Liang, X. Improved Real-Time SPGA Algorithm and Hardware Processing Architecture for Small UAVs. Remote Sens. 2025, 17, 2232. https://doi.org/10.3390/rs17132232

AMA Style

Wang H, Liu Y, Li Y, Li H, Ge X, Xin J, Liang X. Improved Real-Time SPGA Algorithm and Hardware Processing Architecture for Small UAVs. Remote Sensing. 2025; 17(13):2232. https://doi.org/10.3390/rs17132232

Chicago/Turabian Style

Wang, Huan, Yunlong Liu, Yanlei Li, Hang Li, Xuyang Ge, Jihao Xin, and Xingdong Liang. 2025. "Improved Real-Time SPGA Algorithm and Hardware Processing Architecture for Small UAVs" Remote Sensing 17, no. 13: 2232. https://doi.org/10.3390/rs17132232

APA Style

Wang, H., Liu, Y., Li, Y., Li, H., Ge, X., Xin, J., & Liang, X. (2025). Improved Real-Time SPGA Algorithm and Hardware Processing Architecture for Small UAVs. Remote Sensing, 17(13), 2232. https://doi.org/10.3390/rs17132232

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop