Next Article in Journal
Horizontal Attack Against EC kP Accelerator Under Laser Illumination
Previous Article in Journal
Time Series Forecasting via an Elastic Optimal Adaptive GM(1,1) Model
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

GPU-Optimized Implementation for Accelerating CSAR Imaging

1
Institute of Biomedical Engineering, School of Health Science and Engineering, University of Shanghai for Science and Technology, Shanghai 200093, China
2
Terahertz Technology Innovation Research Institute, School of Optical-Electrical and Computer Engineering, University of Shanghai for Science and Technology, Shanghai 200093, China
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Electronics 2025, 14(10), 2073; https://doi.org/10.3390/electronics14102073
Submission received: 31 March 2025 / Revised: 15 May 2025 / Accepted: 16 May 2025 / Published: 20 May 2025

Abstract

:
The direct porting of the Range Migration Algorithm to GPUs for three-dimensional (3D) cylindrical synthetic aperture radar (CSAR) imaging faces difficulties in achieving real-time performance while the architecture and programming models of GPUs significantly differ from CPUs. This paper proposes a GPU-optimized implementation for accelerating CSAR imaging. The proposed method first exploits the concentric-square-grid (CSG) interpolation to reduce the computational complexity for reconstructing a uniform 2D wave-number domain. Although the CSG method transforms the 2D traversal interpolation into two independent 1D interpolations, the interval search to determine the position intervals for interpolation results in a substantial computational burden. Therefore, binary search is applied to avoid traditional point-to-point matching for efficiency improvement. Additionally, leveraging the partition independence of the grid distribution of CSG, the 360° data are divided into four streams along the diagonal for parallel processing. Furthermore, high-speed shared memory is utilized instead of high-latency global memory in the Hadamard product for the phase compensation stage. The experimental results demonstrate that the proposed method achieves CSAR imaging on a 1440 × 100 × 128 dataset in 0.794 s, with an acceleration ratio of 35.09 compared to the CPU implementation and 5.97 compared to the conventional GPU implementation.

1. Introduction

Cylindrical synthetic aperture radar (CSAR), a variant of SAR technology that allows for 360° omnidirectional detection to eliminate blind zones, is widely used in security and non-destructive inspection fields [1,2,3,4]. CSAR can obtain high-resolution three-dimensional (3D) images of the target for detection. However, the execution of the CSAR imaging system on Central Processing Unit (CPU) platforms faces difficult challenges in real-time imaging [5,6].
As shown in Figure 1, Graphics Processing Units (GPUs) with powerful parallel computing capabilities have emerged as powerful accelerators for computationally intensive tasks in SAR imaging [7,8]. Plenty of studies have explored GPU-accelerated strategies to address real-time imaging problems [9,10,11,12,13,14,15,16]. Wu [10] proposed a scheme to accelerate the Back-Projection Algorithm (BPA) by leveraging the combination of CPU and GPU platforms. At an image scale of 4 km × 4 km, this approach achieved a speedup of 163.96 times compared to the CPU-only implementation. Yang [11] proposed a parallel computing method for Range-Doppler and Map Drift algorithms through unifying memory management, shared memory, and cuFFT configuration. For 2048 × 2048 data, the proposed method realized imaging in 135 milliseconds based on an embedded GPU platform, achieving a speedup of 39.78 times over the CPU-based approach. Li [12] proposed an efficient BPA based on time-shifted upsampling-improved cubic spline interpolation, which was accelerated on GPUs using shared memory and 2D thread. The proposed method demonstrated a performance improvement of 1.79 times compared to the approach utilizing 1D threads and global memory. Tan [13] proposed a combined GPU and OpenMp-based approach to greatly accelerate ground-based SAR real-time imaging through host stream, blocking stream, and non-blocking stream multi-stream processing. The proposed method achieved imaging in 0.637 s for 260 × 15360 data, which was 22.8 times faster than conventional imaging. Ren Z. [15] proposed a GPU-based extended two-step approach for full-aperture SAR data processing, consisting of data division and the stream technique. The proposed method processed 32768 × 4996 SAR data in 45.78 s and achieved more than 18.75 times faster than CPU processing. Ding L. [16] proposed a hybrid GPU and CPU parallel computing method to accelerate the Range Migration Algorithm (RMA), which processed 201 × 320 × 320 data 15 times faster than the CPU-only method and 2 times faster than the GPU-only method. Previous works have shown that the powerful parallel computing capability of GPUs significantly improves the processing efficiency of SAR imaging. However, most studies focus on other SAR imaging modalities rather than CSAR. With the increase in dimensionality and changes in application scenarios, it is still an important challenge to fully optimize the CSAR imaging algorithms on the GPU platform.
RMA is commonly used for 3D CSAR imaging due to its ability to effectively eliminate distance migration with low computational complexity and minimal approximation errors [17,18,19,20]. The main stages of this algorithm include 2D Fourier transform (FT), phase compensation, 1D inverse FT, 2D interpolation, and 3D inverse FT. The direct porting of RMA makes it hard to effectively improve the real-time performance due to the significant differences in architecture and programming models of CPU and GPU. Therefore, this paper optimizes the GPU implementation based on the detailed analysis of each stage of the 3D cylindrical RMA.
Operations involving the FTs and inverse FTs can be efficiently implemented using the cuFFT library provided by Computational Unified Device Architecture (CUDA). In the phase-compensation stage, high-speed shared memory is employed to accelerate the Hadamard product instead of global memory with the highest latency. Two-dimensional interpolation constitutes the most computationally intensive stage of RMA due to its high computational complexity and logic-intensive operations required for reconstructing a uniform 2D wave-number domain. To address this, this paper adopts the concentric-square-grid (CSG) interpolation with low computational complexity [21], which transforms the conventional 2D traversal interpolation into two independent 1D interpolations. A key challenge in the GPU implementation of the CSG method is how to efficiently determine the position intervals of the points to be interpolated. A binary search is introduced to enhance search efficiency by halving the search range at each iteration. Furthermore, the proposed method leverages the partition independence of the grid distribution of CSG, allowing the 360° data to be divided into four CUDA streams for parallel processing to improve interpolation efficiency. In this way, the proposed method optimizes the 3D cylindrical RMA by improving memory access, interpolation interval search, and parallel processing. Tested on a hardware setup comprising a 13th Gen Intel® Core™ i9-13900K CPU and NVIDIA GeForce RTX 4090 GPU, the proposed method achieves 3D CSAR imaging of 1682 × 1682 × 128 (azimuth × distance × height) points in 0.794 s for the echo data of 1440 × 100 × 128 (angle × frequency × height) points, which indicates an acceleration ratio of 35.09 compared to the CPU implementation and 5.97 compared to the conventional GPU implementation.

2. Three-Dimensional Cylindrical RMA

The 3D cylindrical imaging scene is shown in Figure 2. The height of the antenna array is L z , scanning circularly with radius R. θ [ π , π ] denotes the azimuth of antenna rotation at a given moment. Let the position coordinates of any sample of the antenna array be ( x , y , z ) = ( R cos θ , R sin θ , z ) , where R and θ are the radius and angle of the sampled antenna in the polar coordinate system. A scattering point on the target is denoted as ( x , y , z ) = ( r cos φ , r sin φ , z ) , where r and φ are the radius and angle of the target point E in the polar coordinate system.
The received reflected echo signal can be expressed as [22,23,24]:
s ( θ , k , z ) = x , y , z σ ( x , y , z ) e j 2 k ( R cos θ x ) 2 + ( R sin θ y ) 2 + ( z z ) 2 d x d y d z ,
where σ ( x , y , z ) corresponds to the backscattering coefficient, c is the speed of light, and k = ω / c is the wave number.
In order to realize the digital processing, discrete sampling is required for the angle, frequency, and height dimensions of the continuous echo signal.
θ m = m · Δ θ m = 0 , 1 , , M f n = n · Δ f n = 0 , 1 , , N z p = p · Δ z p = 0 , 1 , , P
where Δ θ , Δ f , Δ z denote the sampling intervals for the three dimensions and M , N , P denote the numbers of sampling points for the three dimensions. The echo corresponds to the m-th angle, n-th frequency, and the p-th height s ( θ m , k n , z p ) can be rewritten as:
s ( θ m , k n , z p ) = u = 0 U 1 v = 0 V 1 p = 0 P 1 σ ( x u , y v , z p ) e j 2 k n ( R cos θ m x u ) 2 + ( R sin θ m y v ) 2 + ( z p z p ) 2 ,
where U , V , P denote the numbers of points in the azimuth, as well as the distance and height dimensions of the scene of interest, and k n = 2 π f n / c denotes the wave number at the n-th frequency. Based on Equation (3), the details of 3D cylindrical RMA are shown in Algorithm 1.   
Algorithm 1: Three-dimensional cylindrical RMA
  • Input:
    • The echo matrix S C M × N × P is formed by collecting the observations s ( θ m , k n , z p ) in three dimensions, where M , N , and P , respectively, denote the numbers of points in the angle, frequency, and height dimensions of the echo matrix.
      S ( : , k n , : ) = s ( θ 0 , k n , z 0 ) s ( θ 0 , k n , z P 1 ) s ( θ M 1 , k n , z 0 ) s ( θ M 1 , k n , z P 1 )
  • Output:
    • The imaging matrix Q C U × V × P , where U , V , and P , respectively, denote the numbers of points in the angle, frequency, and height dimensions of the image matrix.
1:
Two-dimensional FT: convert time-domain data to wave-number domain along the height and angle dimensions.
S 1 ( : , k n , : ) = F T θ , z S ( : , k n , : )
2:
Phase compensation: correct phase deviations due to the motion of the radar platform and the position of the target.
S 2 ( : , k n , : ) = S 1 ( : , k n , : ) ϕ ( : , k n , : )
ϕ ( : , k n , : ) = ϕ ( θ 0 , k n , z 0 ) ϕ ( θ 0 , k n , z P 1 ) ϕ ( θ M 1 , k n , z 0 ) ϕ ( θ M 1 , k n , k n , z P 1 )
ϕ ( θ m , k n , z p ) = e k r ( n , p ) 2 k θ m 2 + k θ m π + k θ m arcsin k θ m R k r ( n , p )
where ⊙ denotes the Hadamard product, the operation of multiplying two matrices or vectors of the same dimension element by element, and k θ m , k z p , k r ( n , p ) = 4 k n 2 k z p 2 denote the wave-number components corresponding to the m-th angle, p-th height, and radial dimension in the polar coordinate system [25].
3:
Angular-dimensional inverse FT: take the inverse FT along the angle dimension.
S 3 ( : , k n , : ) = I F T k θ S 2 ( : , k n , : )
4:
Two-dimensional interpolation: turn the non-uniform wave-number domain to the uniform wave-number domain.
S 4 ( : , : , k z p ) = i n t e r p k x , k y S 3 ( : , : , k z p ) = S 4 ( k ˜ x 0 , k ˜ y 0 , k z p ) S 4 ( k ˜ x 0 , k ˜ y V 1 , k z p ) S 4 ( k ˜ x U 1 , k ˜ y 0 , k z p ) S 4 ( k ˜ x U 1 , k ˜ y V 1 , k z p )
where k x , k y denote the wave-number components corresponding to the azimuth and distance dimensions, i n t e r p k x , k y · denotes the interpolation function for the distance and azimuth dimensions, and k ˜ x u and k ˜ y v denote the uniformly sampled wave numbers corresponding to the u-th distance and the v-th azimuth.
5:
Three-dimensional inverse FT: take the 3D inverse FT to achieve the imaging matrix.
Q = I F T k ˜ x , k ˜ y , k z S 4

3. The Implementation on GPU

Based on Algorithm 1, it can be observed that the 3D cylindrical RMA primarily consists of the operations of FTs, inverse FTs, the Hadamard product, and 2D interpolation. The flow of the proposed method is shown in Figure 3. The red arrows represent the GPU implementations of the indicated algorithmic stages.

3.1. Fourier Transform

As shown in Figure 3, FTs and inverse FTs are required in several stages. Among the available libraries for FT, the cuFFT library is specifically optimized for GPU architectures, exploiting their massive parallelism to significantly achieve lower execution times compared to CPU-based libraries such as FFTW. Furthermore, the cuFFT library offers seamless integration with the CUDA programming environment, ensuring efficient memory management and minimal overhead. These advantages make the cuFFT library the optimal choice for implementing the Fourier transform stage.

3.2. Phase Compensation

The phase-compensation stage in CUDA is implemented as the Hadamard product, which is highly parallelizable and well suited for GPU-based execution. Conventionally, the Hadamard product is operated in global memory. However, global memory is the memory with the highest latency and lowest access speed, potentially leading to significant performance bottlenecks. Shared memory offers a latency reduction of approximately 20 to 30 times [26,27], making it much more efficient for memory-intensive operations. By utilizing shared memory within the kernel function, the phase-compensation stage can be optimized to improve memory access speed and reduce latency. The detailed operations are shown in Algorithm 2.
Algorithm 2: Phase compensation
  • Input: the array after 2D FT S 1 , the array of phase-compensation terms ϕ .
  • Output: the array after phase compensation S 2 .
  • 1: i d = b l o c k D i m . x × b l o c k I d . x + t h r e a d I d x . x the global thread index
  • 2: extern shared float2 t e m p [ ] the array stored in shared memory
  • 3: float2 ∗shareda = temp
  • 4: float2 ∗sharedb = temp[blockDim.x]
  • 5: if  i d N  then
  • 6:   t e m p [ t h r e a d I d x . x ] = 0.0
  • 7: else
  • 8:   s h a r e d a [ t h r e a d I d x . x ] = S 1 [ i d ]
  • 9:   s h a r e d a [ t h r e a d I d x . x ] = ϕ [ i d ]
  • 10: end if
  • 11: syncthread(); ← synchronizing all threads in the block
  • 12: S 2 [ i d ] = s h a r e d a [ t h r e a d I d x . x ] × s h a r e d b [ t h r e a d I d x . x ] copying data from the shared memory to the global memory
  • 13: return  S 2

3.3. Two-Dimensional CSG Interpolation

The 2D interpolation stage exploits the CSG method [21], which transforms traditional 2D interpolation into two independent 1D interpolations. This method reduces computational complexity without compromising imaging accuracy and makes it easy to implement parallelization on GPU platforms.The principle of CSG interpolation is shown in Figure 4. The original wave-number domain is divided into four intervals along the square diagonal, as shown in Figure 4a. Radial interpolation and distance interpolation are applied to the data in intervals 1 and 3, while radial interpolation and azimuth interpolation are performed on the data in intervals 2 and 4.
In the GPU-based implementation of CSG interpolation, a key challenge is the large number of interval-search operations required for interpolation tasks. The interval search determines the position interval of points to be interpolated, which are sequentially arranged in a fixed and static order. This distribution satisfies the prerequisite condition for binary search, an efficient approach for locating data points. Taking radial interpolation as an example, let the initial radial position array be listed as k n , n [ 0 , 1 , , N 1 ] , and the uniform radial position array as k ˜ l , l [ 0 , 1 , , L 1 ] , where N and L denote the lengths of the arrays. The linear search and binary search are shown in Figure 5, where finding the corresponding position in k n for k ˜ L / 4 is taken as an example, where N and L are even numbers and k N / 4 k ˜ L / 4 k N / 4 + 1 .
As shown in Figure 5a, linear search involves each element in the array k ˜ l being sequentially compared with each two adjacent elements in the array k n to determine the interval where a specific element from k ˜ l . Therefore, the time complexity is O ( N × L ) . Leveraging the parallelism capability of GPU, the comparisons for each element in array k ˜ l can be executed concurrently across multiple threads, reducing the time complexity to O ( N ) . Although the steps of binary search are more tedious than the conventional method, it halves the search range with each iteration until the target interval is located, which reduces the time complexity to O ( log 2 N ) [28,29]. The pseudo-code is shown in Algorithm 3. From Figure 6, it can be seen that the larger the size of the interpolated data, the better the performance of binary search.
Algorithm 3: Binary search
  • Input: the initial position array k , the array of phase-compensation terms k ˜ .
  • Output: the interval array R .
  • 1:  i d = b l o c k D i m . x × b l o c k I d . x + t h r e a d I d x . x the global thread index
  • 2:  left = 1; right = N − 1 N , the length of k
  • 3:  while left ≤ right do
  • 4:  mid = (left + right)/2
  • 5:  if  k ˜ i d k m i d & & k ˜ i d k m i d  then
  • 6:    return R i d
  • 7:  else if  k ˜ i d > k m i d  then
  • 8:    left←mid + 1
  • 9:  else
  • 10:    right←mid − 1 s h a r e d a [ t h r e a d I d x . x ] = ϕ [ i d ]
  • 11:  end if
  • 12: end while
  • 13: return 0
Since the processing of the four groups of data divided based on the partition independence of the grid distribution of CSG is independent of each other, it is well suited for parallel execution. To maximize GPU computational efficiency, the CUDA streaming technique is utilized to assign each group of data to a separate stream, as shown in Figure 7. The upper dashed line shows the conventional interpolation method without the specified streams, while the lower dashed line shows the parallel execution of four streams. Four groups of data divided along the diagonal of the CGS grid are bound to independent CUDA streams to realize coarse-grained pipeline parallelism. Within each stream, CUDA events are used to ensure the sequential execution of kernel functions, including radial zero-padding, radial interpolation, and 1D interpolation (the distance or azimuth interpolation). Meanwhile, the inter-stream asynchronous feature is utilized to realize the parallel processing of the four groups of data. A global synchronization point, established via cudaStreamSynchronize, ensures that all streams complete their tasks before the four groups of data are stitched together for 3D inverse FT to obtain the final imaging.

4. Experimental Results and Analysis

In this section, experiments are conducted to validate the performance of the proposed GPU implementation and to check the imaging differences with the CPU and the conventional GPU implementation. The experimental hardware platforms are 13th Gen Intel® Core™i9-13900K CPU and NVIDIA GeForce RTX 4090 GPU. The experimental software platforms are MATLAB R2024b, Windows Visual Studio 2022 and CUDA 12.4. The parameters of the simulation experiment are enumerated in Table 1 and the detailed implementation of the three methods is shown in Table 2. The single-point simulation experimental imaging results of the CPU implementation and the proposed GPU implementation are shown in Figure 8 and Figure 9.
Figure 8 shows the front, top, and side views of two imaging results obtained from the CPU implementation and the proposed GPU implementation, which are clearly shown and have no obvious differences when observed with the human eye. In order to further confirm the correctness of the proposed GPU-implemented imaging data, the above imaging results are profiled in azimuth, distance, and height dimensions in Figure 9. As shown in Table 3, the imaging quality metrics of the proposed GPU implementation, including peak-side lobe ratio (PSLR), integrated-side lobe ratio (ISLR), and impulse response width (IRW), remain essentially the same as with those of the CPU implementation.
The actual imaging experiments of the CPU and the proposed GPU implementation are performed, where the parameters are the same as those of the simulation experiment. The experimental results are shown in Figure 10. The imaging quality metrics, including structural similarity index (SSIM) and peak signal-to-noise ratio (PSNR), are shown in Table 4. It can be determined that the proposed GPU implementation is capable of correct imaging.
The comparison of the execution time of each for three methods is shown in Figure 11. The total time is specified from the beginning of parameter setting, reading echo data, to the successful construction of the imaging matrix. The computation time of the GPU-CUDA platform is much smaller than that of the CPU-MATLAB platform, both for the whole process and for each stage. The proposed GPU implementation achieves an acceleration ratio of 35.09 compared to the CPU implementation and 5.97 compared to the conventional GPU implementation. It can be seen that the proposed method can significantly improve imaging efficiency.
To further evaluate the acceleration performance of the proposed GPU implementation, the acceleration ratio is tested for processing the echo data with different frequency-dimension points. As shown in Figure 11e, the maximum time-consuming stage of RMA is 2D interpolation. N , the number of the frequency-dimension points which can affect the calculation of 2D interpolation and thus change the size of the imaging matrix, is changed to 25, 50, 75, 100 and 110. The results obtained by processing the echo data with different numbers of frequency-dimension points are shown in Figure 12. When N is less than 50, the acceleration ratio of the proposed GPU implementation over the CPU implementation decreases with the increase in the number of points. When N is greater than 50, the acceleration ratio of the proposed GPU implementation against the CPU implementation stabilizes at around 35. The speedup ratio of the proposed GPU implementation over the conventional GPU implementation increases with the increase in the number of points, but the increase is not significant. Therefore, when the data size increases, the speedup of the proposed method over CPU and conventional GPU stabilizes.
To quantify the contribution of an optimization technique to the overall performance of the proposed GPU implementation, an ablation study is conducted. Binary search and CUDA stream are optimization techniques based on CSG interpolation, so the ablation experiments for these two techniques need to be conducted based on CSG interpolation. As shown in Table 5, the CGS interpolation contributes the most to the performance improvement of the algorithm, where stream parallelism in the CSG optimization technique leads to better performance improvement.

5. Conclusions

This paper presents a comprehensive analysis of the 3D cylindrical RMA and its hardware acceleration with GPU on the CUDA platform. The proposed method exploits four CUDA streams for the parallel execution of the CSG method and employs binary search to improve interval search, significantly reducing computational complexity. Additionally, shared memory is utilized to accelerate the Hadamard product, cutting time consumption during the phase-compensation stage. Tested on a hardware setup comprising a 13th Gen Intel® Core® i9-13900K CPU and NVIDIA GeForce RTX 4090 GPU, the proposed method achieves CSAR imaging in 0.794 s in data of 1440 × 100 × 128 . It is 35.09 times faster compared to CPU implementation and 5.97 times faster than the conventional GPU implementation. This conclusively demonstrates that the proposed method achieves substantial performance improvements over both CPU and conventional GPU implementations at this configuration, validating its effectiveness for real-time 3D CSAR imaging. When the data size increases to a certain threshold, the acceleration achieved by the proposed method tends to stabilize when compared to the CPU and conventional GPU implementations.

Author Contributions

Conceptualization, M.C. and P.L.; methodology, M.C. and P.L.; software, M.C. and P.L.; validation, M.C., P.L. and Z.B.; formal analysis, M.C.; investigation, P.L.; resources, M.C., P.L. and M.X.; data curation, Z.B. and M.X.; writing—original draft preparation, M.C. and L.D.; writing—review and editing, M.C. and P.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

The data can be made available on request.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Lorente, D.; Limbach, M.; Gabler, B.; Esteban, H.; Boria, V.E. Sequential 90 Rotation of Dual-polarized Antenna Elements in Linear Phased Arrays with Improved Cross-polarization Level for Airborne Synthetic Aperture Radar Applications. Remote Sens. 2021, 13, 1430. [Google Scholar] [CrossRef]
  2. Liu, C.A.; Chen, Z.X.; Shao, Y.; Chen, J.S.; Hasi, T.; Pan, H.Z. Research Advances of SAR Remote Sensing for Agriculture Applications: A Review. J. Integr. Agric. 2019, 18, 506–525. [Google Scholar] [CrossRef]
  3. Gao, J.; Deng, B.; Qin, Y.; Wang, H.; Li, X. An Efficient Algorithm for MIMO Cylindrical Millimeter-wave Holographic 3-D Imaging. IEEE Trans. Microw. Theory Tech. 2018, 66, 5065–5074. [Google Scholar] [CrossRef]
  4. Sheen, D.M.; Jones, A.M.; Hall, T.E. Simulation of Active Cylindrical and Planar Millimeter-wave Imaging Systems. In Proceedings of the Passive and Active Millimeter-Wave Imaging XXI, Orlando, FL, USA, 15–19 April 2018; SPIE: Bellingham, WA USA, 2018; pp. 47–57. [Google Scholar]
  5. Alibakhshikenari, M.; Virdee, B.S.; Limiti, E. Wideband Planar Array Antenna Based on SCRLH-TL for Airborne Synthetic Aperture Radar Application. J. Electromagn. Waves Appl. 2018, 32, 1586–1599. [Google Scholar] [CrossRef]
  6. Appleby, R.; Wallace, H.B. Standoff Detection of Weapons and Contraband in the 100 GHz to 1 THz Region. IEEE Trans. Antennas Propag. 2007, 55, 2944–2956. [Google Scholar] [CrossRef]
  7. Kim, B.; Yoon, K.S.; Kim, H.J. Gpu-accelerated Laplace Equation Model Development Based on CUDA Fortran. Water 2021, 13, 3435. [Google Scholar] [CrossRef]
  8. Gao, Y.; Han, J.; Tian, H.; Zhang, R.; Zheng, S.; Wang, H. GPU Acceleration Method Based on Range-Doppler Highly Squint Algorithm. In Proceedings of the 2023 4th China International SAR Symposium (CISS), Xi’an, China, 4–6 December 2023; IEEE: New York, NY, USA, 2023; pp. 1–4. [Google Scholar]
  9. Cui, Z.; Quan, H.; Cao, Z.; Xu, S.; Ding, C.; Wu, J. SAR Target CFAR Detection via GPU Parallel operation. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2018, 11, 4884–4894. [Google Scholar] [CrossRef]
  10. Wu, S.; Xu, Z.; Wang, F.; Yang, D.; Guo, G. An Improved Back-projection Algorithm for GNSS-R BSAR Imaging Based on CPU and GPU Platform. Remote Sens. 2021, 13, 2107. [Google Scholar] [CrossRef]
  11. Yang, T.; Zhang, X.; Xu, Q.; Zhang, S.; Wang, T. An Embedded-GPU-Based Scheme for Real-Time Imaging Processing of Unmanned Aerial Vehicle Borne Video Synthetic Aperture Radar. Remote Sens. 2024, 16, 191. [Google Scholar] [CrossRef]
  12. Li, Z.; Qiu, X.; Yang, J.; Meng, D.; Huang, L.; Song, S. An Efficient BP Algorithm Based on TSU-ICSI Combined with GPU Parallel Computing. Remote Sens. 2023, 15, 5529. [Google Scholar] [CrossRef]
  13. Tan, Y.; Huang, H.; Lai, T. Real-Time Imaging Scheme of Short-Track GB-SAR Based on GPU+ OpenMP. IEEE Sens. J. 2024, 25, 4990–5002. [Google Scholar] [CrossRef]
  14. Gou, L.; Li, Y.; Zhu, D.; Wei, Y. A Real-time Algorithm for Circular Video SAR Imaging Based on GPU. Radar Sci. Technol. 2019, 17, 550–556. [Google Scholar]
  15. Ren, Z.; Zhu, D. A GPU-Based Two-Step Approach for High Resolution SAR Imaging. In Proceedings of the 2021 IEEE 6th International Conference on Signal and Image Processing (ICSIP), Nanjing, China, 22–24 October 2021; IEEE: New York, NY, USA, 2021; pp. 376–380. [Google Scholar]
  16. Ding, L.; Dong, Z.; He, H.; Zheng, Q. A Hybrid GPU and CPU Parallel Computing Method to Accelerate Millimeter-Wave Imaging. Electronics 2023, 12, 840. [Google Scholar] [CrossRef]
  17. Li, J.; Song, L.; Liu, C. The Cubic Trigonometric Automatic Interpolation Spline. IEEE/CAA J. Autom. Sin. 2017, 5, 1136–1141. [Google Scholar] [CrossRef]
  18. Liu, J.; Qiu, X.; Huang, L.; Ding, C. Curved-path SAR Geolocation Error Analysis Based on BP Algorithm. IEEE Access 2019, 7, 20337–20345. [Google Scholar] [CrossRef]
  19. Wang, G.; Qi, F.; Liu, Z.; Liu, C.; Xing, C.; Ning, W. Comparison between Back Projection Algorithm and Range Migration Algorithm in Terahertz Imaging. IEEE Access 2020, 8, 18772–18777. [Google Scholar] [CrossRef]
  20. Miao, X.; Shan, Y. SAR Target Recognition via Sparse Representation of Multi-view SAR Images with Correlation Analysis. J. Electromagn. Waves Appl. 2019, 33, 897–910. [Google Scholar] [CrossRef]
  21. Ding, L.; He, H.; Wang, T.; Chu, D. Cylindrical SAR Imaging Based on a Concentric-square-grid Interpolation Method. J. Electron. Inf. Technol. 2024, 46, 249–257. [Google Scholar]
  22. Soumekh, M. Reconnaissance with Slant Plane Circular SAR Imaging. IEEE Trans. Image Process. 1996, 5, 1252–1265. [Google Scholar] [CrossRef]
  23. Xueming, P.; Wen, H.; Yanping, W.; Kuoye, H.; Yirong, W. Downward Looking Linear Array 3D SAR Sparse Imaging with Wave-front Curvature Compensation. In Proceedings of the 2013 IEEE International Conference on Signal Processing, Communication and Computing (ICSPCC 2013), Kunming, China, 5–8 August 2013; IEEE: New York, NY, USA, 2013; pp. 1–4. [Google Scholar]
  24. Berland, F.; Fromenteze, T.; Decroze, C.; Kpre, E.L.; Boudesocque, D.; Pateloup, V.; Di Bin, P.; Aupetit-Berthelemot, C. Cylindrical MIMO-SAR Imaging and Associated 3-D Fourier Processing. IEEE Open J. Antennas Propag. 2021, 3, 196–205. [Google Scholar] [CrossRef]
  25. Xin, W.; Lu, Z.; Weihua, G.; Pcng, F. Active Millimeter-wave Near-field Cylindrical Scanning Three-dimensional Imaging System. In Proceedings of the 2018 International Conference on Microwave and Millimeter Wave Technology (ICMMT), Chengdu, China, 7–11 May 2018; IEEE: New York, NY, USA, 2018; pp. 1–3. [Google Scholar]
  26. Yang, T.; Xu, Q.; Meng, F.; Zhang, S. Distributed Real-time Image Processing of Formation Flying SAR Based on Embedded GPUs. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2022, 15, 6495–6505. [Google Scholar] [CrossRef]
  27. Tian, H.; Hua, W.; Gao, Y.; Sun, Z.; Cai, M.; Guo, Y. Research on Real-time Imaging Method of Airborne SAR Based on Embedded GPU. In Proceedings of the 2022 3rd China International SAR Symposium (CISS), Shanghai, China, 2–4 November 2022; IEEE: New York, NY, USA, 2022; pp. 1–4. [Google Scholar]
  28. Lin, A. Binary search algorithm. WikiJournal Sci. 2019, 2, 1–13. [Google Scholar] [CrossRef]
  29. Bajwa, M.S.; Agarwal, A.P.; Manchanda, S. Ternary search algorithm: Improvement of Binary Search. In Proceedings of the 2015 2nd International Conference on Computing for Sustainable Global Development (INDIACom), New Delhi, India, 11–13 March 2015; IEEE: New York, NY, USA, 2015; pp. 1723–1725. [Google Scholar]
Figure 1. Comparison of single-precision floating-point computing power between NVIDIA GPUs and Intel CPUs.
Figure 1. Comparison of single-precision floating-point computing power between NVIDIA GPUs and Intel CPUs.
Electronics 14 02073 g001
Figure 2. The model of the 3D cylindrical imaging system.
Figure 2. The model of the 3D cylindrical imaging system.
Electronics 14 02073 g002
Figure 3. A flowchart of the proposed GPU implementation of the 3D cylindrical RMA algorithm.
Figure 3. A flowchart of the proposed GPU implementation of the 3D cylindrical RMA algorithm.
Electronics 14 02073 g003
Figure 4. Wave-number distribution for a given height dimension in the CSG method. (a) Initial wave number; (b) interpolated wave number.
Figure 4. Wave-number distribution for a given height dimension in the CSG method. (a) Initial wave number; (b) interpolated wave number.
Electronics 14 02073 g004
Figure 5. Interval-search methods. (a) Linear search; (b) binary search.
Figure 5. Interval-search methods. (a) Linear search; (b) binary search.
Electronics 14 02073 g005
Figure 6. Time-complexity curves of linear search and binary search.
Figure 6. Time-complexity curves of linear search and binary search.
Electronics 14 02073 g006
Figure 7. Schematic of CUDA stream interpolation parallel processing.
Figure 7. Schematic of CUDA stream interpolation parallel processing.
Electronics 14 02073 g007
Figure 8. Imaging results comparison. (ac) are the front, top, and side views obtained by the CPU implementation; (df) are the front, top, and side views obtained by the proposed GPU implementation.
Figure 8. Imaging results comparison. (ac) are the front, top, and side views obtained by the CPU implementation; (df) are the front, top, and side views obtained by the proposed GPU implementation.
Electronics 14 02073 g008
Figure 9. Imaging profiles for each scenario. (a) azimuth-dimension profile; (b) distance-dimension profile; (c) height-dimension profile.
Figure 9. Imaging profiles for each scenario. (a) azimuth-dimension profile; (b) distance-dimension profile; (c) height-dimension profile.
Electronics 14 02073 g009
Figure 10. Imaging results of the actual experiment. (ac) are the front, top, and side views obtained by the CPU implementation; (df) are the front, top, and side views obtained by the proposed GPU implementation.
Figure 10. Imaging results of the actual experiment. (ac) are the front, top, and side views obtained by the CPU implementation; (df) are the front, top, and side views obtained by the proposed GPU implementation.
Electronics 14 02073 g010
Figure 11. Comparison of time consumption of each stage for three methods. (a) The FT stage; (b) the phase-compensation stage; (c) the 2D interpolation stage; (d) the total time; (e) the percentage of each stage of the proposed GPU.
Figure 11. Comparison of time consumption of each stage for three methods. (a) The FT stage; (b) the phase-compensation stage; (c) the 2D interpolation stage; (d) the total time; (e) the percentage of each stage of the proposed GPU.
Electronics 14 02073 g011
Figure 12. Acceleration ratios by different frequency-dimension points.
Figure 12. Acceleration ratios by different frequency-dimension points.
Electronics 14 02073 g012
Table 1. The parameters of the simulation experiment.
Table 1. The parameters of the simulation experiment.
ParameterParameter SymbolValue
Working frequency rangef30–35 GHz
Frequency-dimension points N 100
Angle-dimension points M 1440
Height-dimension points P 128
Frequency-dimension sampling interval Δ f 50 MHz
Angle-dimension sampling interval Δ θ 0.25 °
Height-dimension sampling interval Δ z 0.005 m
Table 2. Detailed implementation of three methods.
Table 2. Detailed implementation of three methods.
MethodFTPhase Compensation2D Interpolation
The CPU implementationFFTW/CSG method
The conventional GPU implementationcuFFTglobal memorytraversal method
The proposed GPU implementationcuFFTshared memoryCSG method + partitioning parallel processing
Table 3. Quality metrics comparison of the simulation experiment.
Table 3. Quality metrics comparison of the simulation experiment.
DimensionMetricsMethod
CPU ImplementationGPU Implementation
Azimuth dimensionPSLR (dB)−11.0104−11.0071
ISLR (dB)−10.9234−10.9199
IRW (m)0.16620.1663
Distance dimensionPSLR (dB)−12.3573−12.3564
ISLR (dB) −11.0849−11.0705
IRW (m)0.09680.0967
Height dimensionPSLR (dB)−11.0305−11.0273
ISLR (dB)−9.9241−9.9198
IRW (m)0.16660.1663
Table 4. Imaging quality metrics of the actual experiment.
Table 4. Imaging quality metrics of the actual experiment.
ValueFront ViewTop ViewSide View
PSNR (dB)37.9737.5538.21
SSIM0.9090.9220.914
Table 5. Time consumption of individual optimizations.
Table 5. Time consumption of individual optimizations.
Optimization TechniquesTime (s)
the proposed GPU implementation0.794
CSG method + binary search + CUDA stream0.821
Shared memory4.706
Shared memory + CSG method1.737
Shared memory + CSG method + binary search1.492
Shared memory + CSG method + CUDA stream0.902
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Cui, M.; Li, P.; Bu, Z.; Xun, M.; Ding, L. GPU-Optimized Implementation for Accelerating CSAR Imaging. Electronics 2025, 14, 2073. https://doi.org/10.3390/electronics14102073

AMA Style

Cui M, Li P, Bu Z, Xun M, Ding L. GPU-Optimized Implementation for Accelerating CSAR Imaging. Electronics. 2025; 14(10):2073. https://doi.org/10.3390/electronics14102073

Chicago/Turabian Style

Cui, Mengting, Ping Li, Zhaohui Bu, Meng Xun, and Li Ding. 2025. "GPU-Optimized Implementation for Accelerating CSAR Imaging" Electronics 14, no. 10: 2073. https://doi.org/10.3390/electronics14102073

APA Style

Cui, M., Li, P., Bu, Z., Xun, M., & Ding, L. (2025). GPU-Optimized Implementation for Accelerating CSAR Imaging. Electronics, 14(10), 2073. https://doi.org/10.3390/electronics14102073

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop