Next Article in Journal
Insight into the Characteristics and Triggers of Loess Landslides during the 2013 Heavy Rainfall Event in the Tianshui Area, China
Next Article in Special Issue
Enhanced Micro-Doppler Feature Extraction Using Adaptive Short-Time Kernel-Based Sparse Time-Frequency Distribution
Previous Article in Journal
Carotenoid Content Estimation in Tea Leaves Using Noisy Reflectance Data
Previous Article in Special Issue
Multi-Scale Feature Residual Feedback Network for Super-Resolution Reconstruction of the Vertical Structure of the Radar Echo
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Spectral De-Aliasing Method of Micro-Motion Signals Based on a Complex-Valued U-Net Network

Airforce Early Warning Academy, Wuhan 430019, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2023, 15(17), 4299; https://doi.org/10.3390/rs15174299
Submission received: 21 July 2023 / Revised: 29 August 2023 / Accepted: 30 August 2023 / Published: 31 August 2023
(This article belongs to the Special Issue Doppler Radar: Signal, Data and Applications)

Abstract

:
Spectrum aliasing occurs in signal echoes when the sampling frequency does not comply with the Nyquist Sampling Theorem. In this scenario, the extraction of micro-motion parameters becomes challenging. This paper proposes a spectral de-aliasing method for micro-motion signals based on a complex-valued U-Net network. Zero interpolation is employed to insert zeros into the echo, effectively increasing the sampling frequency. After zero interpolation, the micro-motion signal contains both real micro-motion signal frequency components and new frequency components. Short-Time Fourier Transform (STFT) is then applied to transform the zero-interpolated echo from the time domain to the time–frequency domain. Furthermore, a complex-valued U-Net training model is utilized to eliminate redundant frequency components generated by zero interpolation, thereby achieving the frequency reconstruction of micro-motion signal echoes. Finally, the training models are employed to process the measured data. The theoretical analysis, simulations, and experimental results demonstrate that this method is robust and feasible, and is capable of addressing the problem of micro-motion signal echo spectrum aliasing in narrowband radar.

1. Introduction

Micro-motion refers to the vibrations, rotations, precessions, oscillations, and other minor movements of a target or its structural components, which represent the target’s specific motion state details [1]. Micro-motion exists in various target motion forms, such as tank track rolling [2], random ship swinging due to wave fluctuations [3], ballistic missile precession or nutation in mid-flight [4,5], helicopter rotor rotation [6], human posture recognition [7,8], and satellite and space debris spinning [9,10], among others. Consequently, radar target recognition technology based on micro-motion features is considered one of the most promising approaches in the field. In 2000, V. C. Chen introduced the concepts of micro-motion and micro-Doppler (M-D) to the radar domain [1]. Since then, the use of micro-motion features has attracted significant attention from academia and industry [9,10,11,12]. The micro-Doppler effect can reveal the unique micro-motion characteristics and structural features of a target or its components. Assuming that the target’s micro-motion information can be accurately extracted, rapid detection and recognition of the target can be achieved, providing critical military applications for accurately engaging targets.
The critical challenge in micro-motion signal processing is the processing of time-varying signals. Research on target micro-motion feature analysis and extraction primarily involves several aspects. The first aspect is the time–frequency transform method [13], which mainly includes Short-Time Fourier Transform (STFT), Wigner–Ville Distribution (WVD), Continuous Wavelet Transform (CWT), the S-Method (SM), Short-Time Fractional Fourier Transform (STFRFT) [14], and others. The second aspect focuses on signal separation techniques, which predominantly consist of Hilbert–Huang Transform (HHT), Empirical Mode Decomposition (EMD), and their improved methods, such as Complex Empirical Mode Decomposition (CEMD) [15]. In addition to these traditional methods, the statistical histogram approach in paper [16] outperforms the L-statistics in paper [17] and provides a method for separating the echoes of rigid bodies and rotation points. To address the issue of group target echo signal separation, paper [18] achieves group target signal separation using independent component analysis technology. The third aspect involves image-based micro-motion feature extraction techniques. Methods based on Hough Transform and extended Hough Transform are explored in paper [19]. Although this method yields improved results, the calculations involved present numerous challenges. Paper [20] utilizes image rotation correlation obtained from different coherent processing intervals (CPIs) to estimate speed. However, analyzing and extracting the micro-motion features of radar echoes with spectrum aliasing remains difficult due to frequency component aliasing. Orthogonal Frequency Division Multiplexing (OFDM) is proposed in paper [21] to solve the problem of spectrum aliasing. This method necessitates designing the waveform, which is challenging to implement. Translational compensation is performed in paper [22], but this method requires frequency estimation and struggles to effectively compensate targets with complex micro-motion characteristics in order to resolve the issue of spectrum aliasing.
In this paper, the U-Net network is employed to eliminate redundant frequency components caused by zero interpolation [23]. As a machine learning approach [24], deep learning classifies and recurses according to input data. U-Net [25] is a deep learning algorithm used in this paper. The U-Net network model was proposed in 2015 and was initially used for semantic segmentation, which combines lightweight and high performance. In essence, the U-Net network is also a full convolutional neural network model, and is composed of an encoder and a decoder. Its name comes from its architecture shape: the model presents a “U” shape as a whole. It was proposed to solve the problem of semantic segmentation in medical images, but its development in the following years also confirmed that it is a versatile tool for semantic segmentation tasks. Consequently, U-Net shows great potential in applications such as semantic segmentation, and target segmentation and detection, particularly in image segmentation. Currently, various U-Net networks have been developed, including H-Dense U-Net [26], RIC-U-Net [27], and RCNN-U-Net [28].
Spectrum aliasing is common in radar signals [29,30]. The spectral de-aliasing techniques mainly rely on changes in sampling frequency to achieve spectrum de-aliasing. The frequency components are mainly fixed, that is, the frequency does not change over time, while micro-motion signals are generally time-modulated signals that change over time. Aimed at addressing the spectral aliasing problem in micro-motion signals, we propose a method that utilizes zero interpolation and complex-valued U-Net, which is used to resolve the micro-Doppler frequency ambiguity of a helicopter [23]. First, the influence of zero interpolation on the signal sequence is theoretically analyzed, and the modification of frequency components after zero interpolation in the signal sequence is examined. Since zero interpolation in the signal is equivalent to an increase in the sampling frequency, the signal contains real micro-Doppler frequency. However, this method also generates additional frequency components. Then, using the micro-motion model of the rotational signal as an example, the time–frequency results of the rotary micro-motion signal are obtained through STFT, which maintains consistent resolution in the entire time–frequency domain during signal analysis and exhibits strong integrity.
The U-Net is extended to the complex field for model training. The network input consists of the ambiguous time–frequency results of micro-motion signals after zero interpolation. The training model eliminates redundant frequency components in the time–frequency domain, compares the image segmentation effect with the fully convolutional network (FCN) [31] and the fully convolutional residual network (FCRN) [31] training models, and analyzes model performance using the test set. Finally, the experimental micro-motion signal generates time–frequency results after removing clutter through Moving Target Indicator (MTI) processing and STFT. The image segmentation employs a complex-valued U-Net with varying network parameters to remove redundant frequencies after zero interpolation. The performance of the networks is compared, and the influence of parameter changes on the network model’s performance is analyzed to determine the optimal network parameter configuration. Our theoretical analysis and simulation results demonstrate that this spectral de-aliasing method is robust and feasible, effectively addressing spectrum de-aliasing in micro-motion signals.

2. Related Work

2.1. Signal Model of Rotational Micro-Motion

Figure 1 is a diagram illustrating the radar receiving the echo of a rotating target. In Figure 1, a coordinate system with the radar as the origin is established. R is the distance between the radar O and the rotating target’s rotation center, L is the blade length of rotating target, d is the distance between the scattering points, f r o t is the rotating target’s rotation frequency, and α and β are the azimuth angle and the pitch angle between the radar O and the rotating target’s rotation center, respectively. According to [14,32], when transmitting a continuous wave with a wavelength of λ , the maximum micro-Doppler frequency of the rotating target will originate from the tip of the scattering point of the blades. When electromagnetic waves are detected vertically on the blades, the max Doppler frequency of the rotating target is as follows:
f d max = 4 π f r o t L λ cos β
Equation (1) shows that the max Doppler frequency is related to the transmitting wavelength, blade length, rotation speed, and pitch angle. Simultaneously, to ensure that the signal is not aliased, the pulse repetition frequency (PRF) must meet the Nyquist Sampling Theorem:
P R F 2 f d max = 8 π f r o t L λ cos β
Therefore, in the analysis of rotating targets, it is crucial to select an appropriate pulse repetition frequency to avoid the adverse effects of spectrum aliasing.

2.2. Analysis of Zero Interpolation

To address the spectrum aliasing of micro-motion signals, zero interpolation of the sampling signal is equivalent to increasing its sampling frequency [33]. However, new frequency components will appear, enabling the processed signal to satisfy the Nyquist Sampling Theorem and reveal the true frequency in spectral analysis, thereby providing a novel method for resolving spectrum aliasing. The method involves interpolating zeros between the signal samples in order to increase the sampling frequency integer multiple. The signal will add N new frequency components after interpolating N multiples of zeros. Therefore, the sampling frequency is equivalent to f s M = ( N + 1 ) f s , and Table 1 displays the frequency components.
Figure 2 presents the time–frequency results of the rotating target. Figure 2a illustrates the time–frequency results of spectrum aliasing when the sampling frequency does not meet the Nyquist Sampling Theorem. The time–frequency results of the signal in Figure 2b–d correspond to interpolating one zero, two zeros, and three zeros, respectively. These results illustrate that the spectrum distribution becomes more uniform and the true frequency values appear with the ambiguous spectrum components through zero interpolation. The comparison reveals that as the number of interpolated zeros increases, the real frequency components appear, and the frequency components also multiply. In this paper, U-Net is employed for image segmentation to eliminate redundant frequency components.

2.3. Experimental Data Acquisition and Echo Data Processing

To verify the feasibility of the model, an experimental simulation is conducted. We use a specific narrow-band continuous-wave micro-radar. The radar emits sawtooth and triangular waves. Figure 3a displays the radar. Its carrier frequency is 24 GHz, its bandwidth is 20 MHz, its AD sampling rate is 400 kHz, and its pulse repetition frequency is 4000 Hz. Compared to rotation, according to Equation (1), the frequency range of the vibration generally does not produce spectral aliasing. An angular scatterer is rotated to simulate the target motion. The radar receives the echo signal and collects the echo data. Figure 3b illustrates the experimental data acquisition process. The rotation of the angular reflector simulates the target’s rotation, and the radar receives the micro-motion echo signal of the angular reflector.
STFT [17] is a classic time–frequency analysis method, and Equation (3) defines STFT. In Equation (3), s t represents the receiving signal, and w t represents the window function. In this study, STFT with an appropriate window length achieves suitable time–frequency resolution, while sampling reduces the size of the time–frequency transformation result of the complex-valued matrix.
S T F T t , f = s τ w τ t exp j 2 π f τ d τ
Due to clutter in the data acquisition environment, echo signal processing is necessary. Moving Target Indicator (MTI) processing is used in this paper. Figure 4 shows MTI, an essential technology in the radar signal processing field.
Y ( z ) = X ( z ) X ( z 1 )
Equation (4) displays the transfer function of MTI. X ( z ) represents the result of z transformation of the receiving signal. It primarily uses the difference between the moving target echo and the clutter frequency domain to suppress clutter and retain the target echo by designing a filter with a specific notch. Figure 5 presents the time–frequency results after MTI processing. The time–frequency results indicate that MTI processing can suppress environmental clutter.

3. Method and Dataset

3.1. Design of the Complex-Valued U-Net

Based on the time-domain characteristics of micro-motion signals, this paper uses a U-Net in a complex field; the architecture is shown in Figure 6.
The network inputs are the time–frequency results in the complex field after zero interpolation. Every point in the instant frequency results has a real part and an imaginary part. For discrete echo sequences, STFT generates a 200 × 128 matrix time–frequency result, and the matrix is a complex-valued matrix.
To establish the experimental model, each value in the matrix in the time–frequency results is set as follows:
R i j = x i j + i y i j
In Equation (5), x i j represents the real-part value of the complex values in the i row and the j column of the matrix, and y i j represents the imaginary-part value of the complex values in the i row and the j column, i = 1 , 2 , 3 , , 200 , j = 1 , 2 , 3 , , 128 .
Firstly, the maximum value is determined using the following modulus:
R max = max { x i j 2 + y i j 2 }
Then, we normalize the matrix. The real part and the imaginary part are separated, resulting in the following real-part matrix:
R i j r e a l = x i j R max = x 1 , 1 R max x 1 , 128 R max x 200 , 1 R max x 200 , 128 R max
The imaginary-part matrix is as follows:
R i j i n a g = y i j R max = y 1 , 1 R max y 1 , 128 R max y 200 , 1 R max y 200 , 128 R max
Both matrices are substituted into the fully convolutional network for encoding and decoding operations. A superior network model is identified by adjusting the number of network layers and convolution kernel size. According to engineering practice, the convolutional kernels are set to 3 × 3 , 5 × 5 , and 7 × 7 , and the number of layers is set to 5, 7, and 9. Table 2 displays the network structure of the U-Net designed in this paper.
In Table 2, 3conv 3 × 3 ReLU represents three convolution operations, with a convolution kernel size of 3 × 3 and an activation function of ReLU; Max pool 2 × 2 denotes maximum pooling, with a step of 2 × 2 ; and Up-conv 3 × 3 signifies up-convolution, with a convolution kernel size of 3 × 3  Figure 7 illustrates the structure and operation flow.

3.2. The Training Model

In this paper, the time–frequency results of the micro-motion signal echo after zero interpolation serve as the dataset inputs, while the unambiguous time–frequency results of the micro-motion signal obtained by expanding the sampling frequency are the dataset labels. The U-Net in the complex field is trained to generate the model, which is evaluated and compared using the convolutional network FCN and FCRN models. A variety of performance tests are conducted on the designed U-Net to analyze the network performance. Subsequently, the number of layers and convolution kernel sizes of the complex-valued U-Net are changed to generate the training model. After comparing performance, the time–frequency results of the experimental data are tested, and we analyze the influence of the number of layers and the convolution kernel model to solve the problem of spectrum aliasing, and select the best one. Figure 8 demonstrates the training model process and Table 3 shows the training environment.

3.3. Dataset

As can be seen from Equation (2), many factors affect the rotating target echo signal, such as the number of rotating target blades, rotating speed, signal-to-noise ratio, etc. In this paper, the blade tip of the rotor target is equivalent to the scattering point to simulate the rotating target. Through the control variate method, the parameters are set to generate the dataset for training. The initial blade phase of the micro-motion signal is random. The range of the initial phase satisfies the phase Φ [ 0 , 2 π ] . The wavelength is λ = 0.3   m . We set the speed and blade length according to the highest frequency range. For the generation of a simulation dataset, to match with the experimental simulation data, this paper uses a single rotor and narrow-band radar echo model modeling; the rotation speed is set to 4 5   rad / s , the blade length is set to 0 . 36 0 . 46   m , the sampling frequency is 4000   Hz , and the time is 1 s. The STFT is carried out for the echo signal generated by the echo model after interpolating the zeros once. The time–frequency result of the first 0.32 s is taken as the input of the training set. The sampling signal is increased to 8000   Hz , with the other parameters remaining the same, and the time–frequency results generated by the Short-Time Fourier transform of the echo are taken as the labels of the training set. According to the task, 2000 sets of data are generated as training sets. The number of test sets is 10% of the training set. The parameter design of the test set is the same as that of the training set.

3.3.1. Dataset of Zero Interpolation

The dataset does not consider noise, comparing network performance in an ideal case. The dataset input consists of the time–frequency results after interpolating zero once, and Figure 9 shows the training labels.

3.3.2. Dataset under Different SNRs

For the actual micro-motion signal, due to the presence of noise, its spectra are not as clean after STFT. Therefore, a dataset is generated for the initial signals under different SNRs, with the SNR set to 0 10   dB . The dataset input consists of the time–frequency results with SNR = 5 . 584   dB , and Figure 10 shows the training label.

3.3.3. Dataset of Experimental Data

After clutter suppression via MTI processing, the echo of the experimental data is analyzed via STFT to generate the time–frequency results, as shown in Figure 11a. To generate the model for effective image segmentation, the same MTI processing is performed on the input dataset. The time–frequency results of the input are shown in Figure 11b. Compared with the time–frequency results of the micro-motion signal sequences, the intensity of the sinusoidal curves is not uniform, and there are sporadic flicker residues beside the sinusoidal curves.

4. Performance Analysis

4.1. Network Performance Analysis

This paper used the complex-valued U-Net compared with the FCN and the FCRN. Figure 12 shows the training loss curves of these networks. For the FCN and the FCRN, the loss is far greater than that of the complex-valued U-Net. With an increasing number of training epochs, the loss is not stable. The FCRN even shows an overfitting phenomenon as its loss suddenly increases rapidly, and then, after a few epochs, the loss is reduced. However, with increased training epochs, the loss of the complex-valued U-Net remains low, demonstrating excellent performance.
Inputting an image from the test set into the models from the networks, this paper compares the output results. The results show that the FCN and the FCRN exhibit poor micro-motion spectral de-aliasing performance, not only failing to remove the redundant flicker components but also causing the flicker in the initial time–frequency results to become blurred, making the extraction of useful information more challenging. The complex-valued U-Net retains the required flicker and envelope information and effectively removes redundant flicker. Figure 13 displays the outputs of the three network training models on the same image from the test set.
The complex-valued U-Net demonstrates strong image segmentation ability [34]. This paper compares two well-known target image quality indicators, the Peak Signal–Noise Ratio (PSNR) and Structural Similarity Measure (SSIM), to further analyze the model’s performance. The value range of SSIM is between 0 and 1. When two images are perfectly similar, the SSIM value is 1. The closer the SSIM value is to 1, the more similar the two images are. The PSNR value is typically expressed in dB. When the PSNR value is higher than 40 dB, the processed image is virtually indistinguishable from the original image. If it falls between 20 dB and 40 dB, it can be considered that, although the image exhibits distortion, it aligns with the processing objective. If it is below 20 dB, the image processing effect is deemed insufficient to meet the requirements.
Taking 10 samples from the test set, this paper compares the PSNR and SSIM of the three network models. Table 4 and Table 5 display the SSIM and the PSNR for the three networks, respectively. The comparison reveals that the average value of SSIM for the complex-valued U-Net is 0.9468, close to 1, and the average value of PSNR is 43.52 dB, greater than 40 dB, indicating that this model exhibits excellent image segmentation performance. In contrast, the other two networks perform poorly and are unable to complete the task of image segmentation.

4.2. Performance Analysis of the Complex-Valued U-Net

The performance of the complex-valued U-Net is analyzed in this paper using the micro-motion signal after interpolating zero once. The analysis includes two parts: analysis under different SNRs and analysis under different network parameters.

4.2.1. Performance Analysis under Different SNRs

For the complex-valued U-Net used in this paper, the training model is used to test and validate the test set while analyzing its performance under different SNRs. According to the SNR range of the training set, the test set employed in this analysis is separated by 1 dB, and time–frequency results with SNRs from 0 dB to 10 dB are selected for testing. Figure 14 displays input data with 0 dB, 5 dB, and 10 dB from the test set, and Figure 15 shows the output. The results demonstrate that with the increase in the signal-to-noise ratio, the flicker does not change significantly, but the envelope weakens as the signal-to-noise ratio decreases. Compared to the test set input, its envelope is still more pronounced; compared to the test set label, it exhibits some loss. The PSNR and SSIM under different SNR reduction results are provided in Table 6. Through comparison, it is discovered that PSNR and SSIM are significantly reduced in the presence of noise. The randomness of the noise leads to a difference between the input noise and the output noise when generating the dataset. Valuable information primarily resides in the sinusoidal signal. According to the model’s time–frequency results, the model can better retain the sinusoidal signal, and low image index parameters minimally impact subsequent parameter extraction. The performance of the complex-valued U-Net under different SNRs is superior, effectively addressing spectrum aliasing.

4.2.2. Performance Analysis under Different Network Parameters

In this paper, we altered the convolution kernel size and network layers, constructed nine networks, trained the network, generated the training model, used the test set to compare different network performances, and identified the network with the best performance.
The convolution kernel size and network layers directly influence the training results. Figure 16 depicts the loss curves of network training.
Taking the loss curves of networks A, D, and G as examples, the convolution kernel size for the three networks is 3 × 3, but the number of network layers varies, with 5, 7, and 9 layers, respectively. Examining the loss curves of networks A, B, and C, the number of layers for these three networks is the same, and the convolution kernel sizes are 3 × 3, 5 × 5, and 7 × 7, respectively. After comparing, it is found that the networks exhibit lower loss values when the number of network layers is 5, and network A has the largest loss value. This demonstrates the importance of selecting suitable parameters.
Fifty time–frequency results output from the training model are calculated. Figure 17 and Figure 18 show the SSIM and PSNR of the 50 time–frequency results from the test set, respectively. After comparison, it is evident that networks A–I perform differently. Except for network A, the other networks demonstrate good performance.

4.3. Experimental Results

MTI processing is applied to process the experimental data, and the results after STFT are used as inputs to compare the performance of the complex-valued U-Net under different network parameters. Figure 19 presents the output of each training model.
Figure 19 (raw) shows the time–frequency results without processing. Figure 19 (input) shows the time–frequency results of the input network models, and Figure 19a–i correspond to the output results of the models generated by networks A–I. Figure 20 presents the results of each training model after IRadon transform. Figure 20 (raw) shows the results without processing after IRadon transform. Figure 20 (input) shows the results of the input after IRadon transform and Figure 20a–i correspond to the results of networks A–I after IRadon transform.
We use IRadon transform and Hough transform to extract the speed features from the input and output results, and the speed extraction results are shown in in Table 7.
As shown in Figure 19, as the number of network layers and the size of the convolutional kernel increase, the model has a more significant effect on removing excess frequency components from the input time–frequency results. In the transformation results of Figure 20, the focus is also higher. The estimation results in Table 7 also indicate that the proposed method can solve the problem of spectrum aliasing in micro-motion signals and improve the accuracy of micro-motion feature extraction. Upon comparison, the set of network parameters significantly influences the experimental data in addressing the problem of spectrum aliasing. Due to the numerous errors in acquiring experimental data, the obtained data differ considerably from the theoretical analysis, directly affecting the model’s performance. For the measured data, as the number of network layers increases, the effect of removing redundant frequency components improves. This means that the SSIM and PSNR derived from the test set can only serve as references. Selecting the most suitable training parameters through the measured data, generating the training model, and removing redundant frequency components remain necessary.

5. Conclusions

Aiming to address the problem of spectrum aliasing caused by frequency components being higher than the sampling frequency in micro-motion signal echo signals, a complex-valued U-Net is employed to eliminate redundant frequency components generated via zero interpolation. Moreover, the performance of the complex-valued U-Net is compared with that of FCN and FCRN. Simultaneously, different SNR test sets and network parameters are utilized to investigate the performance of the complex-valued U-Net in segmenting micro-motion signals. The measured data are obtained through experiments, and the optimal network parameters are acquired by employing the models under different network parameters, proving that the image segmentation model can be applied in spectral de-aliasing. The spectrum aliasing of micro-motion signals can be effectively addressed using zero interpolation and the complex-valued U-Net. However, due to limitations in the experimental hardware, the amount of experimental data is relatively small, and the model’s performance could be further optimized. In the future, more types of micro-motion signals will be used to test the model’s performance. For the network, the next steps could be to further optimize the network structure, improve the model training dataset, and improve the model’s robustness and usage range.

Author Contributions

Writing—original draft preparation, M.L. (Ming Long); methodology, J.Y.; writing—review and editing, M.L. (Ming Long) and S.X.; data curation, M.L. (Mingjiu Lv), W.C. and B.C. All authors have read and agreed to the published version of the manuscript.

Funding

This work was funded by the National Natural Science Foundation of China under grant number 61671469 and the Hubei Natural Science Foundation Youth Program under grant number 2022CFB901.

Data Availability Statement

The data are openly available in a public repository.

Acknowledgments

The authors wish to extend their sincere thanks to the editors and anonymous reviewers for their careful reading and valuable comments to improve the quality of this paper. This work was supported in part by the Dual Key Project and the National Natural Science Foundation of China under grant number 61671469, and in part by the Hubei Natural Science Foundation Youth Program under grant number 2022CFB901.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Chen, V.C.; Li, F.; Ho, S.S.; Wechsler, H. Analysis of Micro-Doppler Signatures. IEE Proc. Radar Sonar Navig. 2003, 150, 271–276. [Google Scholar] [CrossRef]
  2. Huang, J.; LI, X.; Huang, X.; Jia, F. Micro-Doppler Features Based Parameter Estimation and Identification of Tank. J. Electron. Inf. Technol. 2010, 32, 1050–1055. [Google Scholar]
  3. Wang, X.; Zhang, S.; Zhao, H.; Zhu, L.; Chen, S. Analyze of Ship’s Micro-Doppler Characteristics Based on Hough Transform. In Proceedings of the 2020 IEEE 5th International Conference on Signal and Image Processing (ICSIP), Nanjing, China, 23–25 October 2020; pp. 644–648. [Google Scholar]
  4. Liu, L.; McLernon, D.; Ghogho, M.; Hu, W. Micro-Doppler Extraction from Ballistic Missile Radar Returns Using Time-Frequency Analysis. In Proceedings of the 2010 7th International Symposium on Wireless Communication Systems, York, UK, 19–22 September 2010; pp. 780–784. [Google Scholar]
  5. Tao, Z.; Yang, S. Target Signal Separation of Missile Group Based on Micro-Motion Frequency Correlation. J. Phys. Conf. Ser. 2020, 1607, 012074. [Google Scholar] [CrossRef]
  6. Clemente, C.; Soraghan, J.J. GNSS-Based Passive Bistatic Radar for Micro-Doppler Analysis of Helicopter Rotor Blades. Aerosp. Electron. Syst. IEEE Trans. 2014, 50, 491–500. [Google Scholar] [CrossRef]
  7. Lee, D.; Park, H.; Moon, T.; Kim, Y. Continual Learning of Micro-Doppler Signature-Based Human Activity Classification. IEEE Geosci. Remote Sens. Lett. 2021, 19, 3502405. [Google Scholar] [CrossRef]
  8. Kim, Y.; Ling, H. Human Activity Classification Based on Micro-Doppler Signatures Using a Support Vector Machine. IEEE Trans. Geosci. Remote Sens. 2009, 47, 1328–1337. [Google Scholar] [CrossRef]
  9. Smith, G.E.; Woodbridge, K.; Baker, C.J. Micro-Doppler Signature Classification. In Proceedings of the 2006 CIE International Conference on Radar, Shanghai, China, 16–19 October 2006. [Google Scholar]
  10. Xu, X.; Feng, C.; He, S. A Method for the Micro-Motion Signal Separation and Micro-Doppler Extraction for the Space Precession Target. IEEE Access 2020, 8, 130392–130404. [Google Scholar] [CrossRef]
  11. Chen, V.C.; Lipps, R.D. Time Frequency Signatures of Micro-Doppler Phenomenon for Feature Extraction. In Proceedings of the Wavelet Applications VII, Orlando, FL, USA, 26–28 April 2000. [Google Scholar]
  12. Lei, P.; Sun, J.; Wang, J.; Hong, W. Micromotion Parameter Estimation of Free Rigid Targets Based on Radar Micro-Doppler. IEEE Trans. Geosci. Remote Sens. 2012, 50, 3776–3786. [Google Scholar] [CrossRef]
  13. Shan, T.; Liu, S.; Tao, R.; Zhang, G. Experiment Demonstration of Micro-Doppler Detection of Rotor Blades with Passive Coherent Location Based on Digital Video Broadcast. J. Commun. Technol. Electron. 2014, 59, 1215–1224. [Google Scholar] [CrossRef]
  14. Pang, C.; Han, Y.; Hou, H.; Liu, S.; Zhang, N. Micro-Doppler Signal Time-Frequency Algorithm Based on STFRFT. Sensors 2016, 16, 1559. [Google Scholar] [CrossRef]
  15. Park, J.H.; Yang, W.Y.; Bae, J.W.; Kang, S.C.; Myung, N.H. Extended High Resolution Range Profile-Jet Engine Modulation Analysis with Signal Eccentricity. Prog. Electromagn. Res. 2013, 142, 505–521. [Google Scholar] [CrossRef]
  16. Rui, Z.; Gang, L.; Zhang, Y.D. Micro-Doppler Interference Removal via Histogram Analysis in Time-Frequency Domain. IEEE Trans. Aerosp. Electron. Syst. 2016, 52, 755–768. [Google Scholar] [CrossRef]
  17. Stankovic, L.; Thayaparan, T.; Dakovic, M.; Popovic-Bugarin, V. Micro-Doppler Removal in the Radar Imaging Analysis. IEEE Trans. Aerosp. Electron. Syst. 2013, 49, 1234–1250. [Google Scholar] [CrossRef]
  18. Guo, K.Y.; Li, Q.; Sheng, X.Q. A Precise Recognition Method of Missile Warhead and Decoy in Multi-Target Scene. J. Electromagn. Waves Appl. 2010, 24, 641–652. [Google Scholar] [CrossRef]
  19. Zhang, Q.; Yeo, T.S.; Tan, H.S.; Ying, L. Imaging of a Moving Target With Rotating Parts Based on the Hough Transform. IEEE Trans. Geosci. Remote Sens. 2008, 46, 291–299. [Google Scholar] [CrossRef]
  20. Yeh, C.; Zhou, W.; Lu, Y.; Yang, J. Non-Cooperative Target Imaging and Parameter Estimation with Narrowband Radar Echoes. Sensors 2016, 16, 125. [Google Scholar] [CrossRef]
  21. Lellouch, G.; Tran, P.; Pribic, R.; Genderen, P.V. OFDM Waveforms for Frequency Agility and Opportunities for Doppler Processing in Radar. In Proceedings of the 2008 IEEE Radar Conference (RADAR ’08), Rome, Italy, 26–30 May 2008. [Google Scholar]
  22. Chao, W.; Shuliang, W. Translational compensation of micro-motion with spectrum aliasing for narrow-band radar. Syst. Eng. Electron. 2010, 38, 2731–2737. [Google Scholar]
  23. Long, M.; Yang, J.; Xia, S.; Lv, M.; Wei, X.; Chen, W. A Micro-Doppler Frequency Ambiguity Resolution Method Based on Complex-Valued U-Net. Math. Probl. Eng. 2021, 2021, 2016816. [Google Scholar] [CrossRef]
  24. Schmidhuber, J. Deep Learning in Neural Networks: An Overview. Neural Netw. 2015, 61, 85–117. [Google Scholar] [CrossRef]
  25. Ronneberger, O.; Fischer, P.; Brox, T. U-Net: Convolutional Networks for Biomedical Image Segmentation. In Medical Image Computing and Computer-Assisted Intervention—MICCAI 2015, Proceedings of the 18th International Conference, Munich, Germany, 5–9 October 2015; Navab, N., Hornegger, J., Wells, W.M., Frangi, A.F., Eds.; Springer International Publishing: Cham, Switzerland, 2015; pp. 234–241. [Google Scholar]
  26. Li, X.; Chen, H.; Qi, X.; Dou, Q.; Fu, C.-W.; Heng, P.-A. H-DenseUNet: Hybrid Densely Connected UNet for Liver and Tumor Segmentation From CT Volumes. IEEE Trans. Med. Imaging 2018, 37, 2663–2674. [Google Scholar] [CrossRef]
  27. Zeng, Z.; Xie, W.; Zhang, Y.; Lu, Y. RIC-Unet: An Improved Neural Network Based on Unet for Nuclei Segmentation in Histology Images. IEEE Access 2019, 7, 21420–21428. [Google Scholar] [CrossRef]
  28. Yang, X.; Li, X.; Ye, Y.; Lau, R.Y.K.; Zhang, X.; Huang, X. Road Detection and Centerline Extraction Via Deep Recurrent Convolutional Neural Network U-Net. IEEE Trans. Geosci. Remote Sens. 2019, 57, 7209–7220. [Google Scholar] [CrossRef]
  29. Borys, A. The Problem of Aliasing and Folding Effects in Spectrum of Sampled Signals in View of Information Theory. Int. J. Electron. Telecommun. 2022, 68, 315–322. [Google Scholar] [CrossRef]
  30. Jana, S.; Rakshit, G.; Maitra, A. Aliasing Effect Due to Convective Rain in Doppler Spectrum Observed by Micro Rain Radar at a Tropical Location. Adv. Space Res. 2018, 62, 2443–2453. [Google Scholar] [CrossRef]
  31. Zhou, L.; Kong, X.; Gong, C.; Zhang, F.; Zhang, X. FC-RCCN: Fully Convolutional Residual Continuous CRF Network for Semantic Segmentation. Pattern Recognit. Lett. 2018, 130, 54–63. [Google Scholar] [CrossRef]
  32. Zhou, L.; Kong, X.; Gong, C.; Zhang, F.; Zhang, X. Micro-Doppler Extraction Based on Short-Time Iterative Adaptive Approach and Inverse Radon Transform. Acta Electonica Sin. 2016, 44, 505. [Google Scholar] [CrossRef]
  33. Ming, L.; Jianchao, Z.; Xu, W.; Jun, Y. Influence of zero interpolation of radar echo series on signal frequency domain characteristics. J. Air Force Early Warn. Acad. 2021, 35, 20–24+29. [Google Scholar]
  34. Tong, Y.B.; Zhang, Q.S.; Qi, Y.P. Image Quality Assessing by Combining PSNR with SSIM. J. Image Graph. 2006, 1758–1763. [Google Scholar]
Figure 1. Diagram illustrating radar receiving echo of rotating target.
Figure 1. Diagram illustrating radar receiving echo of rotating target.
Remotesensing 15 04299 g001
Figure 2. Time–frequency results of micro-motion signal after using zero interpolation. (a) The result when the sampling frequency does not meet the Nyquist Sampling Theorem. (b) The result after interpolating the zeros once. (c) The result after interpolating the zeros twice. (d) The result after interpolating the zeros three times.
Figure 2. Time–frequency results of micro-motion signal after using zero interpolation. (a) The result when the sampling frequency does not meet the Nyquist Sampling Theorem. (b) The result after interpolating the zeros once. (c) The result after interpolating the zeros twice. (d) The result after interpolating the zeros three times.
Remotesensing 15 04299 g002aRemotesensing 15 04299 g002b
Figure 3. Experimental configuration. (a) Radar. (b) Experimental operation.
Figure 3. Experimental configuration. (a) Radar. (b) Experimental operation.
Remotesensing 15 04299 g003
Figure 4. MTI processing.
Figure 4. MTI processing.
Remotesensing 15 04299 g004
Figure 5. Time–frequency results. (a) Raw signal. (b) Result after MTI processing.
Figure 5. Time–frequency results. (a) Raw signal. (b) Result after MTI processing.
Remotesensing 15 04299 g005
Figure 6. The U-Net.
Figure 6. The U-Net.
Remotesensing 15 04299 g006
Figure 7. Structure and operation flow of complex-valued U-Net-D.
Figure 7. Structure and operation flow of complex-valued U-Net-D.
Remotesensing 15 04299 g007
Figure 8. Training model flow.
Figure 8. Training model flow.
Remotesensing 15 04299 g008
Figure 9. Dataset of zero interpolation. (a) Input. (b) Label.
Figure 9. Dataset of zero interpolation. (a) Input. (b) Label.
Remotesensing 15 04299 g009
Figure 10. Dataset under different signal-to-noise ratios. (a) Input. (b) Label.
Figure 10. Dataset under different signal-to-noise ratios. (a) Input. (b) Label.
Remotesensing 15 04299 g010
Figure 11. Dataset of experimental data. (a) Input. (b) Label.
Figure 11. Dataset of experimental data. (a) Input. (b) Label.
Remotesensing 15 04299 g011
Figure 12. Training loss curves of the networks.
Figure 12. Training loss curves of the networks.
Remotesensing 15 04299 g012
Figure 13. Output of training models. (a) Test set. (b) Output of the FCN. (c) Output of the FCRN. (d) Output of the complex-valued U-Net.
Figure 13. Output of training models. (a) Test set. (b) Output of the FCN. (c) Output of the FCRN. (d) Output of the complex-valued U-Net.
Remotesensing 15 04299 g013
Figure 14. Test set under different SNRs. (a) SNR = 0   dB . (b) SNR = 5   dB . (c) SNR = 10   dB .
Figure 14. Test set under different SNRs. (a) SNR = 0   dB . (b) SNR = 5   dB . (c) SNR = 10   dB .
Remotesensing 15 04299 g014
Figure 15. Outputs under different SNRs. (a) SNR = 0   dB . (b) SNR = 5   dB . (c) SNR = 10   dB .
Figure 15. Outputs under different SNRs. (a) SNR = 0   dB . (b) SNR = 5   dB . (c) SNR = 10   dB .
Remotesensing 15 04299 g015aRemotesensing 15 04299 g015b
Figure 16. Loss curves of network training under different network parameters.
Figure 16. Loss curves of network training under different network parameters.
Remotesensing 15 04299 g016
Figure 17. SSIM of 50 time–frequency results.
Figure 17. SSIM of 50 time–frequency results.
Remotesensing 15 04299 g017
Figure 18. PSNR of 50 time–frequency results.
Figure 18. PSNR of 50 time–frequency results.
Remotesensing 15 04299 g018
Figure 19. Outputs of training models. (raw) Time–frequency results of the raw micro-motion signal. (input) Input. (a) Result of network A. (b) Result of network B. (c) Result of network C. (d) Result of network D. (e) Result of network E. (f) Result of network F. (g) Result of network G. (h) Result of network H. (i) Result of network I.
Figure 19. Outputs of training models. (raw) Time–frequency results of the raw micro-motion signal. (input) Input. (a) Result of network A. (b) Result of network B. (c) Result of network C. (d) Result of network D. (e) Result of network E. (f) Result of network F. (g) Result of network G. (h) Result of network H. (i) Result of network I.
Remotesensing 15 04299 g019aRemotesensing 15 04299 g019b
Figure 20. Results of outputs after IRadon transform. (raw) Result of the raw micro-motion signal. (input) Result of the input. (a) Result of network A. (b) Result of network B. (c) Result of network C. (d) Result of network D. (e) Result of network E. (f) Result of network F. (g) Result of network G. (h) Result of network H. (i) Result of network I.
Figure 20. Results of outputs after IRadon transform. (raw) Result of the raw micro-motion signal. (input) Result of the input. (a) Result of network A. (b) Result of network B. (c) Result of network C. (d) Result of network D. (e) Result of network E. (f) Result of network F. (g) Result of network G. (h) Result of network H. (i) Result of network I.
Remotesensing 15 04299 g020aRemotesensing 15 04299 g020b
Table 1. Frequency components after using zero interpolation.
Table 1. Frequency components after using zero interpolation.
Interpolating Zero Multiples N Including Frequency Component
N = 0 f 1
N = 1 f 1 , f 1 f s
N = 2 p f 1 , f 1 f s , f 1 + f s , , f 1 p f s , f 1 + p f s
N = 2 p + 1 f 1 , f 1 f s , f 1 + f s , , f 1 + p f s , f 1 ( p + 1 ) f s
p = 1 , , M , M = N 2 , denotes the rounding down symbol.
Table 2. U-Net parameter configuration.
Table 2. U-Net parameter configuration.
ABCDEFGHI
5 Weight
Layers
5 Weight
Layers
5 Weight
Layers
7 Weight
Layers
7 Weight
Layers
7 Weight
Layers
9 Weight
Layers
9 Weight
Layers
9 Weight
Layers
3Conv 3 × 3 , ReLU3Conv 5 × 5 , ReLU3Conv 7 × 7 , ReLU3Conv 3 × 3 , ReLU3Conv 5 × 5 , ReLU3Conv 7 × 7 , ReLU3Conv 3 × 3 , ReLU3Conv 5 × 5 , ReLU3Conv 7 × 7 , ReLU
Max pool 2 × 2
2Conv 3 × 3 , ReLU
Max pool 2 × 2
2Conv 5 × 5 , ReLU
Max pool 2 × 2
2Conv 7 × 7 , ReLU
Max pool 2 × 2
2Conv 3 × 3 , ReLU
Max pool 2 × 2
2Conv 5 × 5 , ReLU
Max pool 2 × 2
2Conv 7 × 7 , ReLU
Max pool 2 × 2
2Conv 3 × 3 , ReLU
Max pool 2 × 2
2Conv 5 × 5 , ReLU
Max pool 2 × 2
2Conv 7 × 7 , ReLU
Max pool 2 × 2
2Conv 3 × 3 , ReLU
Max pool 2 × 2
2Conv 5 × 5 , ReLU
Max pool 2 × 2
2Conv 7 × 7 , ReLU
Max pool 2 × 2
2Conv 3 × 3 , ReLU
Max pool 2 × 2
2Conv 5 × 5 , ReLU
Max pool 2 × 2
2Conv 7 × 7 , ReLU
Max pool 2 × 2
2Conv 3 × 3 , ReLU
Max pool 2 × 2
2Conv 5 × 5 , ReLU
Max pool 2 × 2
2Conv 7 × 7 , ReLU
Up-conv 2 × 2
Conv 3 × 3 , ReLU
Copy and Corp
2Conv 3 × 3 , ReLU
Up-conv 2 × 2
Conv 5 × 5 , ReLU
Copy and Corp
2Conv 5 × 5 , ReLU
Up-conv 2 × 2
Conv 7 × 7 , ReLU
Copy and Corp
2Conv 7 × 7 , ReLU
Max pool 2 × 2
2Conv 3 × 3 , ReLU
Max pool 2 × 2
2Conv 5 × 5 , ReLU
Max pool 2 × 2
2Conv 7 × 7 , ReLU
Max pool 2 × 2
2Conv 3 × 3 , ReLU
Max pool 2 × 2
2Conv 5 × 5 , ReLU
Max pool 2 × 2
2Conv 7 × 7 , ReLU
Up-conv 2 × 2
Conv 3 × 3 , ReLU
Copy and Corp
2Conv 3 × 3 , ReLU
Conv 1 × 1 , ReLU
Copy and Corp
Up-conv 2 × 2
Conv 5 × 5 , ReLU
Copy and Corp
2Conv 5 × 5 , ReLU
Conv 1 × 1 , ReLU
Copy and Corp
Up-conv 2 × 2
Conv 7 × 7 , ReLU
Copy and Corp
2Conv 7 × 7 , ReLU
Conv 1 × 1 , ReLU
Copy and Corp
Up-conv 2 × 2
Conv 3 × 3 , ReLU
Copy and Corp
2Conv 3 × 3 , ReLU
Up-conv 2 × 2
Conv 5 × 5 , ReLU
Copy and Corp
2Conv 5 × 5 , ReLU
Up-conv 2 × 2
Conv 7 × 7 , ReLU
Copy and Corp
2Conv 7 × 7 , ReLU
Max pool 2 × 2
2Conv 3 × 3 , ReLU
Max pool 2 × 2
2Conv 5 × 5 , ReLU
Max pool 2 × 2
2Conv 7 × 7 , ReLU
Up-conv 2 × 2
Conv 3 × 3 , ReLU
Copy and Corp
2Conv 3 × 3 , ReLU
Up-conv 2 × 2
Conv 5 × 5 , ReLU
Copy and Corp
2Conv 5 × 5 , ReLU
Up-conv 2 × 2
Conv 7 × 7 , ReLU
Copy and Corp
2Conv 7 × 7 , ReLU
Up-conv 2 × 2
Conv 3 × 3 , ReLU
Copy and Corp
2Conv 3 × 3 , ReLU
Up-conv 2 × 2
Conv 5 × 5 , ReLU
Copy and Corp
2Conv 5 × 5 , ReLU
Up-conv 2 × 2
Conv 7 × 7 , ReLU
Copy and Corp
2Conv 7 × 7 , ReLU
Up-conv 2 × 2
Conv 3 × 3 , ReLU
Copy and Corp
2Conv 3 × 3 , ReLU
Conv 1 × 1 , ReLU
Copy and Corp
Up-conv 2 × 2
Conv 5 × 5 , ReLU
Copy and Corp
2Conv 5 × 5 , ReLU
Conv 1 × 1 , ReLU
Copy and Corp
Up-conv 2 × 2
Conv 7 × 7 , ReLU
Copy and Corp
2Conv 7 × 7 , ReLU
Conv 1 × 1 , ReLU
Copy and Corp
Up-conv 2 × 2
Conv 3 × 3 , ReLU
Copy and Corp
2Conv 3 × 3 , ReLU
Up-conv 2 × 2
Conv 3 × 3 , ReLU
Copy and Corp
2Conv 5 × 5 , ReLU
Up-conv 2 × 2
Conv 3 × 3 , ReLU
Copy and Corp
2Conv 7 × 7 , ReLU
Up-conv 2 × 2
Conv 3 × 3 , ReLU
Copy and Corp
2Conv 3 × 3 , ReLU
Up-conv 2 × 2
Conv 5 × 5 , ReLU
Copy and Corp
2Conv 5 × 5 , ReLU
Up-conv 2 × 2
Conv 7 × 7 , ReLU
Copy and Corp
2Conv 7 × 7 , ReLU
Up-conv 2 × 2
Conv 3 × 3 , ReLU
Copy and Corp
2Conv 3 × 3 , ReLU
Conv 1 × 1 , ReLU
Copy and Corp
Up-conv 2 × 2
Conv 5 × 5 , ReLU
Copy and Corp
2Conv 5 × 5 , ReLU
Conv 1 × 1 , ReLU
Copy and Corp
Up-conv 2 × 2
Conv 7 × 7 ReLU
Copy and Corp
2Conv 7 × 7 , ReLU
Conv 1 × 1 , ReLU
Copy and Corp
Table 3. Configuration of deep learning platform in training.
Table 3. Configuration of deep learning platform in training.
ConfigurationContent
CPUAMD Ryzen 7 5700U
GPURTX 3050
Framework of deep learningTensorFlow 2.6.0
SystemWindows 10
Table 4. SSIM of the three networks.
Table 4. SSIM of the three networks.
FCNFRCNU-Net
10.7160.6590.947
20.7780.7370.944
30.6910.6360.957
40.7490.7040.945
50.7160.6590.956
60.7430.6910.941
70.7700.7280.956
80.7880.7590.940
90.7600.7120.930
100.7790.7400.952
avg0.7490.70250.9468
Table 5. PSNR of the three networks.
Table 5. PSNR of the three networks.
FCNFRCNU-Net
120.6818.9543.33
220.7519.2642.71
320.6419.0043.51
420.6519.0943.16
520.7119.0244.69
620.6919.0942.53
720.7019.2044.97
820.7319.5544.51
920.6819.1841.34
1020.7319.2344.46
avg20.7019.1643.52
Table 6. SSIM and PSNR of complex-valued U-Net.
Table 6. SSIM and PSNR of complex-valued U-Net.
0 dB1 dB2 dB3 dB4 dB5 dB6 dB7 dB8 dB9 dB10 dB
SSIM0.3220.3590.3020.4480.4550.4240.4430.4460.4150.4800.392
PSNR (dB)20.4821.0817.9522.8524.0224.4925.2326.3126.4827.5928.19
Table 7. The Speed ( rad / s ) extraction results.
Table 7. The Speed ( rad / s ) extraction results.
AlgorithmRawInputABCD
IRadon3.72.812.862.852.862.83
Hough3.962.192.472.482.473.96
AlgorithmEFGHI
IRadon2.842.832.862.872.86
Hough2.192.482.472.192.19
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Long, M.; Yang, J.; Xia, S.; Lv, M.; Cheng, B.; Chen, W. Spectral De-Aliasing Method of Micro-Motion Signals Based on a Complex-Valued U-Net Network. Remote Sens. 2023, 15, 4299. https://doi.org/10.3390/rs15174299

AMA Style

Long M, Yang J, Xia S, Lv M, Cheng B, Chen W. Spectral De-Aliasing Method of Micro-Motion Signals Based on a Complex-Valued U-Net Network. Remote Sensing. 2023; 15(17):4299. https://doi.org/10.3390/rs15174299

Chicago/Turabian Style

Long, Ming, Jun Yang, Saiqiang Xia, Mingjiu Lv, Bolin Cheng, and Wenfeng Chen. 2023. "Spectral De-Aliasing Method of Micro-Motion Signals Based on a Complex-Valued U-Net Network" Remote Sensing 15, no. 17: 4299. https://doi.org/10.3390/rs15174299

APA Style

Long, M., Yang, J., Xia, S., Lv, M., Cheng, B., & Chen, W. (2023). Spectral De-Aliasing Method of Micro-Motion Signals Based on a Complex-Valued U-Net Network. Remote Sensing, 15(17), 4299. https://doi.org/10.3390/rs15174299

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop