Next Article in Journal
Examining the Potential of Sentinel Imagery and Ensemble Algorithms for Estimating Aboveground Biomass in a Tropical Dry Forest
Previous Article in Journal
Editorial on the Advances, Innovations and Applications of UAV Technology for Remote Sensing
Previous Article in Special Issue
An SAR Image Automatic Target Recognition Method Based on the Scattering Parameter Gaussian Mixture Model
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Space Targets with Micro-Motion Classification Using Complex-Valued GAN and Kinematically Sifted Methods

Air and Missile Defense College, Air Force Engineering University, Xi’an 710038, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2023, 15(21), 5085; https://doi.org/10.3390/rs15215085
Submission received: 3 September 2023 / Revised: 30 September 2023 / Accepted: 9 October 2023 / Published: 24 October 2023

Abstract

:
Space target classification based on micro-motion characteristics has become a subject of great interest in the field of radar, particularly when using deep learning techniques. However, in practical applications, the ability of deep learning is hampered by the available radar datasets. As a result, obtaining a sufficient amount of the training dataset is a daunting challenge. To address this issue, this paper presents a novel framework for space target classification, consisting of three distinct modules: dataset generation, the kinematically sifted module, and classification. Initially, the micro-motion model of cone-shaped space targets is constructed to analyze target characteristics. Subsequently, the dataset generation module employs a complex-valued generative adversarial network (CV-GAN) to generate a large number of time-range maps. These maps serve as the foundation for training the subsequent modules. Next, the kinematically sifted module is introduced to eliminate images that do not align with the micro-motion characteristics of space targets. By filtering out incompatible images, the module ensures that only relevant and accurate dataset is utilized for further analysis. Finally, the classification model is constructed using complex-valued parallel blocks (CV-PB) to extract valuable information from the target. Experimental results validate the effectiveness of the proposed framework in space micro-motion target classification. The main contribution of the framework is to generate a sufficient amount of high-quality training data that conforms to motion characteristics, and to achieve accurate classification of space targets based on their micro-motion signatures. This breakthrough has significant implications for various applications in space target classification.

1. Introduction

1.1. Background and Motivation

Space targets fly in outer space and are only affected by gravity and drag force [1]. With the development of space technology, it is thus very important to carry out the measuring, tracking, and recognition of space targets [2,3,4].
The motion of space targets consists of two parts: high-speed translation and micro-motion [5]. As an inherent characteristic of the target, micro-motion can be chosen as a recognition basis [6]. Specifically, micro-motion can cause doppler modulation, commonly known as the micro-doppler (m-D) [7]. The typical micro-motions of space targets include wobbling of components, precession of warheads, tumbling of decoys, etc. The micro-motion characteristic and the m-D effect have attracted significant attention. Ali Hanif et al. present the recent application of micro-motion, including ground-moving target recognition, air target recognition, clutter rejection from wind turbines, space target recognition, and home automation [8]. Boyu Zhou et al. present a simulator for generating synthetic data for monitoring human activity and classification based on the m-D effect [9]. Xiaobin Liu et al. designed a method for dynamic measurement of precession targets in the anechoic chamber [10].
In recent years, through the wide use of deep learning (DL), micro-motion target classification has developed vigorously [11,12,13,14,15,16]. In [11], long short-term memory (LSTM) units were used to extract time-frequency spectrogram sequential features, and the recognition of six human bodies with micro-motion was realized. S. Z. Gurbuz et al. summarize the extensive application of DL in human micro-motion recognition [12]. Ref. [13] propose a continuous human motion signal processing method based on empirical mode decomposition (EMD). The multiscale squeeze-and-excitation network (MSENet), stacked bidirectional long short-term memory (BLSTM), and connectionist temporal classification (CTC) algorithm was used to designed the recognition network. Ref. [14] proposed a target-Aware Recurrent Attentional Network (TARAN), and the temporal dependence between range cells was used to realize the recognition of plane targets. Yizhe Wang et al. proposed an image preprocessing method, which can realize space micro-motion target recognition at a low signal-to-noise ratio (SNR) [15]. In [16], a rotation-invariant attention mechanism (RAM) module was added to the classification network design. Although the above methods have made some achievements in the field of space target classification, a central problem remains largely unsolved: space target is a typical non-cooperative target that lacks training datasets for network training. Recent studies have shown that a small dataset causes the problem of overfitting and limits the depth of DNNs that are implementable. Therefore, obtaining a sufficient amount of training data is a daunting task for space target classification.
Traditional recognition networks based on electromagnetic datasets fail to cover all micro-motion parameters and have single target size parameters, leading to clear limitations and deficiencies when recognizing space target radar images which are heavily influenced by micro-motion and geometric parameters. Generative adversarial networks (GAN) have recently been introduced for synthesizing credible radar images [17,18,19,20]. Ibrahim Alnujaim et al. proposed using GAN to increase the human micro-Doppler signatures training dataset [19]. In [20], auxiliary classifier generative adversarial networks (ACGAN) were used for the generation of radar images that are adapted to different environmental conditions. However, the complex-valued characteristics of radar images have not been considered. In fact, in the field of complex signal processing such as radar and sonar, the importance of complex-valued networks has been noticed [21,22,23]. Zhimian Zhang and colleagues introduced a complex-valued Convolutional Neural Network (CV-CNN) based on the complex backpropagation algorithm to utilize the amplitude and phase information of complex SAR imagery. The experiments demonstrated the effectiveness of the CNN in SAR image interpretation [21]. Jingkun Gao and colleagues introduced an enhanced radar imaging method using a CV-CNN. The CV-CNN was modified with the cReLU activation for radar imaging tasks [22]. A hybrid conditional random field model based on a complex-valued 3-D Convolutional Neural Network (CV 3-D CNN) was also proposed [23]. This model effectively exploits both the amplitude and phase information in PolSAR data from various polarimetric channels.
Table 1 provides a brief summary of the main attributes of the relevant methods and the proposed method.

1.2. Main Contributions

The above research results of the micro-motion target classification and GAN have inspired us. To establish a space target database that conforms to the kinematics principle and achieves a space target classification, we propose a framework that includes a dataset generation module, a kinematically sifted module, and a classification module.
To conclude, this paper has three main contributions:
(1)
This paper systematically explores the entire process of target classification based on micro-motion characteristics. The results of the simulation are utilized to show the effectiveness of the proposed framework.
(2)
To address the issue of limited training datasets, we design a GAN based on complex-valued residual dense blocks. This innovative approach allows us to generate synthetic radar echo data that can be used to augment the training dataset. The complex-valued residual dense block in the GAN enables it to capture and learn the intricate electromagnetic information present in radar echoes. This novel technique significantly enhances the performance of the classification model by providing more diverse and representative training data.
(3)
We propose a novel kinematically sifted method for target classification. This method is based on continuous curve detection and curve periodic judgment. By considering the physics characteristics of the generated space targets dataset, we are able to filter out irrelevant or misleading dataset, improving the accuracy of the classification results. The experiment results demonstrate the importance of incorporating the physical properties of the targets into the dataset-sifted process.
In summary, this paper contributes to the field of target classification by systematically exploring the use of micro-motion characteristics, designing a CV-GAN for data generation, and introducing a novel kinematically sifted method.

1.3. Organization

For the sake of readability, Table 2 summarized the variable notations used in this paper.
The remainder of this paper is organized as follows. Space targets’ micro-motion model and characteristics analysis are introduced in Section 2. The classification framework is depicted in Section 3. The simulation experiment verifies the accuracy of the proposed method in Section 4. Conclusions are drawn in the last section.

2. Space Target Micro-Motion Characteristics Analysis

Here, the representative micro-motions for space targets are considered: (1) Wobbling, see Figure 1a. (2) Tumbling, see Figure 1b. (3) Precession, see Figure 1c. and (4) nutation, see Figure 1d [24]. The mass of the cone-shaped space target is set at the origin of the o x y z coordinate system, the elevation angle of the target symmetry axis is θ , the angle between the radar line-of-sight (LOS) and the o z axis is β . The initial coning matrix can be expressed as:
I n i t = 1 0 0 0 cos θ sin θ 0 sin θ cos θ
Taking the wobbling motion as an example to analyze, assuming ω w represents the wobbling angular velocity, the micro-motion matrix at time t is
w ( t ) = I + U w sin ( ω w t ) + U w 2 [ 1 cos ( ω w t ) ]
where I is the identity matrix and ω w represents the wobble angular velocity. The skew-symmetric matrix U w can be defined by
U w = 0 ω w z ω w y ω w z 0 ω w x ω w y ω w x 0
where ω w x , ω w y , and ω w z are the projection of ω w onto the x , y , and z axis , respectively.
Then, the micro-range of scattering centers by wobbling are
r w ( t ) = w ( t ) · i n i t · r 0 · n
where r 0 is the initial position of scattering centers and n is the unit vector of LOS.
Similarly, at time t , the tumbling matrix can be expressed as
t ( t ) = I + U t sin ( ω t t ) + U t 2 [ 1 cos ( ω t t ) ]
where ω t represents the tumble angular velocity. U t represents the skew-symmetric matrix of tumbling angular velocity ω t .
Precession can be considered as the combination of spinning and coning motion, and nutation can be considered as the combination of spinning, coning, and wobbling [25].
Assume c and s represent the coning matrix and the spinning matrix, respectively. Then, the precession matrix and nutation matrix can be expressed as
p ( t ) = c s n ( t ) = w c s
Then, the micro-range of scattering centers by tumbling, precession, and nutation are
r t ( t ) = t ( t ) · i n i t · r 0 · n r p ( t ) = p ( t ) · i n i t · r 0 · n r n ( t ) = n ( t ) · i n i t · r 0 · n
In conclusion, the micro-range characteristics of the wobbling target conform to the law of sinusoidal modulation. If the wobbling angle exceeds π , the target will rotate in a specific direction around its center of mass, entering the tumbling state. Precession and nutation belong to compound micro-motions which can be regarded as a composite of multiple primary micro-motions. Consequently, the micro-range characteristics of the precession and nutation target will undergo additional modulation on a sinusoidal basis.
Given that the wideband radar emits a linear frequency-modulated (LFM) signal, consequently, the echo signal may be defined as follows
s w ( t ^ , t s ) = k = 1 K σ k rect t ^ T k t s τ · exp j 2 π f c t T k t s + μ 2 t ^ T k t s 2
where t ^ and t s represent the fast time and slow time, respectively. μ is the frequency modulation slope of LFM signal, f c denotes the wideband radar carrier frequency, τ is the time duration, and c represents the speed of light. σ k and T k t s = 2 R k t s / c denote the scattering coefficient and the echo delay time at t s of the kth scattering center, respectively.
After applying the “dechirp” processing and compensating for the residual video phase, a Fourier transform (FT) is performed on the echo in the fast time domain [26]. Subsequently, r = c / 2 μ f is substituted to obtain the expression for the echo in the domain of the time-range, which can be defined as follows:
S ( r , t s ) = k = 1 K σ k τ s i n c 2 B c r + R Δ k ( t s ) · exp j 4 π f c c R Δ k ( t s )
where B is the radar bandwidth, and r denotes the motion range. R Δ k ( t s ) = R k ( t s ) R r e f ; symbol R r e f is reference range. s i n c · represents the sinc function.
In simulation, the time-range maps can be obtained by Equation (9). Assuming the wideband radar carrier frequency f c is 10 GHz, the bandwidth B is 2 GHz, and the radar observation time is 2 s. The height and radius of the cone target are 3 m and 0.64 m, respectively.
Figure 2 presents simulation time-range maps of the four micro-motion forms including wobbling, tumbling, precession, and nutation. In wobbling state, the wobble angular velocity 2 π rad/s, the maximum angle of wobbling is 30 ° , and the simulation result is shown in Figure 2a. In the tumbling state, the tumble angular velocity ω t = 1.2 π rad/s, which is the simulation result, as shown in Figure 2b. The cone rotation angular velocity of precession state is 4 π rad/s, and the angle β is 12 ° . The simulation result of the precession as shown in Figure 2c. For nutation target, the cone rotation angular velocity, wobble angular velocity, β and maximum wobbling angle are 4 π rad/s, 2 π rad/s, 12 ° and 20 ° , respectively. The simulation result of nutation is shown in Figure 2a.
As shown in Figure 2, the time-range maps of different micro-motion forms have obvious periodicity, but their variation laws are different, which can be used as an important basis for distinguishing different targets.

3. Proposed Framework

Figure 3 is the schematic diagram of the proposed framework. The basic structure of our proposed framework includes a dataset generation module, a kinematically sifted module, and a classification module. The dataset generation module uses a complex-valued generative adversarial network (CV-GAN) to generator dataset. The module for the kinematic sifted module verifies the compliance of the generated images with the kinematic laws and eliminates any images that violate these laws. The classification module implements target recognition function based on time-range maps. Below are the implementation methods of the three modules.

3.1. Dataset Generation Module

GAN is commonly utilized for various tasks such as generating images, style conversion, and achieving image super-resolution capabilities. Unlike traditional networks, GAN consists of two connected parts, including the generator and the discriminator. The generator is responsible for creating new data samples from random noise. Its role is to generate different but highly similar samples compared to real samples. On the other hand, the discriminator is employed to differentiate between real and generated samples. The generator and the discriminator play a min-max game, and the final optimization goal is to reach Nash equilibrium [27]. The process can be defined as:
min G max D V G , D = E y P d a t a log D y + E z P z log 1 D G z
where y is the samples input, and z is the noise vector. D and G defined the generator and the discriminator, respectively. P d a t a and P z defined the distribution of real data and generated data, respectively.
In the objective function, the discriminator D manages to maximize its ability to distinguish between fake and real images. The generator G aims to minimize the JS divergence between the distribution of generated data and real data.
In this paper, this is the first time that the concept of complex values has been used in GAN construction. We use the complex-valued generative adversarial network (CV-GAN) to generate large time-range maps. Similarly, CV-GAN consists of two parts. The generator and discriminator in this paper were composed of complex-valued residual dense blocks (CV-RDB). The structure of CV-RDB is shown in Figure 4.
Figure 4 illustrates the architecture of the CV-RDB, with blue blocks representing the complex-valued convolution layer, gray blocks representing the complex-valued batch normalization (CV BN) layer, green blocks representing the complex-valued activation function, “C” denoting the concatenation operator, and “+” indicating the summation. CV-RDB can be considered as a combination of the residual block and dense block, which can fully extract features and widen the network.
Complex-valued convolution can be defined as
M W = M W M W = W R W I W I W R M R M I
where · and subscript “ R ” represents the real part of complex values, · and subscript “ I ” represents the imaginary part of the complex value. M = M R + i M I is a complex filter matrix, W = W R + i W I is a complex vector, and * denotes the convolution operation.
According to Liouville’s theory [28], bounded entire functions must be constants. In a real-valued network, the activation function is often required to be a differentiable function, but in the complex-valued field, a bounded complex activation function cannot be complexly differentiable at the same time. Hence, it is not possible to use the real-valued activation function in a complex-valued network. There is still no agreed consensus on the use of complex-valued activation functions. In this paper, we use complex-valued ReLU (CReLU) [29] as activation function, which can be defined as
C R e L U z = R e L U z + i R e L U z
where z , R e L U · represents the rectified linear unit. θ z is the phase of z . When θ z 0 , π 2 or θ z π , 3 π 2 , real and imaginary parts are either strictly positive or strictly negative at the same time.
Complex-valued networks generally rely upon the CV BN process to accelerate learning. The CV BN can be written as
x ˜ = C 1 2 x E x
where x is complex-valued data, x ˜ represents normalized complex-valued data, and C is a covariance matrix, which is given by:
C = C o v x , x C o v x , x C o v x , x C o v x , x
where C o v represents the covariance operation.
The generator and the discriminator in a dynamic process of continuous confrontation, and finally, CV-GAN can produce many fake images which are sufficiently close to real images. The optimization function of CV-GAN is defined as follows:
min G max D V D , G = E y P d a t a log D y + D y + E z P z log 1 D G z + G z

3.2. Kinematically Sifted Module

It is important to note that the phenomenology of radar data exhibits different properties from optical imagery. In GAN synthetic image conventional evaluation, the visual effect is often regarded as the main evaluation metric. Visual effect evaluation involves assessing the quality of generated images in terms of details, colors, textures, clarity, and other aspects through subjectively judging their visual impact. However, some of the authentic synthetic images may be visually similar, but kinematically impossible. Therefore, using visual effects as the sole evaluation metric for synthetic images may not be accurate.
In this paper, it is necessary to observe the targets continuously over one micro-motion periodic to obtain the periodic and continuous curve images of the targets. We proposed a generation data-sifted method based on physics and kinematics, which can be described as follows:
(1)
Preprocessing: The pixel-wise adaptive Wiener filter [15] was used to preprocess.
(2)
Continuous Curve Detection: Due to the constraints of the radar line of sight, the number of scattering points of cone-shaped targets should not exceed three, which is also reflected in the time-range map, indicating that the number of continuous curves should be no more than three. This paper uses the ShuffleNet [30] to recognize the number of scattering point curves, as shown in Figure 5. The basic unit of the ShuffleNet consists of a group convolution with channel shuffling, as depicted in Figure 5a. Specifically, the original group is divided into multiple subgroups, which are then recombined to form new groups and subsequently undergo group convolution. As illustrated in Figure 5b, the number of groups after channel shuffle is three. The main branch of the basic unit first performs channel-level signal convolution via a 1 × 1 group convolution to the feature map, followed by shuffling the obtained feature map with channel shuffling technology. Subsequently, a 3 × 3 convolution kernel is used to additionally learn the target feature map, and a 1 × 1 group convolution is used to learn the obtained feature map. Moreover, a direct connection structure is adopted in the residual block to guarantee that the quality of the features produced by the next layer of the network does not fall below the quality of the features produced by the previous layer.
To train the network, time-range maps with different numbers of scattering points were generated according to the motion characteristics of different micro-motion targets presented in Section 2. Four labels were constructed to represent images containing one, two, three, and more than three scattering point curves, respectively. The network is trained using the simulation database, and then the generated database is classified to remove images with labels indicating more than three scattering point curves, thus enabling a screening of the number of scattering point curves.
(3)
Curve periodic judgment. We choose t i as the initial time point. Time-range map sequence I i at t i as the reference sequence. Then, we calculate the value of the correlation coefficient between I i and other sequence S i , which can be described as:
s i m = i I i I ¯ S i S ¯ i I i I ¯ 2 i S i S ¯ 2
where I ¯ and S ¯ denote the average of I i and S i , respectively.

3.3. Classification Module

Based on the characteristics of the time-range maps, we use a series of complex-valued parallel block (CV-PB) structures to construct a complex-valued parallel network (CV-PNet). The overall architecture of the CV-PNet is shown in Figure 4.
Assuming the parallel blocks have N parallel branches, the parallel block contains 2 N 1 convolutional blocks. In Figure 6, where N = 4 , it contains 15 convolutional blocks. Stacking M parallel block and a complex-valued convolution layer yields the parallel network whose total depth, measured in terms of convolution layers, is M · 2 N 1 + 1 .
There are N different types of complex-valued convolutional layers in the parallel block, which are, respectively, 1 × 1 , 3 × 3 , , 2 N 1 × 2 N 1 . In order to reduce computational complexity and expand the network depth, the structural design of parallel blocks follows fractal thinking. First, the 1 × 1 complex-valued convolutional layer was used as the initial layer. Then, we connect two 1 × 1 complex-valued convolutional layers in series and combine them with 3 × 3 complex-valued convolutional layers to obtain the second framework. Then, concatenating two second layer frameworks with 5 × 5 complex-valued convolutional layers yields the third framework. This continues until two N 1 t h frameworks are concatenated and then combined with complex-valued 2 N 1 × 2 N 1 convolutional layers to obtain the N t h framework.
Parallel structures, as opposed to independent convolutional kernels that learn a single feature, have the ability to automatically extract diverse features and demonstrate an increased capacity for handling complex tasks. Different processing branches sample different convolutional kernels, enabling the module to capture features at different scales.

4. Simulation Results and Discussion

4.1. Electromagnetic Computation Dataset

It is technically very hard to obtain real radar echo data of the space cone-shaped target [15,31]. Therefore, to acquire a sufficiently large dataset to drive the synthetic radar image generation with CV-GAN, the electromagnetic computation data was used as the CV-GAN training dataset. The steps for obtaining the CV-GAN training dataset are as follows:
Step 1: Calculation of the radar cross-section (RCS) value for the space target, as shown in Figure 7. Begin by creating a simulation model of the space target. The geometric structure is illustrated in Figure 7a. Next, the simulation model is imported into the FEKO 2019 software, and the Physical Optics (PO) method is applied to calculate the RCS of the target. The parameters used for electromagnetic computation are provided in Table 3. Figure 7b demonstrates the RCS effect.
The RCS value of the target experiences a sudden change at pitch angles of approximately 79.2 and 180 . This phenomenon occurs due to the mirror reflection between the generatrix and the base of the cone target, resulting in a significant increase in RCS. Furthermore, due to the occlusion effect, the cone target comprises three scattering centers, and the visibility of these centers is dependent on the pitch angle.
Step 2: Micro-motion parameter setting. According to the RCS of the target, we can obtain the CV-GAN training dataset by setting different target motion parameters, an initial angle, and LOS. Table 4 presents the settings of the micro-motion parameters.
Step 3: Acquire the echo signal of the space target. Integrate the expressions for the target micro-range under different micro-motions and place the echo expression into the time-range domain in order to generate the time-range maps.
Finally, there are 3840 samples in this EM computation dataset.

4.2. Dataset Generation and Kinematically Sifted Module

The structure of CV-RDB is shown in Table 5. For a simple presentation, the CV BN layer and CReLU layer are not represented.
The network structure of CV-GAN is shown in Table 6.
In this paper, the adaptive moment estimation (Adam) algorithm is employed as the CV-GAN optimizer method, with an exponential decay of 0.5 for the first-order momentum and 0.999 for the second-order momentum. The learning rate is 0.0002. The size of complex-valued convolutional layers in CV-RDB is 3 × 3 , and the step size is 1. The CV BN with the momentum of 0.8.
In this paper, the CV-GAN experiments were implemented using PyCharm, while the kinematically sifted experiments were implemented using MATLAB. The configuration details of the PC platform were as follows: CPU Intel(R) Xeon(R) Gold [email protected] Ghz and RAM 256 GB, and the GPU was an NVIDIA Quadro GV100.
Each training image in CV-GAN is scaled to 128 × 128 × 3 . The minibatch size is 16, and the noise vector of the generator is 128 × 1 (dimensionality of the latent space). Each time the number of channels in a CV-RDB layer is halved, the length and width are doubled.
A sample of six time-range maps generated by GAN [19], ACGAN [20], and CV-GAN are depicted in Figure 8. It can be observed that the GAN-generated images appear to be blurry, the ACGAN-generated images are comparatively clear, and the images generated by CV-GAN have the highest imaging quality.
In our evaluation of the quality of generated images, we employed Peak Signal to Noise Ratio (PSNR), an objective evaluation indicator for evaluating noise levels or image distortion that plays a crucial role in ensuring the quality of generated images. PSNR can be defined as:
P S N R = 10 log 10 M A X I 2 M S E
where M A X I defined the maximum value of image color and M S E is the error of the mean square.
The larger the PSNR, the less distortion and the better the quality of the generated image. In Figure 9, we observe that the images generated using CV-GAN exhibit significantly larger PSNR values, indicating superior image quality compared to other methods.
To quantitatively assess the performance of different methods, Fréchet inception distance (FID) is employed to measure the quality of generated data [32]. FID is mathematically formulated as
FID ( P e , P s ) = μ e μ s 2 2 + Tr e + s 2 e s 1 / 2
where P r and P s denote the distribution of the electromagnetic computation image and synthetic image, respectively. The value of FID is dependent on means μ e , μ s , and covariances e , s . Tr represents the trace of the matrix.
The FID of the time-range maps generated by different methods was calculated. As shown in Figure 10, the generated images based on CV-GAN have a lower FID value, indicating an improved diversity of the images.
To demonstrate the kinematic sifting method presented in this paper, we randomly selected 16 synthetic images generated by CV-GAN. As illustrated in Figure 11, the green label indicates that the generated image adheres to the motion law, while the red label indicates that the generated image is unsuccessful.
Upon examining this random selection of 16 synthetic images, we discovered that three images did not adhere to the kinematic laws and should be excluded from the generated database. This process was repeated to filter all generated data. As a result, 1273 images of wobbling, 1360 images of tumbling, 1185 images of precession, and 977 images of nutation that did not meet the criteria were eliminated.

4.3. Classification Result

The structure and hyperparameters of the CV-PNet are shown in Table 7.
The number of parallel branches is 4 and the size of complex-valued convolutional layers in CV-PB are 1 × 1 , 3 × 3 , 5 × 5 , and 7 × 7 . The step size of all complex-valued convolutional layers is 1.
The performance of the network can be assessed using the classification accuracy, F1 [33], and kappa coefficient [16], which are all values between 0 and 1, with higher values indicating better performance. The definition of the evaluation metric is as follows
ACC = TP + TN TP + TN + FP + FN
MAP = 1 Y i = 1 Y T P i T P i + F P i
MF 1 = 1 Y i = 1 Y 2 × T P i T P i + F P i × T P i T P i + F N i T P i T P i + F P i + T P i T P i + F N i
KC = ACC p e 1 p e
where Y is the number of categories. The abbreviations are as follows: TP (true positive), TN (true negative), FN (false negative), and FP (false positive). p e is the relative misclassification number. Assuming that the total number of samples is N , then
p e = T P + F N × T P + F P + F N + T N T N + F P N 2
We set the electromagnetic computation dataset as Data1, the generated dataset as Data2, and the generated dataset after sifting as Data3. To ensure the preciseness of the experiment, we set the database size of each target micro-motion type in Data2 and Data3 to 8640. The following experiments were designed to verify the effectiveness of the proposed method:
(1)
Experiment 1: We used Data1 as the network training database, with 20% of Data2 allocated as the classification validation database and 20% as the classification testing database.
(2)
Experiment 2: We used Data1 and 60% of Data2 as the classification training database. The settings of the validation database and testing database are kept consistent with Experiment 1.
(3)
Experiment 3: We used Data1 and 60% of Data3 as the classification training database. The settings of the validation database and testing database are kept consistent with Experiment 1.
The test confusion matrixes are depicted in Figure 12.
Then, we compute the value of the evaluation metric based on the confusion matrix. The confusion matrix of experiment 1, experiment 2 and experiment 3 is shown in Figure 12a–c. Figure 13 illustrates that the four indicators of the experiment 3 are the best and the experiment 2 is slightly inferior, whereas experiment 1 is poor. The worst recognition performance of experiment 1 was attributed to the small number of samples in the electromagnetic dataset, while the best performance accuracy of experiment 3 verified the importance of kinematically sifted experiments.
To further verify the effectiveness of the method, we compared the proposed networks with the following state-of-the-art target recognition algorithms. (1) TARAN [14]; (2) 2D CNN [15]; (3) 19-layer CNN [19].
Table 8 shows that under the same dataset, the proposed networks achieve a better performance in terms of ACC, MAP, MF1, and KC.
Then, based on the initial network structure, we studied the influence of the number of parallel blocks (M) and the number of parallel branches (N) in the proposed CV-PNet. The Data3 dataset for each class was randomly divided into 60% for training, 20% for validation, and 20% for testing. The comparison results are illustrated in Table 9 and Table 10.
As shown in Table 9, the number of parallel blocks (M) is increased from 4 to 12 in an interval of 2. As the value of M increased, the evaluation metric values gradually increased. It is noteworthy that the recognition effect of CV-PNet significantly improved when M was between 4 and 10. However, beyond M = 10, the recognition effect showed a slower improvement. One can observe from Table 10 that as the number of parallel branches increases, so does the recognition ability of CV-PNet.

5. Conclusions

This research introduces a comprehensive framework for classifying space micro-motion targets. The framework is composed of three key modules: a dataset generation module that utilizes CV-GAN to produce high-quality time-range maps, a kinematically sifted module that ensures compliance with kinematics laws, and a classification module designed to accurately recognize space micro-motion targets. Experimental results have unequivocally demonstrated the effectiveness of this proposed framework.
In future studies, we plan to expand the database generation method to include the time-frequency domain and range-doppler domain. Additionally, we will explore the impact of various radar images on the classification performance, aiming to further enhance the classification accuracy. These advancements will contribute to a more versatile and reliable classification framework for space micro-motion targets.

Author Contributions

Conceptualization, C.F.; methodology, L.H.; software, L.H.; validation, C.F. and L.H.; formal analysis, L.H.; investigation, X.H.; resources, X.H.; data curation, X.H.; writing—original draft preparation, L.H.; visualization, L.H.; supervision, C.F.; project administration, C.F.; funding acquisition, C.F. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the Natural Science Basic Research Plan in Shaanxi Province of China (No. 2023-JC-YB-491) and The Youth Innovation Team of Shaanxi Universities.

Data Availability Statement

Not applicable.

Acknowledgments

The authors would like to thank the editor and the reviewers for their helpful comments to improve this paper.

Conflicts of Interest

No conflict of interest exist in the submission of this manuscript, and the manuscript is approved by all authors for publication. I would like to declare on behalf of my co-authors that the work described is original research that has not been published previously and is not under consideration for publication elsewhere, in whole or in part. All the authors listed have approved the manuscript that is enclosed.

References

  1. Choi, I.O.; Park, S.H.; Kim, M.; Kang, K.B.; Kim, K.T. Efficient Discrimination of Ballistic Targets with Micromotions. IEEE Trans. Aerosp. Electron. Syst. 2020, 56, 1243–1261. [Google Scholar] [CrossRef]
  2. Xie, P.F.; Zhang, L.; Du, C.; Wang, X.Q.; Zhong, W.J. Space Target Attitude Estimation From ISAR Image Sequences With Key Point Extraction Network. IEEE Signal Process. Lett. 2021, 28, 1041–1045. [Google Scholar] [CrossRef]
  3. Tian, X.D.; Bai, X.R.; Xue, R.H.; Qin, R.Y.; Zhou, F. Fusion Recognition of Space Targets with Micromotion. IEEE Trans. Aerosp. Electron. Syst. 2022, 58, 3116–3125. [Google Scholar] [CrossRef]
  4. Kang, L.; Liang, B.S.; Luo, Y.; Zhang, Q. Sparse Imaging for Spinning Space Targets with Short Time Observation. IEEE Sens. J. 2021, 21, 9090–9098. [Google Scholar] [CrossRef]
  5. Zhang, W.P.; Li, K.L.; Jiang, W.D. Micro-motion frequency estimation of radar targets with complicated translations. Aeu-Int. J. Electron. Commun. 2015, 69, 903–914. [Google Scholar] [CrossRef]
  6. Gong, J.K.; Yan, J.; Li, D.R.; Chen, R.Z.; Tian, F.Y.; Yan, Z. Theoretical and Experimental Analysis of Radar Micro-Doppler Signature Modulated by Rotating Blades of Drones. IEEE Antennas Wirel. Propag. Lett. 2020, 19, 1659–1663. [Google Scholar] [CrossRef]
  7. Qiao, X.S.; Amin, M.G.; Shan, T.; Zeng, Z.X.; Tao, R. Human Activity Classification Based on Micro-Doppler Signatures Separation. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5105014. [Google Scholar] [CrossRef]
  8. Hanif, A.; Muaz, M.; Hasan, A.; Adeel, M. Micro-Doppler Based Target Recognition with Radars: A Review. IEEE Sens. J. 2022, 22, 2948–2961. [Google Scholar] [CrossRef]
  9. Zhou, B.; Lin, Y.; Le Kernec, J.; Yang, S.; Fioranelli, F.; Romain, O.; Zhao, Z. Simulation framework for activity recognition and benchmarking in different radar geometries. IET Radar Sonar Navig. 2021, 15, 390–401. [Google Scholar] [CrossRef]
  10. Liu, X.; Zhao, F.; Xu, Z.; Ai, X.; Xiao, S. Design and experimental research of precession target micro-Doppler measurement using pulse signal in anechoic chamber. Measurement 2021, 184, 109948. [Google Scholar] [CrossRef]
  11. Wang, M.Y.; Zhang, Y.M.D.; Cui, G.L. Human motion recognition exploiting radar with stacked recurrent neural network. Digit. Signal Prog. 2019, 87, 125–131. [Google Scholar] [CrossRef]
  12. Gurbuz, S.Z.; Amin, M.G. Radar-Based Human-Motion Recognition with Deep Learning Promising applications for indoor monitoring. IEEE Signal Process. Mag. 2019, 36, 16–28. [Google Scholar] [CrossRef]
  13. Zhao, R.N.; Ma, X.L.; Liu, X.H.; Li, F.M. Continuous Human Motion Recognition Using Micro-Doppler Signatures in the Scenario with Micro Motion Interference. IEEE Sens. J. 2021, 21, 5022–5034. [Google Scholar] [CrossRef]
  14. Xu, B.; Chen, B.; Wan, J.; Liu, H.; Jin, L. Target-Aware Recurrent Attentional Network for Radar HRRP Target Recognition. Signal Process. 2019, 155, 268–280. [Google Scholar] [CrossRef]
  15. Wang, Y.Z.; Feng, C.Q.; Hu, X.W.; Zhang, Y.S. Classification of Space Micromotion Targets with Similar Shapes at Low SNR. IEEE Geosci. Remote Sens. Lett. 2022, 19, 3504305. [Google Scholar] [CrossRef]
  16. Zhang, Y.; Yuan, H.X.; Li, H.B.; Chen, J.Y.; Niu, M.Q. Meta-Learner-Based Stacking Network on Space Target Recognition for ISAR Images. IEEE J. Sel. Top. Appl. Earth Observ. Remote Sens. 2021, 14, 12132–12148. [Google Scholar] [CrossRef]
  17. Huang, Y.; Wen, Y.; Shi, L.C.; Ding, X.H. Recognition-Aware HRRP Generation with Generative Adversarial Network. IEEE Geosci. Remote Sens. Lett. 2022, 19, 3503705. [Google Scholar] [CrossRef]
  18. Lang, Y.; Hou, C.P.; Ji, H.R.; Yang, Y. A Dual Generation Adversarial Network for Human Motion Detection Using Micro-Doppler Signatures. IEEE Sens. J. 2021, 21, 17995–18003. [Google Scholar] [CrossRef]
  19. Alnujaim, I.; Oh, D.; Kim, Y. Generative Adversarial Networks for Classification of Micro-Doppler Signatures of Human Activity. IEEE Geosci. Remote Sens. Lett. 2020, 17, 396–400. [Google Scholar] [CrossRef]
  20. Erol, B.; Gurbuz, S.Z.; Amin, M.G. Motion Classification Using Kinematically Sifted ACGAN-Synthesized Radar Micro-Doppler Signatures. IEEE Trans. Aerosp. Electron. Syst. 2020, 56, 3197–3213. [Google Scholar] [CrossRef]
  21. Zhang, Z.M.; Wang, H.P.; Xu, F.; Jin, Y.Q. Complex-Valued Convolutional Neural Network and Its Application in Polarimetric SAR Image Classification. IEEE Trans. Geosci. Remote Sens. 2017, 55, 7177–7188. [Google Scholar] [CrossRef]
  22. Gao, J.K.; Deng, B.; Qin, Y.L.; Wang, H.Q.; Li, X. Enhanced Radar Imaging Using a Complex-Valued Convolutional Neural Network. IEEE Geosci. Remote Sens. Lett. 2019, 16, 35–39. [Google Scholar] [CrossRef]
  23. Zhang, P.; Tan, X.F.; Li, B.B.; Jiang, Y.Y.; Song, W.Y.; Li, M.; Wu, Y. PolSAR Image Classification Using Hybrid Conditional Random Fields Model Based on Complex-Valued 3-D CNN. IEEE Trans. Aerosp. Electron. Syst. 2021, 57, 1713–1730. [Google Scholar] [CrossRef]
  24. Choi, I.O.; Park, S.H.; Kang, K.B.; Lee, S.H.; Kim, K.T. Efficient Parameter Estimation for Cone-Shaped Target Based on Distributed Radar Networks. IEEE Sens. J. 2019, 19, 9736–9747. [Google Scholar] [CrossRef]
  25. Xu, X.G.; Feng, C.Q.; Han, L.X. Classification of Radar Targets with Micro-Motion Based on RCS Sequences Encoding and Convolutional Neural Network. Remote Sens. 2022, 14, 5863. [Google Scholar] [CrossRef]
  26. Zhuo, Z.Y.; Zhou, Y.; Du, L.; Ren, K.; Li, Y. A Noise Robust Micro-Range Estimation Method for Precession Cone-Shaped Targets. Remote Sens. 2021, 13, 1820. [Google Scholar] [CrossRef]
  27. Lucas, A.; Lopez-Tapia, S.; Molina, R.; Katsaggelos, A.K. Generative Adversarial Networks and Perceptual Losses for Video Super-Resolution. IEEE Trans. Image Process. 2019, 28, 3312–3327. [Google Scholar] [CrossRef]
  28. Lee, C.; Hasegawa, H.; Gao, S.C. Complex-Valued Neural Networks: A Comprehensive Survey. IEEE-CAA J. Autom. Sin. 2022, 9, 1406–1426. [Google Scholar] [CrossRef]
  29. Su, H.N.; Bao, Q.L.; Chen, Z.P. Parameter Estimation Processor for Chirp Signals Based on a Complex-Valued Deep Neural Network. IEEE Access 2019, 7, 176278–176290. [Google Scholar] [CrossRef]
  30. Rezaee, K.; Mousavirad, S.J.; Khosravi, M.R.; Moghimi, M.K.; Heidari, M. An Autonomous UAV-Assisted Distance-Aware Crowd Sensing Platform Using Deep ShuffleNet Transfer Learning. IEEE Trans. Intell. Transp. Syst. 2022, 23, 9404–9413. [Google Scholar] [CrossRef]
  31. Zhang, Y.P.; Zhang, Q.; Kang, L.; Luo, Y.; Zhang, L. End-to-End Recognition of Similar Space Cone-Cylinder Targets Based on Complex-Valued Coordinate Attention Networks. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5106214. [Google Scholar] [CrossRef]
  32. Qu, L.L.; Wang, Y.T.; Yang, T.H.; Sun, Y.P. Human Activity Recognition Based on WRGAN-GP-Synthesized Micro-Doppler Spectrograms. IEEE Sens. J. 2022, 22, 8960–8973. [Google Scholar] [CrossRef]
  33. Dong, Y.F.; Liu, J.L.; Yan, W.J. Dynamic Hand Gesture Recognition Based on Signals From Specialized Data Glove and Deep Learning Algorithms. IEEE Trans. Instrum. Meas. 2021, 70, 2509014. [Google Scholar] [CrossRef]
Figure 1. Cone target with micro-motion.
Figure 1. Cone target with micro-motion.
Remotesensing 15 05085 g001aRemotesensing 15 05085 g001b
Figure 2. Time-range maps of the four micro-motion forms.
Figure 2. Time-range maps of the four micro-motion forms.
Remotesensing 15 05085 g002
Figure 3. Schematic diagram of the proposed framework.
Figure 3. Schematic diagram of the proposed framework.
Remotesensing 15 05085 g003
Figure 4. The structure of complex-valued residual dense block.
Figure 4. The structure of complex-valued residual dense block.
Remotesensing 15 05085 g004
Figure 5. The structure of the ShuffleNet. (a) The basic unit of the ShuffleNet. (b) Channel shuffle.
Figure 5. The structure of the ShuffleNet. (a) The basic unit of the ShuffleNet. (b) Channel shuffle.
Remotesensing 15 05085 g005
Figure 6. Complex-valued parallel network.
Figure 6. Complex-valued parallel network.
Remotesensing 15 05085 g006
Figure 7. Target electromagnetic scattering data. (a) The geometric structure of cone target. (b) Effect pictures of RCS.
Figure 7. Target electromagnetic scattering data. (a) The geometric structure of cone target. (b) Effect pictures of RCS.
Remotesensing 15 05085 g007
Figure 8. Randomly chosen images generated by different methods.
Figure 8. Randomly chosen images generated by different methods.
Remotesensing 15 05085 g008
Figure 9. PSNR of the generated data obtained by different method.
Figure 9. PSNR of the generated data obtained by different method.
Remotesensing 15 05085 g009
Figure 10. FID of the generated data obtained by different method.
Figure 10. FID of the generated data obtained by different method.
Remotesensing 15 05085 g010
Figure 11. Kinematically sifted time-range maps generated by CV-GAN.
Figure 11. Kinematically sifted time-range maps generated by CV-GAN.
Remotesensing 15 05085 g011
Figure 12. Classification results. (a) The confusion matrix of experiment 1. (b) The confusion matrix of experiment 2. (c) The confusion matrix of experiment 3.
Figure 12. Classification results. (a) The confusion matrix of experiment 1. (b) The confusion matrix of experiment 2. (c) The confusion matrix of experiment 3.
Remotesensing 15 05085 g012
Figure 13. The value of evaluation metric with the 3 proposed experiments.
Figure 13. The value of evaluation metric with the 3 proposed experiments.
Remotesensing 15 05085 g013
Table 1. Summary of the main attributes of the diverse relevant approaches and our method.
Table 1. Summary of the main attributes of the diverse relevant approaches and our method.
Method ImprovementSpecific MethodsRadar ImageReferenceYear
Recognition networkLSTMTime-frequency spectrogram[11]2019
MSENet + BLSTM + CTCTime-frequency spectrogram[13]2021
TARANHRRP[14]2019
Denoised + CNNTime-range map[15]2022
RAM + CNNISAR Image[16]2021
Dataser generation networkGANHRRP[17]2022
Dual GANTime-frequency spectrogram[18]2021
GANTime-frequency spectrogram[19]2020
ACGANTime-frequency spectrogram[20]2020
Complex-valued networkCV-CNNSAR image[21]2017
CV-CNNISAR Image[22]2019
CV 3-D CNNPolarimetric SAR image[23]2021
Table 2. List of the Variable Notations Used in this Paper.
Table 2. List of the Variable Notations Used in this Paper.
VariableDescriptionVariableDescription
ω w The wobbling angular velocity R r e f Reference range
t Time s i n c · The sinc function
I The identity matrix y The samples input
ω w The wobble angular velocity z The noise vector
U w The skew-symmetric matrix D The generator
r 0 The initial position of scattering centers G The discriminator
n The unit vector of LOS · , R The real part of complex-valued
t ^ Fast time · , I The imaginary part of complex-valued
t s Slow time C o v The covariance operation
μ The frequency modulation slope of LFM signal P r The distribution of the electromagnetic computation image
f c The wideband radar carrier frequency P s The distribution of synthetic image
τ The time duration Tr The trace of the matrix
c The speed of light Y The number of categories
B The radar bandwidth p e The relative misclassification number
r The motion range N The total number of samples
Table 3. Parameters of EM Model.
Table 3. Parameters of EM Model.
Start frequency10 GHzCarrier frequency12 GHz
End frequency12 GHzPolarizationVV
Range of LOS 0 180 Conic node radius0.075 m
Bottom radius0.64 mHeight3 m
Table 4. Micro-motion parameters of EM computation dataset.
Table 4. Micro-motion parameters of EM computation dataset.
Micro-Motion StatesLOS (°)Angular Velocity (rad/s)Other Parameters
Wobbling30:13:147 ω w : 0.5 π :0.1 π :1.6 π Wobbling amplitude (°): 10:3:31
Tumbling30:5:145 ω t : 0.6 π :0.1 π :4.5 π
Precession30:13:147 ω c : 0.9 π :0.1 π :2 π Precession angle (°): 10:3:31
Nutation30:13:147 ω c : 1 π :0.2 π :2 π ω w : 0.4 π :0.2 π :1 π Wobbling amplitude (°): 10:6:28
Table 5. CV-RDB Architecture.
Table 5. CV-RDB Architecture.
CV-RDB
CV Conv2d, stride = 1, size = 3 × 3
CV Conv2d, stride = 1, size = 3 × 3
CV Conv2d, stride = 1, size = 3 × 3
Depth Concatenation
CV Conv2d, stride = 1, size = 1 × 1
Addition
Table 6. CV-GAN Architecture.
Table 6. CV-GAN Architecture.
GeneratorDiscriminator
Noise, size = 128Image, size = 128 × 128 × 3
Fully connectCV Conv2d, stride = 2, filters = 64, size = 3 × 3
CV-RDB, 512CV-RDB, 128
CV-RDB, 256CV-RDB, 256
CV-RDB, 128CV-RDB, 512
CV BNFully connect
CV Conv2d, stride = 2, filter s= 64, size = 3 × 3CReLU
CReLU
Table 7. Classification module architecture and hyperparameters.
Table 7. Classification module architecture and hyperparameters.
CV-PNetHyperparameters
Input, size = 128 × 128 × 3Optimizer: Adam
CV Conv2d, stride = 2, filters = 64, size = 7 × 7Epochs: 30
Max pooling, stride = 1, size = 5 × 5Dropout: 0.005
CV-PB (number = 10)Initial learning rate: 0.001
Global average poolingLearning rate decay factor: 0.5/5 epoch
Fully connectBath size: 12
SoftMax
Output
Table 8. Performance comparison with different networks.
Table 8. Performance comparison with different networks.
ACCMAPMF1KC
TARAN [14]0.94470.94510.94490.9263
2D CNN [15]0.92840.92920.92880.9045
19-layer CNN [19]0.91790.91780.91790.8906
CV-PNet0.95650.95690.95670.9419
Table 9. Effects of parallel block.
Table 9. Effects of parallel block.
ACCMAPMF1KC
M = 4 0.84110.84670.84390.7882
M = 6 0.89190.89250.89220.8559
M = 8 0.93620.93670.93640.9149
M = 10 0.95650.95690.95670.9419
M = 12 0.96480.96580.96530.9531
Table 10. Effects of parallel branches.
Table 10. Effects of parallel branches.
ACCMAPMF1KC
N = 2 0.78780.78820.78800.7170
N = 3 0.86590.88190.87380.8212
N = 4 0.95650.95690.95670.9419
N = 5 0.96350.96490.96420.9514
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Han, L.; Feng, C.; Hu, X. Space Targets with Micro-Motion Classification Using Complex-Valued GAN and Kinematically Sifted Methods. Remote Sens. 2023, 15, 5085. https://doi.org/10.3390/rs15215085

AMA Style

Han L, Feng C, Hu X. Space Targets with Micro-Motion Classification Using Complex-Valued GAN and Kinematically Sifted Methods. Remote Sensing. 2023; 15(21):5085. https://doi.org/10.3390/rs15215085

Chicago/Turabian Style

Han, Lixun, Cunqian Feng, and Xiaowei Hu. 2023. "Space Targets with Micro-Motion Classification Using Complex-Valued GAN and Kinematically Sifted Methods" Remote Sensing 15, no. 21: 5085. https://doi.org/10.3390/rs15215085

APA Style

Han, L., Feng, C., & Hu, X. (2023). Space Targets with Micro-Motion Classification Using Complex-Valued GAN and Kinematically Sifted Methods. Remote Sensing, 15(21), 5085. https://doi.org/10.3390/rs15215085

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop