Next Article in Journal
The Identification and Driving Factor Analysis of Ecological-Economi Spatial Conflict in Nanjing Metropolitan Area Based on Remote Sensing Data
Next Article in Special Issue
Intelligent Radar Jamming Recognition in Open Set Environment Based on Deep Learning Networks
Previous Article in Journal
A Survey of Computer Vision Techniques for Forest Characterization and Carbon Monitoring Tasks
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Classification of Radar Targets with Micro-Motion Based on RCS Sequences Encoding and Convolutional Neural Network

Air and Missile Defense College, Air Force Engineering University, Xi’an 710038, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2022, 14(22), 5863; https://doi.org/10.3390/rs14225863
Submission received: 10 September 2022 / Revised: 23 October 2022 / Accepted: 25 October 2022 / Published: 19 November 2022
(This article belongs to the Special Issue Targets Characterization by Radars)

Abstract

:
Radar cross section (RCS) sequences, an easy-to-obtain target feature with small data volume, play a significant role in radar target classification. However, radar target classification based on RCS sequences has the shortcomings of limited information and low recognition accuracy. In order to overcome the shortcomings of RCS-based methods, this paper proposes a spatial micro-motion target classification method based on RCS sequences encoding and convolutional neural network (CNN). First, we establish the micro-motion models of spatial targets, including precession, swing and rolling. Second, we introduce three approaches for encoding RCS sequences as images. These three types of images are Gramian angular field (GAF), Markov transition field (MTF) and recurrence plot (RP). Third, a multi-scale CNN is developed to classify those RCS feature maps. Finally, the experimental results demonstrate that RP is best at reflecting the characteristics of the target among those three encoding methods. Moreover, the proposed network outperforms other existing networks with the highest classification accuracy.

1. Introduction

Micro-motion is an important feature of space targets, which is of great significance for target parameters estimation [1,2]. Over the past decade, scholars have carried out extensive research on the micro-motion of space radar targets, and relevant results have been successfully applied to the characteristic analysis of satellites, ballistic targets (BTs), space debris and other targets [3,4,5].
BT classification is a typical application of micro-motion [6,7,8,9]. The micro-motion of the warhead is precession due to the release of the decoy. Lacking an attitude control system, the micro-motion of the heavy decoy is swing, and the micro-motion of the light decoy is rolling. Furthermore, the micro-motions of boosters and the debris are usually rolling. The micro-motions of BTs are different and unique [7,10,11], providing an effective tool for BT classification.
Scholars have conducted various research works on the radar echoes received from BTs. The micro-Doppler (MD) time–frequency graphs (TFGs), the high-resolution range profile (HRRP) and the high-resolution range profile sequences (HRRPs) and the RCS sequences are three important features of echoes and have been widely used for discriminating BTs [12]. Correspondingly, BT classification methods based on micro-motion features are generally classified into the following three categories. The first class mainly uses the MD features on the TFGs of the target. Choi extracted the fundamental frequency, the bandwidth and the sinusoidal moment of TFGs to distinguish the precession/nutation warheads from the wobble decoys [13]. Jung designed a CNN to classify warheads and decoys based on TFGs and a cadence velocity diagram [14]. Zhang presented a complex-valued coordinate attention network to classify the cone–cylinder precession targets with different MD parameters [15]. The second category mainly depends on HRRP/HRRPs to discriminate BTs. Zhou employed HRRPs to estimate the parameters of BTs and thus realized the classification of the warhead, the heavy decoy and the light decoy according to the differences of their parameters [16]. Persico conducted an Iradon transform on HRRP to generate a new feature and input the feature into the K-nearest neighbor classifier (KNN) to classify warheads and false targets [17]. Wang created a novel processing flow based on HRRPs and used CNN to classify five types of micro-motions with similar shapes [18]. The third category mainly relies on RCS sequences. Cai discriminated warheads from debris based on several statistics of RCS sequences, such as the expectation, the variance and the period [19]. Ye developed a gated recurrent unit (GRU) network for radar target recognition based on several statistics of RCS sequences [20]. Choi proposed a multi-feature fusion framework based on RCS sequence for BTs recognition [12]. Chen input RCS sequences into a one-dimensional CNN to realize the intelligent identification of warheads and decoys [21]. In addition to the above three types, methods based on feature fusion have also been a promising technique [22]. Chen fused TFGs and HRRPs of the flying bird and drone to generate a new feature map and then input them to a modified multi-scale CNN to classify the targets [4]. Features based on a combination of RCS, TF, HRRP and RID were adopted to achieve a decision-level fusion, and thereby, a high-accuracy recognition was achieved for space targets with micro-motion [23].
Although the above methods can detect warheads from BTs, there are some significant drawbacks that limit their widespread application. (1) HRRP/HRRPS can only be acquired by wideband radars, and narrowband radars do not have this capability. (2) Only when the repetition frequency of the radar is greater than twice that of the MD can the valid TFGs be generated. Otherwise, Doppler ambiguity will occur in TFGs. (3) The information reflected by RCS sequences is relatively abstract, resulting in a low accuracy for the classification task. Therefore, research works on simple and high-precision methods for BTs recognition have drawn ever-increasing attention.
RCS sequences are easily processed due to the small data volume. Moreover, both wideband radars and narrowband radars can acquire RCS sequences. Therefore, how to mine advanced features from RCS sequences has become the key point to promote the application of RCS-based methods. This paper proposes a BTs classification structure based on RCS sequence encoding and multi-scale convolutional neural network. We introduce MTF, GAF and RP to convert RCS sequences to images, so as to improve the richness of RCS sequences. We develop a multi-scale CNN to extract the features of these images and discriminate four different BTs. Moreover, several experiments are conducted to evaluate the performance of the encoding methods and the proposed network.
The rest of this paper is organized as follows. Section 2 presents the micro-motion of different BTs. Section 3 describes three encoding methods for RCS sequences. A multi-scale CNN is proposed in Section 4. Section 5 comprises the simulation results and performance analysis. Section 6 is the conclusion.

2. The Micro-Motion Model of BTs

Before establishing the micro-motion model, we first introduce the concept of RCS. RCS is a measure of an object’s ability to reflect the electromagnetic wave, defined as 4 π times that of the ratio of the reflected power of the target to the incident power. Set RCS as σ , and the definition is written as
σ = lim R 4 π R 2 E s E i 2 = 4 π E f a r , H 2 + E f a r , V 2 E i 2 = σ H + σ V
where E i represents the intensity of the incident electric field from the radar, E s represents the intensity of the scattering electric field from the target, R is the range from the radar to the target, R indicates that the observation scene is a far-field condition, and the incident wave is regarded as the uniform plane wave. E f a r , H and E f a r , V represent the electric field horizontal component and vertical component, respectively.
Micro-motion will change the attitude of the target, causing fluctuations in RCS. According to the change of RCS sequences, we can derive the micro-motion type and the shape of the target [20]. The above description is the basis of the recognition methods based on RCS sequences.
Rotationally symmetric structures are commonly adopted for BTs, including cone, cone–cylinder, ellipse and so on. For these shapes, the angle between the symmetry axis of the target and the radar line of sight (LOS) (defined as the aspect angle) determines the value of RCS.
In the remainder of this section, we first establish the models of different micro-motions in Figure 1 and then derive the expression of the aspect angle for every motion.
Swing: As shown in Figure 1a, the target swings at a small angle along o z in y o z . Set the amplitude of swing as θ b , the initial phase as θ b _ 0 and the swing frequency as f b . Setting the angle between the symmetry axis and o z as a n g s w ( t ) , we denote a n g s w ( t ) as
a n g s w ( t ) = θ b sin ( 2 π f b t + θ b _ 0 )
Based on a n g s w ( t ) , the unit vector of the symmetry axis is given as
n s w ( t ) = [ 0 , sin ( a n g s w ( t ) ) , cos ( a n g s w ( t ) ) ]
Rolling: As shown in Figure 1b, the target rolls in y o z . Set the initial phase as θ r _ 0 and the rolling frequency as f r . Set the angle between the symmetry axis and o z as a n g r o l ( t ) , and it is written as
a n g r o l ( t ) = θ r _ 0 + 2 π f r t
According to Equation (3), the unit vector of the symmetry axis is written as
n r o l ( t ) = [ 0 , sin ( a n g r o l ( t ) ) , cos ( a n g r o l ( t ) ) ]
Precession: As shown in Figure 1c, precession is a combination of spinning and conning. For rotationally symmetric targets, spinning does not modulate the echo. Therefore, only the modulation of the conning is considered. Set the precession frequency as f p and the initial phase as θ p _ 0 . The azimuth angle of the symmetry axis in y o z is written as
a n g p r e ( t ) = 2 π f p t + θ p _ 0
Set the precession angle as θ p , and the unit vector of the symmetry axis is expressed as
n p r e ( t ) = [ cos ( a n g p r e ( t ) ) sin ( θ p ) , sin ( a n g p r e ( t ) ) sin ( θ p ) , cos ( θ p ) ]
Moreover, we also establish the observation model in Figure 1d. Set α as the azimuth angle of the LOS and the angle between LOS and o z as β . The unit vector of the LOS is expressed as
n L O S = [ cos α sin β , sin α sin β ,   cos β ]
Set ψ ( t ) as the average aspect angle, and it is solved as
ψ ( t ) = c o s 1 ( n L O S × n s ( t ) )
where
n s ( t ) [ n s w ( t ) , n r o l ( t ) , n p r e ( t ) ]

3. Methods for RCS Sequences Encoding

The raw RCS sequence is a 1D signal, which has a relatively poor ability to present the targets’ characteristics and is not conducive to target classification. We believe that converting RCS sequences to 2D images can show the features more abundantly and intuitively, which is an effective and efficient approach to visualize the details and features. In the radar signal processing field, the time–frequency analysis is the most popular encoding method. However, time–frequency analysis has high requirements for sequence length and repetition frequency, so its application scope is very limited.
Based on the above analysis, we introduce three time series encoding methods to map RCS into two-dimensional space, so as to enhance the feature representation of RCS and lay the foundation for high-precision target recognition. These three methods are MTF, GAF and RP. As far as we know, the research on RCS analysis using these three time series coding methods is very limited, and this is exactly one of the important contributions of this paper.

3.1. MTF

The MTF can effectively capture the state transition information of the time series, thereby improving the feature expression [24,25].
Denoting Y = ( y 1 , y 2 , , y n , , y N ) , n [ 1 ,   N ] as a RCS sequence with the length being N , identify Q quantile bins of Y , and every y n is assigned into the related bins q i ( i [ 1 ,   Q ] ) . Counting the first-order transitions among the different bins along the time axis, we will obtain a Q × Q weighted adjacency matrix W , where w i , j is the normalized transition probability of q i q j . Nevertheless, W is insensitive to the step size of the time series. To enhance the expressiveness of temporal information, MTF is designed based on W , and the expression is shown as
M = w i , j | y 1 q i , y 1 q j w i , j | y 1 q i , y N q j w i , j | y N q i , y 1 q j w i , j | y N q i , y N q j
Compared to W , M contains not only the step size information of the time series but also the state transition information. As a result, MTF is widely used for time series analysis.

3.2. GAF

GAF displays the temporal correlation of the sequence in a 2D image, where the motion information of the sequence is represented as the change from the upper left corner to the lower right corner [26,27]. GAF is based on a transformation from the Cartesian coordinate system to the polar coordinate system. The generation process mainly includes three steps.
Step 1. Rescale Y to Y ˜ via (11), so that y n falls between −1 and 1.
y ˜ n = ( y n max ( Y ) ) + ( y n max ( Y ) ) max ( Y ) min ( Y )
Step 2. Transform the scaled sequence from the Cartesian coordinate system ( t , y n ) to the polar coordinate system ( r n , ϕ n ) via Equation (12).
ϕ n = arcos ( y ˜ n ) r n = t n N 0
where t n denotes the time stamp, and N 0 is a constant factor to regularize the span of the polar coordinate system.
Step 3. Extract the angle information in polar coordinates and generate GAF via Equations (13) and (14).
G A S F = cos ( ϕ 1 + ϕ 1 ) cos ( ϕ 1 + ϕ N ) cos ( ϕ N + ϕ 1 ) cos ( ϕ N + ϕ N ) = Y ˜ T Y ˜ ( I Y ˜ 2 ) T I Y ˜ 2
G A D F = sin ( ϕ 1 ϕ 1 ) sin ( ϕ 1 ϕ N ) sin ( ϕ N ϕ 1 ) sin ( ϕ N ϕ N ) = I Y ˜ 2 Y ˜ Y ˜ I Y ˜ 2
where GASF denotes the Gramian angular summation fields, GADF denotes the Gramian angular difference fields, I represents the identity matrix, ( · ) T represents the transpose of the elements.
GASF presents the correlation between two moments based on the sum of cosine functions, while GADF presents the correlation by the difference of sine functions. After conducting GAF on the time series, the correlations of the time series are significantly enhanced.

3.3. RP

As a nonlinear system analysis tool, RP could detect nonlinear features and visualize the recurrent behavior in the time series [28,29].
Set the delay time as τ , the embedding dimension as m , and the phase space matrix of the time series is calculated as
Z = [ z 1 , z 2 , , z M ] = y 1 y 2 y M y 1 + τ y 2 + τ y M + τ y 1 + ( m 1 ) τ y 2 + ( m 1 ) τ y M + ( m 1 ) τ
where
M = N ( m 1 ) τ
Denoting R as the RP, R can be solved as
R ( p , q ) = Θ ( ε - z p z q ) = 1 ; ε z p z q 0 ;   ε < z p z q
where · denotes a norm, and the L 2 norm is selected in this paper; ε is a threshold to determine the state of RP, and p , q [ 1 , M ] .
The generation of RP involves three parameters: τ , m and ε . As for the selection of these three parameters, this paper follows the following rules.
(1) ε . There is a disagreement over the selection of ε . According to some scholars, ε is related to the signal-to-noise ratio (SNR), and according to others, some ε is related to the phase space radius. An unsuitable ε can easily lose the details of the sequence. To preserve the details of the RP, this paper improves the RP in Equation (16) into a non-threshold RP without taking ε into consideration. The non-threshold RP is expressed as
R ( p , q ) = y p y q 2 2
(2) τ and m . Many methods have been utilized to select these two parameters, but different methods may yield different results [30]. After consulting many literature works, we have still not found an acceptable method. Therefore, different values of τ and m are empirically selected to test the performance based on Refs. [28,31,32,33].
These three encoding methods have their own characteristics, and each method contains different information of the RCS sequence. The MTF emphasizes the state transition of the sequences; GAF is good at representing the temporal relation; and RP is able to display the chaos characteristics. By evaluating the classification performance of these methods, the most suitable encoding method can be determined.

4. Proposed Network

With the development of the deep-learning theory, CNNs have been widely used in image classification, pattern recognition and parameter regression [34]. Instead of manually extracting the features, CNNs learn the advanced features by themselves and show advanced performance in those tasks. Typical CNNs include Alexnet, Googlenet and Resnet [4,35]. To achieve an efficient classification, this paper designs a multi CNN. The details of the network are as follows.

4.1. Res2Net

ResNet is a commonly used module in CNN to solve the gradient problem (shown in Figure 2a). On the basis of ResNet, Gao developed a multi-scale module called Res2Net by building hierarchical residual-like connections within one single residual block [36]. Res2net is an effective multi-scale technique, which further explores the multi-scale features and extends the range of receptive fields [37].
The structure of Res2Net is shown in Figure 2b. For the output feature map of the previous layer, the 1 × 1 convolution is adopted to adjust the number of channels; then, the feature map is split into s feature map subsets x i , i [ 1   s ] , and s is the scale factor of Res2Net. The process of channel splitting and convolution can be formulated as
y i = x i i = 1 K i ( x i ) i = 2 K i ( x i + y i 1 ) 2 < i s
where K i ( · ) represents the 3 × 3 convolution for x i . In addition to x 1 , each x i is summed with the output of the previous layer, and the obtained feature map is input into a 3 × 3 convolution to obtain y i . K i ( · ) could receive the feature from other scales before i th , making Res2Net possess a greater receptive field. Such structure can effectively capture the global and local features of the feature maps, so that the feature extraction ability is improved.
Finally, we stack y i along the channel dimension and input the stacked feature map into a 1 × 1 convolution to match the size of the input.

4.2. Channel Attention

To take advantage of these features from different channels, Woo developed a module called the channel attention module (CAM) [38]. CAM assigns attention weights to every channel of the input to obtain a more effective feature map. The structure of CAM is shown in Figure 2c.
For the input ( W × H × C ) , a global average/maximum pooling layer is first conducted along the channel dimension to obtain a vector ( 1 × 1 × C ) . Then, the vector is transformed into the attention weights by a shared MLP (containing two fully connected (FC) layers and a ReLU activation function) and a Sigmoid function. The first FC is utilized to compress the vector ( 1 × 1 × C / s ) , and the second FC is used to restore the vector ( 1 × 1 × C ) . s is the squeeze ratio and is set to 16 in this paper.
CAM can be plugged into the Res2net to fuse the channel-wise information, and thereby, the CAM–Res2net module is generated.

4.3. The Activation Function (AF) and the Loss Function

The most regular AF in CNNs is the ReLU AF, which is very simple to solve. For input x f , the output through the ReLU AF is given as
ReLU ( x f ) = max ( 0 ,   x f )
The ReLU AF will output the negative samples of x f as 0, thereby destroying the transmission of information. The Mish AF is a novel AF used in YOLOv4 for object detection [39]. Based on a combination of the Tanh AF and the Softplus AF, the Mish AF is expressed as
Mish ( x f ) = x f × tanh ( ln ( 1 + e x f ) )
The Mish AF is a continuous function, which has a nonlinear representation effect on positive values and does not induce gradient disappearance on negative values. The Mish AF has a strong regularization limitation and can effectively improve the feature expression ability of the model. However, the complexity of the Mish AF will consume the computing power of the computer. Therefore, both the Mish AF and the ReLU AF are contained in the proposed network, where the Mish AF is used for the deep features, and the ReLU AF is used for the shallow features.
Moreover, we choose the cross entropy loss as the cost function of the proposed network. The cross entropy loss is given as
L = n p n log ( c n ) + ( 1 p n ) log ( 1 c n )
where p n and c n represent the predicted label and the real label, respectively.
In summary, the structure of the proposed network is shown in Table 1.
It needs to be noted that regular convolution is replaced with depthwise separable convolution (DS-Conv) due to fewer parameters. Because DS-Conv is an existing technique, the details are presented in the Appendix A instead of the main text.
As for the motivation of the configuration in Table 1, we will explain as follows.
Several research works have proven that the multi-scale CNN performs well in the classification task. As a result, we propose a multi-scale CNN to achieve high-accuracy classification in this manuscript. Res2net, channel attention, DS-Conv and other details form the proposed network. Actually, the configurations of the existing classical CNNs provide us with a meaningful reference, and those configurations are selected based on the existing CNNs and our own understanding of the CNN’s configuration. Some details are explained as follows.
Due to fewer parameters and more nonlinearity, the kernel size of the convolution is set to 3 × 3 instead of 5 × 5, 7 × 7 and other sizes. There are more convolution kernels in the deeper position of the network (changing from 32, 64, 768, 1024), which shows the emphasis on deeper features. Due to the better validity and reliability of the deeper features, several Res2net/CAM–Res2net modules are adopted to learn the deeper features instead of the shallow features. Ds-Conv is adopted to reduce the number of parameters. Maxpooling and avgpooling are adopted to reduce the dimension of the feature map and strengthen the nonlinearity of the network.
In fact, the convolution layer is always followed by a batch normalization layer and a nonlinear activation layer in this paper. It should be noted that the batch normalization layer and nonlinear activation layer are not exhibited in Table 1. Moreover, the ReLU function exists in parts 1–7, and the Mish function exists in the rest of the network.

5. Experiments

According to Refs [20,21,40], we establish the following four geometric models to represent the warhead, the heavy decoy, the light decoy and the booster (Figure 3).
In the classification task for micro-motion targets, it is difficult to classify objects with similar motions and similar structures. In this paper, the warhead’s structure is similar to that of the weight decoy’s, and the light decoy’s motion type is similar to that of the booster’s. This configuration is used to verify that the proposed network can effectively address this classification difficulty.
The micro-motion parameters of the above targets are shown in Table 2.
For each type of target in Table 2, there are 480 groups of different parameters to generate the RCS of the target.
For BTs classification, datasets based on the measured data are difficult to obtain. As a result, datasets based on electromagnetic calculation are always adopted in these research works [15]. This paper also uses electromagnetic calculation to simulate RCS sequences. The generation process of the training dataset is shown as follows.
Step 1. Based on the real structure of BTs, we establish the simplified geometric model by using AutoCAD (Autodesk Computer Aided Design, a drawing tool software for 2D drawing, detailed drawing and basic 3D design).
Step 2. We import the geometric model into FEKO [20,41] (a software widely used for electromagnetic calculation) to solve the RCS. Set the source as plane wave source, the observation condition as far field, the frequency as 8 GHz (this setting can be thought of as the signal type of the radar being a single frequency signal) and the polarization as the vertical linear polarization. Moreover, the physical optics (PO) algorithm is selected as the electromagnetic calculation method in FEKO.
Step 3. Based on the configuration above, we change the value of the pitching angle of the incident wave from 0 to 180 with a step size of 0.001 and solve the static RCS σ 0 180 of the object.
It needs to be noted that because the polarization of the incident wave is vertical linear polarization, we can thus rewrite Equation (1) as the following equation to solve σ 0 180 .
σ T o t a l σ V e r t i c a l = 4 π | E f a r , V | 2 E 0 2
Step 4. According to the micro-motion parameters in Table 2, we solve the ψ ( t ) t = 0 3 s via Equations (2)–(9) for each group parameter of the target.
Moreover, during the calculation of ψ ( t ) t = 0 3 s , the repetition frequency of the radar is set to 256 Hz, and the duration time is 3 s.
Step 5. Based on the value of ψ ( t ) t = 0 3 s and σ 0 180 , the method of interpolation is employed to solve the RCS sequence σ ψ ( t ) .
Step 6. For every σ ψ ( t ) , two sequences will be extracted with stage 0–2 s and stage 0.25–2.25 s. The length of one RCS sequence is 512.
Step 7. Perform different encoding methods on the RCS sequence, and hence, the encoded images are generated. Adjust the size of the encoded image to 64 × 64 × 1 to match the input size of the proposed network, and thereby, the dataset for BT classification is generated.
As a result, there are four types of micro-motion targets in the dataset, and the number of samples for each type is 960. If we want to add noise to the RCS, the noise should be added to the E f a r , V and then solve the value of σ ψ ( t ) via Equation (22). For each fixed SNR, there are 3840 samples in the dataset.
To visualize the modulation of micro-motion on RCS sequences and on encoded images, we randomly selected one sample for every target. The RCS sequences and the corresponding encoded images of the four targets are shown in Figure 4, Figure 5, Figure 6 and Figure 7.
It needs to be noted that the RP is generated with τ = 4 , m = 5 , and the MTF is generated with Q = 8 . According to Figure 4, Figure 5, Figure 6 and Figure 7, we find that different targets will induce different RCS sequences. The RCS of the precession warhead is relatively complex, which means that precession is a complex motion. The RCS of the rolling light decoy is simple, and the period is obvious. The structure of the rolling booster is the most complex among the four targets, and the RCS sequence of the swing weight decoy is relatively complex. As for the encoding methods, the textures on the RP are the sharpest, which means that the chaotic characteristic of the RCS sequence is strong. Textures on the other three images are very irregular, which may induce low classification accuracy.
As for the splitting of the dataset, the random selection method is used to split the dataset to guarantee the generalization ability of the network. We divide the dataset into a training subset ( 60 % ), a validation subset ( 20 % ) and a test subset ( 20 % ). The training subset is adopted to train the network; the validation subset is used to evaluate the training process and select suitable hyperparameters; and the testing subset is adopted to evaluate the performance of the network. After conducting the performance analysis of several training processes with different hyperparameters, the following parameters are selected as the hyperparameters to achieve a relatively high accuracy classification. We use the SGDM method to train the proposed network; the initial learning rate is set to 0.02 and will drop by a factor of 0.5 every 4 epochs; the total epoch is set to 20 , and the batch size is set to 32 . Moreover, the computer graphics card is NVIDIA GeForce RTX 3070. The hyperparameters of the second and third experiments are the same as those of the first experiment.
Accuracy, Precision, Recall and F1 are the four indicators widely used for evaluating the classification task. The performance analysis of the proposed method is based on these indicators. TP represents how many positive samples predict positively; FP denotes how many negative samples predict positively; TN stands for the number of positive samples, which are predicted to be negative; and FN stands for the number of positive samples whose predictions are negative. These four indicators are solved via Equation (23).
A c c u r a c y = T P + T N T P + T N + F P + F N P r e c i s i o n = T P T P + F P R e c a l l = T P T P + F N F 1   = 2 × P r e c i s i o n × R e c a l l P r e c i s i o n + R e c a l l
To evaluate the performance of the encoding methods and the proposed network, three group experiments are conducted in this paper. The first experiment is conducted to show the accuracy of the different encoding methods. The result is shown in Table 3, Table 4 and Table 5.
Table 3 presents the performance of the proposed network based on MTF. Different values of Q mean different numbers of the quantile bins. With the increase in Q , the accuracy goes up first but then goes down. This tendency indicates that either too large or too small Q is unfavorable for feature expression, and thus, the poor accuracy occurs. When Q is set to 8, the accuracy is the best, and the recall for the three targets is also the best. Setting Q to 8 is the most reasonable choice for MTF-based classification methods.
According to Table 4, although there are differences in the discrimination ability for different targets, the overall classification ability of GASF and GADF is similar. We think that the reason for this phenomenon is that the generation ways of GASF and GADF are similar.
Seven groups with different τ and m are conducted to generate different RPs. Overall, the difference will induce a different accuracy for the proposed network, but the difference is not obvious. The difference between the best accuracy and the worst accuracy is only 0.0034. However, compared with MTF and GAF, the RP-based network performs much better than the other two encoding methods. The difference between the RP-based network and the other two networks is approximately 0.06, and such difference is significant. Such a result is consistent with our previous statement, as RP intuitively displays more details of RCS sequences than other methods.
To make the classification results more accurate and clear, the confusion matrices of different encoding methods are presented in Figure 8. It needs to be noted that the confusion matrices give a synthesis of the classification results obtained from 10 trials.
In Figure 8, T1, T2, T2 and T4 represent the warhead, the heavy decoy, the light decoy and the booster, respectively. According to the accuracy of these three encoding methods, we draw the conclusion that RP is the optimal encoding method, GAF is the second, and MTF is the worst. For these four targets, the booster is the easiest to identify; the classification difficulties of the light decoy and the warhead are close; and the weight decoy is the most difficult to identify.
The second experiment is conducted to evaluate how the proportion of the training set affects the classification accuracy. In general, we believe that the number of training sets will have an impact on the performance of the network. As a result, we conduct the experiment to analyze how strong the impact is. Taking the RP as the input, different proportions of the training set to the total samples are employed to train the network, and the accuracy of the test set is shown in Figure 9.
Since this experiment aims to analyze the effects of the number of training sets, all the experiments should be performed with the same hyperparameters. As a result, the parameters are kept the same as those in the first experiment. Moreover, we only split the datasets into the training subset and the testing subset, without taking the validation subset into consideration. With the proportion value changing from 0.1 to 0.8, the accuracy of the proposed network is constantly increasing. This is a normal trend because the larger the number of training sets, the richer the features the network can learn. However, when the proportion changes from 0.8 to 0.9, there is a small drop emerging in the accuracy. This phenomenon is very strange and incredible. After consulting several literature works, we found a reasonable explanation for this phenomenon. More training sets mean fewer test sets are available, which makes it difficult for the network to match the test set with the prediction labels. Within certain limits, the impact of a small test set is stronger than the impact of a large training set. This conclusion is very important, and it will provide an important reference for our future work.
To evaluate the robustness of the proposed method, experiments are conducted with different SNRs. Furthermore, three CNNs are employed as comparisons to verify the superiority of the proposed method. It should be noted that the proposed method, Resnet50 and Alexnet are based on the RP ( τ = 4 , m = 5 ), while the 1D CNN is based on the RCS sequences. Based on the optimal result of the second experiment, the proportion of the training set is set to 0.8, and the proportion of the testing set is set to 0.2. By using the random selection method, the dataset for the third experiment is generated. Taking the accuracy and F1 score as the criteria, the results are shown in Figure 10 and Table 6.
According to the accuracy and F1 score, the algorithm in this paper has the best performance among the four algorithms in most cases. The advantages of the algorithm in this paper are mainly reflected in two aspects. (1) Compared with AlexNet, Resnet50 and Googlenet, the datasets of the proposed algorithm are the same as those in the three networks, but the network structure is different. A comparison of these three algorithms shows the effectiveness of the proposed network. (2) Compared with 1D CNN, the superiority of the proposed method in this paper is obvious, which indicates that compared with the original RCS sequence, the encoding of the RCS sequence can also improve the recognition rate of the image. However, there is one factor, which is common to all algorithms. When the SNR is low, the performance of the RCS-based method is relatively poor, which proves that RCS is sensitive to the noise.
In addition to the accuracy, time consumption is also an important indicator for performance analysis. We calculate the prediction time for one sample of the five methods, and the result is presented in Table 7.
As shown in Table 7, the prediction time of the proposed method is the highest among the five methods. Obviously, the high accuracy comes at the cost of the network’s prediction time for the proposed method. As a result, one valuable aim is to reduce the prediction time of the proposed method.

6. Conclusions

In this paper, we develop a framework for BTs classification. The framework first converts the RCS sequences into images with three methods and then inputs those images to the proposed multi-scale CNN. The experimental results show that the RP-based method has the best accuracy among these three encoding methods. Moreover, the proposed network outperforms other existing networks with better accuracy and robustness. However, several parameters are empirically selected, including the parameters involved in the encoding methods and hyperparameters of the network. If those parameters are selected more precisely and scientifically, the performance of the proposed method may benefit from a significant improvement. In our future work, we will study the method to adaptively select those parameters to make the proposed method more effective. However, the performance of the presented network is based on a large number of labeled samples, which is rarely applicable in the real-world radar target classification tasks [42]. As such, developing a few-shot learning method for radar target classification will be a meaningful and promising technique [43,44]. In the future, we will pay more attention to the few-shot learning methods to make the proposed methods more available.

Author Contributions

Writing—original draft, X.X.; Writing—review & editing, X.X., C.F. and L.H. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the Natural Science Basic Research Program of Shaanxi (No. 2021-JQ-361).

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A. Depthwise Separable Convolution (DS-Conv)

In 2017, F. Chollet of Google Labs proposed a new CNN named Xception [45]. The most important highlight of Xception is DS-Conv, which reduces the complexity of the network without a loss of accuracy.
DS-Conv replaces regular convolution with depthwise convolution followed by pointwise convolution (standard convolution with 1 × 1 kernel) [46]. First, every channel of the input is convoluted through depthwise convolution, so that spatial correlations are achieved. Then, the features of each channel are combined through a pointwise convolution, and channel correlations are achieved. Set the kernel size K × K for DS-Conv, and the diagram is shown in Figure A1.
Figure A1. The diagram of DS-Conv.
Figure A1. The diagram of DS-Conv.
Remotesensing 14 05863 g0a1
Mathematically, depthwise convolution is formulated as
G ( x d , y d , j d ) = u = 1 k v = 1 k K ( u , v , j d ) × F i n ( x d + u k + 1 2 , y d + v k + 1 2 , j d )
where K represents the convolution operation, k denotes the kernel size, F i n represents the input ( h × w × c i n ). x d and y d is the position of every pixel, u and v is the step order of the convolution kernel, j d is the order of the channel, and G is the output feature map. The parameter quantity involved in depthwise convolution is K × K × c i n .
Then, pointwise convolution is calculated as
F o u t ( x d , y d , l ) = j = 1 c i n G ( x d , y d , j d ) × P ( j d , l )
where P ( j d , l ) is the 1 × 1 convolution, l [ 1 ,   c o u t ] . The number of parameters involved in this process is 1 × 1 × c i n × c o u t . For regular convolution, the parameter quantity is K × K × c i n × c o u t [46], which is bigger than that of DS-Conv’ ( c i n × c o u t + K × K × c i n ).
To reduce the parameter quantity, this paper uses separable convolution many times in the proposed network.

References

  1. Tahmoush, D.; Ling, H.; Stankovic, L.; Thayaparan, T.; Narayanan, R. Current Research in Micro-Doppler: Editorial for the Special Issue on Micro-Doppler. IET Radar Sonar Nav. 2015, 9, 1137–1139. [Google Scholar] [CrossRef]
  2. Choi, I.; Jung, J.; Kim, K.; Park, S. Effective Discrimination between Warhead and Decoy in Mid-Course Phase of Ballistic Missile. J. Korean Inst. Electromagn. Eng. Sci. 2020, 31, 468–477. [Google Scholar] [CrossRef]
  3. Zhang, W.P.; Li, K.L.; Jiang, W.D. Parameter Estimation of Radar Targets with Macro-Motion and Micro-Motion Based on Circular Correlation Coefficients. IEEE Signal Proc. Lett. 2015, 22, 633–637. [Google Scholar] [CrossRef]
  4. Chen, X.L.; Zhang, H.; Song, J.; Guan, J.; Li, J.F.; He, Z.W. Micro-Motion Classification of Flying Bird and Rotor Drones via Data Augmentation and Modified Multi-Scale CNN. Remote Sens. 2022, 14, 1107. [Google Scholar] [CrossRef]
  5. Kim, B.K.; Kang, H.S.; Park, S.O. Drone Classification Using Convolutional Neural Networks With Merged Doppler Images. IEEE Geosci. Remote S 2017, 14, 38–42. [Google Scholar] [CrossRef]
  6. Zhu, N.; Xu, S.; Li, C.; Hu, J.; Fan, X.; Wu, W.; Chen, Z. An Improved Phase-Derived Range Method Based on High-Order Multi-Frame Track-Before-Detect for Warhead Detection. Remote Sens. 2022, 14, 29. [Google Scholar] [CrossRef]
  7. Zhu, N.; Hu, J.; Xu, S.; Wu, W.; Zhang, Y.; Chen, Z. Micro-Motion Parameter Extraction for Ballistic Missile with Wideband Radar Using Improved Ensemble EMD Method. Remote Sens. 2021, 13, 3545. [Google Scholar] [CrossRef]
  8. He, F.Y.; Xiao, Z.Y. Micro-motion modelling and analysis of extended ballistic targets based on inertial parameters. Electron. Lett. 2013, 49, 129–130. [Google Scholar] [CrossRef]
  9. Gao, H.; Xie, L.; Wen, S.; Kuang, Y. Micro-Doppler Signature Extraction from Ballistic Target with Micro-Motions. IEEE Trans. Aerosp. Electron. Syst. 2010, 46, 1969–1982. [Google Scholar] [CrossRef]
  10. Han, L.; Feng, C. Micro-Doppler-Based Space Target Recognition with a One-Dimensional Parallel Network. Int. J. Antennas Propag. 2020, 2020, 8013802. [Google Scholar] [CrossRef]
  11. Hanif, A.; Muaz, M.; Hasan, A.; Adeel, M. Micro-Doppler Based Target Recognition With Radars: A Review. IEEE Sens. J. 2022, 22, 2948–2961. [Google Scholar] [CrossRef]
  12. Choi, I.O.; Park, S.H.; Kim, M.; Kang, K.B.; Kim, K.T. Efficient Discrimination of Ballistic Targets With Micromotions. IEEE Trans. Aerosp. Electron. Syst. 2020, 56, 1243–1261. [Google Scholar] [CrossRef]
  13. Choi, I.O.; Jung, J.H.; Kim, K.T.; Riggs, L.; Kim, S.H.; Park, S.H. Efficient 3DFV for improved discrimination of ballistic warhead. Electron Lett. 2018, 54, 1452–1453. [Google Scholar] [CrossRef]
  14. Jung, K.; Lee, J.I.; Kim, N.; Oh, S.; Seo, D.W. Classification of Space Objects by Using Deep Learning with Micro-Doppler Signature Images. Sensors 2021, 21, 4365. [Google Scholar] [CrossRef]
  15. Zhang, Y.P.; Zhang, Q.; Kang, L.; Luo, Y.; Zhang, L. End-to-End Recognition of Similar Space Cone–Cylinder Targets Based on Complex-Valued Coordinate Attention Networks. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–14. [Google Scholar] [CrossRef]
  16. Daiying, Z.; Jianhui, Y. Ballistic target recognition based on micro-motion characteristics using sequential HRRPs. In Proceedings of the 2013 IEEE 4th International Conference on Electronics Information and Emergency Communication, Beijing, China, 15–17 November 2013; pp. 165–168. [Google Scholar]
  17. Persico, A.R.; Ilioudis, C.V.; Clemente, C.; Soraghan, J.J. Novel Classification Algorithm for Ballistic Target Based on HRRP Frame. IEEE Trans. Aerosp. Electron. Syst. 2019, 55, 3168–3189. [Google Scholar] [CrossRef] [Green Version]
  18. Wang, Y.Z.; Feng, C.Q.; Hu, X.W.; Zhang, Y.S. Classification of Space Micromotion Targets With Similar Shapes at Low SNR. IEEE Geosci. Remote Sens. Lett. 2021, 19, 1–5. [Google Scholar] [CrossRef]
  19. Cai, T.; Sheng, Y.; He, Z.; Sun, W. Classification and recognition of ballistic microcephalus based on deep neural network. In Proceedings of the 2019 International Applied Computational Electromagnetics Society Symposium, Nanjing, China, 8–11 August 2019; pp. 1–2. [Google Scholar]
  20. Ye, L.; Hu, S.B.; Yan, T.T.; Meng, X.; Zhu, M.Q.; Xu, R.Z. Radar target shape recognition using a gated recurrent unit based on RCS time series’ statistical features by sliding window segmentation. IET Radar Sonar Navig. 2021, 15, 1715–1726. [Google Scholar] [CrossRef]
  21. Chen, J.; Xu, S.Y.; Chen, Z.P. Convolutional neural network for classifying space target of the same shape by using RCS time series. IET Radar Sonar Navig. 2018, 12, 1268–1275. [Google Scholar] [CrossRef]
  22. Noori, F.M.; Riegler, M.; Uddin, M.Z.; Torresen, J. Human Activity Recognition from Multiple Sensors Data Using Multi-fusion Representations and CNNs. ACM Trans. Multimedia Comput. Commun. Appl. 2020, 16, 1–19. [Google Scholar] [CrossRef]
  23. Tian, X.; Bai, X.; Xue, R.; Qin, R.; Zhou, F. Fusion Recognition of Space Targets With Micromotion. IEEE Trans. Aerosp. Electron. Syst. 2022, 58, 3116–3125. [Google Scholar] [CrossRef]
  24. Zhao, X.; Sun, H.; Lin, B.; Zhao, H.; Niu, Y.; Zhong, X.; Wang, Y.; Zhao, Y.; Meng, F.; Ding, J.; et al. Markov Transition Fields and Deep Learning-Based Event-Classification and Vibration-Frequency Measurement for φ-OTDR. IEEE Sens. J. 2022, 22, 3348–3357. [Google Scholar] [CrossRef]
  25. Esmael, A.A.; da Silva, H.H.; Ji, T.; Torres, R.D. Non-Technical Loss Detection in Power Grid Using Information Retrieval Approaches: A Comparative Study. IEEE Access 2021, 9, 40635–40648. [Google Scholar] [CrossRef]
  26. Paulo, J.R.; Pires, G.; Nunes, U.J. Cross-Subject Zero Calibration Driver’s Drowsiness Detection: Exploring Spatiotemporal Image Encoding of EEG Signals for Convolutional Neural Network Classification. IEEE Trans. Neural Syst. Rehabil. Eng. 2021, 29, 905–915. [Google Scholar] [CrossRef] [PubMed]
  27. Wang, Z.; Oates, T. Spatially Encoding Temporal Correlations to Classify Temporal Data Using Convolutional Neural Networks. arXiv 2015, arXiv:1509.07481. [Google Scholar]
  28. Zhang, Y.; Hou, Y.; OuYang, K.; Zhou, S. Multi-scale signed recurrence plot based time series classification using inception architectural networks. Pattern Recognit. 2022, 123, 108385. [Google Scholar] [CrossRef]
  29. Dias, D.; Pinto, A.; Dias, U.; Lamparelli, R.; Le Maire, G.; Torres, R.D. A Multirepresentational Fusion of Time Series for Pixelwise Classification. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2020, 13, 4399–4409. [Google Scholar] [CrossRef]
  30. Faustine, A.; Pereira, L.; Klemenjak, C. Adaptive Weighted Recurrence Graphs for Appliance Recognition in Non-Intrusive Load Monitoring. IEEE Trans. Smart Grid 2021, 12, 398–406. [Google Scholar] [CrossRef]
  31. Deng, Q.Q.; Lu, H.Z.; Hu, M.F.; Zhao, B.D. Exo-atmospheric infrared objects classification using recurrence-plots-based convolutional neural networks. Appl. Optics 2019, 58, 164–171. [Google Scholar] [CrossRef]
  32. Pham, T.D. Pattern analysis of computer keystroke time series in healthy control and early-stage Parkinson’s disease subjects using fuzzy recurrence and scalable recurrence network features. J. Neurosci Meth. 2018, 307, 194–202. [Google Scholar] [CrossRef]
  33. Pham, T.D.; Wardell, K.; Eklund, A.; Salerud, G. Classification of Short Time Series in Early Parkinson’s Disease With Deep Learning of Fuzzy Recurrence Plots. IEEE-CAA J. Autom. Sin. 2019, 6, 1306–1317. [Google Scholar] [CrossRef]
  34. Zhang, Z.; Luo, H.; Wang, C.; Gan, C.; Xiang, Y. Automatic Modulation Classification Using CNN-LSTM Based Dual-Stream Structure. IEEE Trans. Veh. Technol. 2020, 69, 13521–13531. [Google Scholar] [CrossRef]
  35. Xie, X.J.; Peng, S.L.; Yang, X. Deep Learning-Based Signal-To-Noise Ratio Estimation Using Constellation Diagrams. Mob. Inf Syst. 2020, 2020, 1–9. [Google Scholar] [CrossRef]
  36. Gao, S.H.; Cheng, M.M.; Zhao, K.; Zhang, X.Y.; Yang, M.H.; Torr, P. Res2Net: A New Multi-Scale Backbone Architecture. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 43, 652–662. [Google Scholar] [CrossRef] [Green Version]
  37. Que, Y.; Li, S.L.; Lee, H.J. Attentive Composite Residual Network for Robust Rain Removal from Single Images. IEEE Trans. Multimedia 2021, 23, 3059–3072. [Google Scholar] [CrossRef]
  38. Woo, S.H.; Park, J.; Lee, J.Y.; Kweon, I.S. CBAM: Convolutional Block Attention Module. In Proceedings of the 15th European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 3–19. [Google Scholar]
  39. Dubey, S.R.; Singh, S.K.; Chaudhuri, B.B. Activation functions in deep learning: A comprehensive survey and benchmark. Neurocomputing 2022, 503, 92–108. [Google Scholar] [CrossRef]
  40. Tang, W.; Yu, L.; Wei, Y.; Tong, P. Radar Target Recognition of Ballistic Missile in Complex Scene. In Proceedings of the 2019 IEEE International Conference on Signal, Information and Data Processing (ICSIDP), Chongqing, China, 11–13 December 2019; pp. 1–6. [Google Scholar]
  41. Kim, I.H.; Choi, I.S.; Chae, D.Y. A Study on the Performance Enhancement of Radar Target Classification Using the Two-Level Feature Vector Fusion Method. J. Electromagn. Eng. Sci. 2018, 18, 206–211. [Google Scholar] [CrossRef] [Green Version]
  42. Che, J.; Wang, L.; Bai, X.; Liu, C.; Zhou, F. Spatial-Temporal Hybrid Feature Extraction Network for Few-shot Automatic Modulation Classification. IEEE Trans. Veh. Technol. 2022, 1–6. [Google Scholar] [CrossRef]
  43. Cheng, M.; Wang, H.L.; Long, Y. Meta-Learning-Based Incremental Few-Shot Object Detection. IEEE Trans. Circuits Syst. Video Technol. 2022, 32, 2158–2169. [Google Scholar] [CrossRef]
  44. Quan, J.N.; Ge, B.Z.; Chen, L. Cross attention redistribution with contrastive learning for few shot object detection. Displays 2022, 72, 102162. [Google Scholar] [CrossRef]
  45. Chollet, F. Xception: Deep Learning with Depthwise Separable Convolutions. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 1800–1807. [Google Scholar]
  46. Muhammad, W.; Aramvith, S.; Onoye, T. Multi-scale Xception based depthwise separable convolution for single image super-resolution. PLoS ONE 2021, 16, 20. [Google Scholar] [CrossRef] [PubMed]
Figure 1. (a) Swing; (b) Rolling; (c) Precession; (d) Observation models of different micro-motions.
Figure 1. (a) Swing; (b) Rolling; (c) Precession; (d) Observation models of different micro-motions.
Remotesensing 14 05863 g001
Figure 2. The Res2net module and CAM module. (a) Resnet module; (b) Res2net/CAM–Res2net module; (c) CAM module.
Figure 2. The Res2net module and CAM module. (a) Resnet module; (b) Res2net/CAM–Res2net module; (c) CAM module.
Remotesensing 14 05863 g002
Figure 3. Models of the BTs. (a) Warhead; (b) Heavy decoy; (c) Light decoy; (d) Booster.
Figure 3. Models of the BTs. (a) Warhead; (b) Heavy decoy; (c) Light decoy; (d) Booster.
Remotesensing 14 05863 g003
Figure 4. (a) RCS; (b) RP; (c) MTF; (d) GASF; (e) GADF. RCS sequences encoding of the warhead.
Figure 4. (a) RCS; (b) RP; (c) MTF; (d) GASF; (e) GADF. RCS sequences encoding of the warhead.
Remotesensing 14 05863 g004aRemotesensing 14 05863 g004b
Figure 5. (a) RCS; (b) RP; (c) MTF; (d) GASF; (e) GADF. RCS sequences encoding of the heavy decoy.
Figure 5. (a) RCS; (b) RP; (c) MTF; (d) GASF; (e) GADF. RCS sequences encoding of the heavy decoy.
Remotesensing 14 05863 g005
Figure 6. (a) RCS; (b) RP; (c) MTF; (d) GASF; (e) GADF. RCS sequences encoding of the light decoy.
Figure 6. (a) RCS; (b) RP; (c) MTF; (d) GASF; (e) GADF. RCS sequences encoding of the light decoy.
Remotesensing 14 05863 g006
Figure 7. (a) RCS; (b) RP; (c) MTF; (d) GASF; (e) GADF. RCS sequences encoding of the booster.
Figure 7. (a) RCS; (b) RP; (c) MTF; (d) GASF; (e) GADF. RCS sequences encoding of the booster.
Remotesensing 14 05863 g007
Figure 8. (a) MTF-8; (b) GASF; (c) GADF; (d) RP (4, 5). Confusion matrices of different encoding methods.
Figure 8. (a) MTF-8; (b) GASF; (c) GADF; (d) RP (4, 5). Confusion matrices of different encoding methods.
Remotesensing 14 05863 g008
Figure 9. The performance with different proportions of the training set.
Figure 9. The performance with different proportions of the training set.
Remotesensing 14 05863 g009
Figure 10. The classification performance for different methods.
Figure 10. The classification performance for different methods.
Remotesensing 14 05863 g010
Table 1. The configuration of the proposed multi-scale CNN for BTs classification.
Table 1. The configuration of the proposed multi-scale CNN for BTs classification.
OrderShortcutCycleOperationOutput Size
1Input64 × 64 × 1
2Conv 32.3 × 3, stride = 162 × 62 × 32
3Conv 128.3 × 3, stride = 162 × 62 × 128
4Maxpooling, 3 × 3, stride = 230 × 30 × 128
5Res2net 128-256-12830 × 30 × 128
6Conv 1 × 1, stride = 2DS-Conv 768.3 × 3, stride = 1
DS-Conv 768.3 × 3, stride = 1
Maxpooling, 3 × 3, stride = 2
15 × 15 × 768
72Res2net15 × 15 × 768
8 CAM-Res2net15 × 15 × 768
9Conv 1 × 1, stride = 2DS-Conv 1024.3 × 3, stride = 1
DS-Conv 1024.3 × 3, stride = 1
Maxpooling, 3 × 3, stride = 2
8 × 8 × 1024
10CAM-Res2net8 × 8 × 1024
11DS-Conv 256.3 × 3, stride = 18 × 8 × 256
12Global Avgpooling1 × 1 × 256
13Dropout 0.21 × 1 × 256
14Fc 41 × 1 × 4
15OutputPredicted label
Table 2. The parameter settings of the datasets.
Table 2. The parameter settings of the datasets.
WarheadHeavy DecoyLight DecoyBooster
PrecessionSwingRollingRolling
α ( ) 20 : 20 : 60 20 : 20 : 60 15 : 10 : 65 15 : 10 : 65
β ( ) 30 : 10 : 60 20 : 15 : 155 20 : 15 : 155 15 : 10 : 65
f c ( Hz ) 0.5 : 1 : 3.5
f b ( Hz ) 0.5 : 1 : 3.5
f r ( Hz ) 0.5 : 0.4 : 3.3 0.5 : 0.4 : 3.3
θ p ( ) 6 : 1 : 15
θ b ( ) 6 : 3 : 15
Table 3. The classification evaluation based on MTFs.
Table 3. The classification evaluation based on MTFs.
Q WarheadWeight DecoyLight DecoyBooster
RecallAccuracy
40.88540.74740.89640.91250.8604
80.92190.76040.91150.94270.8841
160.90830.77240.88280.93230.8740
320.89060.73800.85730.88750.8434
640.79530.65100.82240.79170.7651
Table 4. The classification evaluation based on GAFs.
Table 4. The classification evaluation based on GAFs.
WarheadWeight DecoyLight DecoyBooster
RecallAccuracy
GASF0.93230.79170.91150.97400.9023
GADF0.91670.77080.94790.91670.8880
Table 5. The classification evaluation based on RPs.
Table 5. The classification evaluation based on RPs.
( τ , m ) WarheadWeight DecoyLight DecoyBooster
RecallAccuracy
(1, 1)0.95730.88910.95990.99380.9500
(1, 3)0.95260.88440.96200.99230.9478
(1, 5)0.96350.87340.96090.99320.9478
(4, 3)0.95470.87340.97400.99530.9493
(7, 3)0.96200.89900.94840.99530.9512
(4, 5)0.96350.90100.95310.97920.9492
(4, 7)0.96720.87760.95990.99580.9501
Table 6. The performance evaluation based on F1 score.
Table 6. The performance evaluation based on F1 score.
20 dB15 dB10 dB5 dB0 dB
Resnet500.98020.97520.94930.86970.6822
Googlenet0.94700.93870.88470.78380.6306
Alexnet0.93730.92260.86720.78450.6387
1D CNN0.93900.93700.91530.88150.6938
Proposed0.98680.97580.95070.86390.7013
Table 7. The performance evaluation based on prediction time for one sample.
Table 7. The performance evaluation based on prediction time for one sample.
Resnet50GooglenetAlexnet1D CNNProposed
Time(us)17.3812.665.37014.354633.79
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Xu, X.; Feng, C.; Han, L. Classification of Radar Targets with Micro-Motion Based on RCS Sequences Encoding and Convolutional Neural Network. Remote Sens. 2022, 14, 5863. https://doi.org/10.3390/rs14225863

AMA Style

Xu X, Feng C, Han L. Classification of Radar Targets with Micro-Motion Based on RCS Sequences Encoding and Convolutional Neural Network. Remote Sensing. 2022; 14(22):5863. https://doi.org/10.3390/rs14225863

Chicago/Turabian Style

Xu, Xuguang, Cunqian Feng, and Lixun Han. 2022. "Classification of Radar Targets with Micro-Motion Based on RCS Sequences Encoding and Convolutional Neural Network" Remote Sensing 14, no. 22: 5863. https://doi.org/10.3390/rs14225863

APA Style

Xu, X., Feng, C., & Han, L. (2022). Classification of Radar Targets with Micro-Motion Based on RCS Sequences Encoding and Convolutional Neural Network. Remote Sensing, 14(22), 5863. https://doi.org/10.3390/rs14225863

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop