Next Article in Journal
Combined Layer/Particle Approaches in Surface Molecular Imprinting of Proteins: Signal Enhancement and Competition
Previous Article in Journal
RESTOP: Retaining External Peripheral State in Intermittently-Powered Sensor Systems
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Radar HRRP Target Recognition Based on Stacked Autoencoder and Extreme Learning Machine

College of Electronic Science, National University of Defense Technology, Changsha 410073, China
*
Author to whom correspondence should be addressed.
Sensors 2018, 18(1), 173; https://doi.org/10.3390/s18010173
Submission received: 2 December 2017 / Revised: 30 December 2017 / Accepted: 8 January 2018 / Published: 10 January 2018
(This article belongs to the Section Remote Sensors)

Abstract

:
A novel radar high-resolution range profile (HRRP) target recognition method based on a stacked autoencoder (SAE) and extreme learning machine (ELM) is presented in this paper. As a key component of deep structure, the SAE does not only learn features by making use of data, it also obtains feature expressions at different levels of data. However, with the deep structure, it is hard to achieve good generalization performance with a fast learning speed. ELM, as a new learning algorithm for single hidden layer feedforward neural networks (SLFNs), has attracted great interest from various fields for its fast learning speed and good generalization performance. However, ELM needs more hidden nodes than conventional tuning-based learning algorithms due to the random set of input weights and hidden biases. In addition, the existing ELM methods cannot utilize the class information of targets well. To solve this problem, a regularized ELM method based on the class information of the target is proposed. In this paper, SAE and the regularized ELM are combined to make full use of their advantages and make up for each of their shortcomings. The effectiveness of the proposed method is demonstrated by experiments with measured radar HRRP data. The experimental results show that the proposed method can achieve good performance in the two aspects of real-time and accuracy, especially when only a few training samples are available.

1. Introduction

Radar target recognition based on high-resolution range profile (HRRP) has become a research hotspot due to the acquisition and processing of HRRP data being relatively easy [1,2,3,4,5,6,7]. However, the non-cooperative recognition [8,9] with limited training samples is a challenging task. In the non-cooperative situation, such as at the battle with time, the amount of data under the test is usually huge but the training data is limited. This is because the radar system cannot be guaranteed to detect and track the non-cooperative targets for a long period of time, which will cause HRRP data to be lost or not observed. Therefore, it is very important to study the generalization performance of the recognition model and obtain good recognition performance under conditions of fewer training samples.
It is generally known that feature extraction is a key step in radar target recognition. The quality of the extracted features determines the performance of target recognition. Therefore, many scholars [3,6,10,11,12,13,14,15] have spent a lot of effort studying the methods of HRRP feature extraction. In [3], the principal component analysis (PCA) subspace model is utilized to minimize the reconstruction error. The multitask learning truncated stick-breaking hidden Markov model (MTL TSB-HMM) proposed in [6] is used to characterize the fast fourier transform (FFT) magnitude features of HRRP. Some other researchers [10,11] have used complicated statistic models to extract features from HRRP that have specific physical meaning, such as the target size, the center of gravity, the number of peaks, and so on. By using the RELAX and other super-resolution algorithms, the precise location and intensity information of radar HRRP scatterers can be extracted [12,13]. Manifold learning is used in target recognition of radar HRRP to reduce the feature dimensions [14]. Dictionary learning is adopted in [15] to extract the noise-robust and highly discriminative features of the HRRPs. These methods can work well on some occasions, but all of them are shallow architectures that cannot effectively characterize the radar HRRP. What is more, the features are mostly artificially designed and they need to rely on the experience of the researchers. If we do not have sufficient prior knowledge, the extracted features would be incomplete. Therefore, how to automatically extract the deep abstract features that are beneficial for target recognition has become an important issue.
The deep learning [16] theory put forward by Hinton can effectively solve the above problem. The essence of deep learning is to construct a neural network containing multiple hidden layers to map the data in order to obtain the deep essential characteristics [17]. A deep belief network is used in [18] to solve the non-cooperative target recognition with an imbalanced training dataset. As an important component of the deep learning structure, the stacked autoencoder (SAE) plays an important role in unsupervised learning and nonlinear feature extraction and it has also been applied in many fields [19,20,21,22]. The discriminant deep autoencoders (DDAEs) proposed in [23] are used to enhance the recognition performance where there are few training samples. Stacked corrective autoencoders (SCAEs) are proposed in [24], which employ the average profile of each HRRP frame as the correction term. In [25], a novel robust variational autoencoder model (RVAE) is proposed to explore the latent representations of HRRP. In these applications, SAE is used for feature learning to obtain the hierarchical abstract representation of the target. In addition, to implement recognition, we need to add a classifier to the top encoding layer of SAE and softmax regression is usually chosen. The last step of training is to fine-tune the parameters of all layers to achieve the desired recognition performance. This process will take a lot of time. Replacing the softmax regression with the extreme learning machine (ELM) as a classifier can improve the training speed.
ELM [26] is a new learning algorithm for single hidden layer feedforward neural networks (SLFNs). Its network topology is the same as that of back propagation (BP) [27] neural networks. It is also composed of an input layer, a hidden layer, and an output layer. Although the network structure is the same, the training method of ELM is quite different from that of the BP [28,29,30]. The BP network needs to use gradient descent algorithms to solve the network weights through multiple iterations, while the ELM solves the output weights by randomly generating the input weights and hidden biases. ELM has been widely studied by many scholars [31,32,33,34,35,36] because of its characteristics of fewer training parameters, fast learning speed, and good generalization ability. The regularized ELM is studied in [31], and the experimental results show that the addition of a regularization term can enhance the robustness and generalization performance of ELM. An online sequential ELM (OS-ELM) is proposed in [32], which can learn the training data one by one or chunk by chunk. Error minimized extreme learning machine (EM-ELM) is studied in [33]. In [34], the researchers extend the ELM algorithm from the real domain to the complex domain and propose a fully complex ELM (C-ELM). The enhanced incremental extreme learning machine (EI-ELM) is also studied in [35]. ELM based on the kernel method [36] is faster and more generalized than the support vector machine (SVM) [37]. Because of its own advantages, ELM has been widely used in many aspects, such as image processing [38], clustering [39], traffic signal recognition [40], fault detection [41], and so on. However, the existing ELM [31,32,34,35,36,37,38,41] does not make better use of the target category information when dealing with the recognition tasks. To solve this problem, a regularized ELM based on target class information is proposed in this paper. Besides, due to the random selection of input weights and hidden biases, the ELM tends to need more hidden nodes to achieve better generalization performance [29,39,42], which makes the network structure complex. In this paper, SAE is used to optimize the input weights and hidden biases of ELM, which then achieves better results with fewer hidden layer nodes. The features of the proposed method are summarized as follows:
(a) The proposed model is “end to end”, the input is the original radar HRRP data, and the output is the target class.
(b) This paper proposes a combination of SAE and regularized ELM, which can improve the recognition performance by making full use of the advantages of SAE and ELM. Compared with the shallow learning algorithms such as PCA [3], MTL TSB-HMMs [6], ELM [26], and so on, the proposed algorithm can extract the inherent characteristics of the target. Since the network is not required to be fine-tuned, the proposed algorithm is faster than the other deep learning models [18,23,24,25].
(c) The proposed method does not only improve the training speed but also gets good performance when the training sample is small.
The rest of this paper is organized as follows: Section 2 introduces the relevant theoretical knowledge of SAE and ELM. In Section 3, we present the regularized ELM, then we also introduce the learning process of SAE-ELM. Experimental results are analyzed in Section 4, and in Section 5 the paper is summarized.

2. Theoretical Background

2.1. Description of HRRP

HRRP can be regarded as the amplitude of the coherent summations of the complex time returns from target scatters in each range cell [3], which represents the projection of the complex returned echoes from the target scattering centers onto the radar line-of-sight (LOS) [4]. The illustration of an HRRP sample from a plane target is shown in Figure 1. Since HRRP contains the target-important structural features such as target size and the distribution of scattering centers, etc., radar HRRP target recognition has drawn much attention from the radar automatic target recognition community [3,4,5,6,7].

2.2. Stacked Autoencoder

An autoencoder (AE) is an unsupervised learning algorithm. Figure 2 shows a simple model structure for an AE:
Given an unlabeled dataset { x ( i ) } i = 1 m , each of its training data x ( i ) is encoded by an encoder and the feature representation y ( i ) of the hidden layer can be obtained: y ( i ) = f θ ( x ( i ) ) = s ( W x ( i ) + b ) where θ = ( W , b ) is the network parameter, W is the weight matrix, b is the bias vector, and s ( x ) is the activation function; the sigmoid function is selected here and s ( x ) = 1 / ( 1 + e x ) . Then, the feature representation y ( i ) of the hidden layer is decoded by the decoder and the reconstruction vector z ( i ) can be obtained: z ( i ) = g θ ( y ( i ) ) = s ( W y ( i ) + b ) , where θ = ( W , b ) , W is the weight matrix and W = W T . In fact, the optimization of the model parameters is to minimize the reconstruction error [16]:
θ * , θ * = arg min θ , θ i = 1 m J ( x ( i ) , z ( i ) ) = arg min θ , θ i = 1 m J ( x ( i ) , g θ ( f θ ( x ( i ) ) ) )
where m is the sample number and J is the cost function. The expression for J is J ( x , z ) = 1 2 z x 2 .
For a dataset containing m samples, the total cost function is:
J 1 = [ 1 m i = 1 m ( 1 2 z i x i 2 ) ] + λ 2 l = 1 n l 1 i = 1 s l j = 1 s l + 1 ( W j i ( l ) ) 2
where W j i ( l ) is the connection weights between the i-th neurons of layer l and the j-th neurons of layer l + 1 ; n l and s l indicate the number of network layers and the number of neurons of layer l , respectively. The first part of Equation (2) is a mean squared error term and the second part is a weight decay term, which can be seen as a way to compromise between the small weights and minimized cost function [21]. The second term of Equation (2) is intended to prevent overfitting [19].
If the number of hidden layer nodes is large, and even more than the number of input layer nodes, the sparsity constraint needs to be added on the hidden units [19]. Hidden units are constrained to be zero most of the time when the activation function is selected as a sigmoid function [43]. This is motivated by the structure of the brain in which most of the neurons are inactive most of the time. By forcing the hidden units to have mostly zero activations/values, interesting representations can be learned. Then, the overall cost function is expressed as follows:
J s = J 1 + η j = 1 s 2 KL ( ρ ρ ^ j )
where the second part of Equation (3) represents the sparse penalty term and the penalty term used in this paper is based on Kullback-Leibler (KL) divergence [44]. KL indicates the relative entropy [24] between the two Bernoulli random variables with the mean of ρ and the mean of ρ ^ j , and KL ( ρ ρ ^ j ) = ρ log ρ ρ ^ j + ( 1 ρ ) log 1 ρ 1 ρ ^ j . If ρ ^ j = ρ , KL ( ρ ρ ^ j ) reaches the minimum value of 0, and if ρ ^ j approaches 0 or 1, the KL ( ρ ρ ^ j ) increases dramatically. s 2 is the number of neurons in the hidden layer. η is the weight of the sparsity penalty.
SAE is a neural network consisting of multiple layers of autoencoders, and the structure of an SAE is shown in Figure 3. We can use a greedy layer-wise training method to train SAE; that is, the output of each layer is wired to the input of the successive layer. Then, the BP algorithm is used to fine-tune the whole network.

2.3. Extreme Learning Machine

Given a set of N training datasets ( x i , t i ) where i = 1 , 2 , , N , x i = [ x i 1 , x i 2 , , x i n ] T R n , and t i = [ t i 1 , t i 2 , , t i m ] T R m , x i is an n-dimensional input vector and t i is the expected output. The output function of ELM with L hidden nodes is represented as follows:
i = 1 L β i g ( w i x j + b i ) = o j j = 1 , 2 , , N
where w i = [ w i 1 , w i 2 , , w i n ] T R n is the weight vector of input nodes to hidden nodes and b i is the bias of i-th hidden node; β i = [ β i 1 , β i 2 , , β i m ] R m is the weight vector between hidden nodes and the output nodes; g ( x ) is the activation function of the hidden layer; and o j is the output vector.
If the SLFNs with L hidden nodes can approximate the N samples with zero error, we know that Equation (4) can be converted to the following formula [26,45]:
{ i = 1 L β i g ( w i x 1 + b i ) = t 1     i = 1 L β i g ( w i x N + b i ) = t N
The above equations can be written as:
H β = T
where
H ( w 1 , , w L , b 1 , , b L , x 1 , , x N ) = [ g ( w 1 x 1 + b 1 ) g ( w L x 1 + b L )       g ( w 1 x N + b 1 ) g ( w L x N + b L ) ] N × L
β = [ β 1 T   β L T ] L × m T = [ t 1 T   t N T ] N × m
So training the SLFNs corresponds to finding the norm least-squares solution β ^ , which can be shown as follows:
β ^ = H + T
where H + is the Moore–Penrose generalized inverse [46,47] of hidden layer output matrix H .
Then, Equation (9) can be converted to:
β = { ( I C + H T H ) 1 H T T , N > L H T ( I C + H H T ) 1 T , N < L
where I is the unit matrix and C is the regularization coefficient.
ELM can also be explained using the optimization method. The ELM theory aims to reach the smallest training error H β T 2 and the smallest norm of the output weights β [28,31]. Then, the solution of Equation (6) can be obtained by:
Minimize :   L P E L M = 1 2 β 2 + C 1 2 i = 1 N ξ i 2 Subject   to :   h ( x i ) β = t i T ξ i T , i = 1 , , N
where ξ i is the training error vector of the m output nodes corresponding to training sample x i , and h ( x i ) is the hidden layer output vector of i-th sample x i . According to the Karush–Kuhn–Tucker (KKT) theorem [48], the same solution as Equation (10) can be obtained.
Thus, the learning steps of the ELM can be summarized as Algorithm 1:
Algorithm 1: ELM
Input: training sets { x i , t i } , ( x i R n , t i R m , i = 1 , 2 , , N ), activation function g ( x ) and hidden nodes L .
Output: output weight vector β
(1): Set random values to the input weights w i and the hidden layer biases b i ;
(2): Calculate the hidden layer output matrix H according to Equation (7);
(3): Calculate the output weight vector β according to Equation (9).

3. Stacked Autoencoder-Regularized Extreme Learning Machine

As we know the sample data has similar attributes and distribution features, we can use the similar relationships to enhance the generalization performance of ELM. Therefore, in this section, we propose a regularized ELM based on the class information of the target. Optimizing the output weights by maximizing the within-class scatter degree and by minimizing the inter-class scatter degree can make the ELM have better recognition and generalization ability. In addition, due to the random selection of input weights and hidden biases, ELM tends to need more hidden nodes to achieve better generalization performance, which makes the network structure complex. To address this issue, SAE is used to optimize the input weights and hidden biases of ELM; this achieves better results with fewer hidden layer nodes.

3.1. Regularized ELM Based on the Class Information of the Target

Given a set of sample sets { x i ( j ) ; j = 1 , 2 , , c ; i = 1 , 2 , , n j } , the number of classes is c and the ω j class contains n j samples. The inter-class scatter matrix of class ω j is defined as
S W ( j ) = 1 n j i = 1 n j ( x i ( j ) m j ) ( x i ( j ) m j ) T , j = 1 , 2 , , c
where m j is the mean of ω j class samples and m j = 1 n j i = 1 n j x i ( j ) , j = 1 , 2 , , c .
The total inter-class scatter matrix is defined as
S W = j = 1 c n j N S W ( j )
The within-class scatter matrix is defined as
S B = j = 1 c n j N ( m j m ) ( m j m ) T
where m is the mean of all samples and m = 1 N i = 1 N x i .
To improve the recognition performance, we should maximize the within-class scatter matrix and minimize the inter-class scatter matrix [49]. Therefore, we define the matrix S as shown below:
S = S W τ S B   ( 0 < τ 1 )
Then, the optimization formula of regularized ELM can be written as:
Minimize :   L P E L M = 1 2 β T S β + C 1 2 i = 1 N ξ i 2 Subject   to :   h ( x i ) β = t i T ξ i T , i = 1 , , N
We can solve the above problem by defining the Lagrange function:
min :   L E L M = 1 2 β T S β + C 1 2 i = 1 N ξ i 2 i = 1 N j = 1 m α i , j ( h ( x i ) β j t i , j + ξ i , j )
then
L E L M β j = 0 β S = H T α L E L M ξ i = 0 α i = C ξ i , i = 1 , , N L E L M α i = 0 h ( x i ) β t i T + ξ i T = 0 , i = 1 , , N
where α i = [ α i , 1 , , α i , m ] T and α = [ α 1 , , α N ] T .
Then, the solution to Equation (16) is:
β = { ( S C + H T H ) 1 H T T , N > L H T ( S C + H H T ) 1 T , N < L
Thus, the learning steps of the regularized ELM can be summarized as Algorithm 2:
Algorithm 2: Regularized ELM
Input: training sets { x i , t i } , ( x i R n , t i R m , i = 1 , 2 , , N ), activation function g ( x ) and hidden nodes L .
Output: output weight vector β
(1): Calculate S W and S B , then calculate S according to Equation (15);
(2): Set random values to the input weights w i and the hidden layer biases b i ;
(3): Calculate the hidden layer output matrix H according to Equation (7);
(4): Calculate the output weight vector β according to Equation (19).

3.2. SAE-ELM

In order to implement recognition, we need to add a classifier to the top encoding layer of SAE. In this section, we propose that using ELM instead of softmax as a classifier can effectively improve the network training speed. In addition, we can get the appropriate ELM network parameters by training SAE. The SAE–ELM system architecture is shown in Figure 4, and the illustration of the structure is shown in Figure 5. The learning process of SAE–ELM is as follows:
(1) Establish the first layer of AE network and, as described in Section 2.2, use the gradient descent method to train the network. Then, we can obtain the output H 1 of the first hidden layer and the network parameters θ 1 . H 1 is the characteristic representation of the input data and θ 1 = ( W 1 , b 1 ) .
(2) Establish the second layer of the AE network. The first layer output H 1 is input as the second layer. We use the gradient descent method to train the network. Then, the output H 2 of the second hidden layer and the network parameters θ 2 are available and θ 2 = ( W 2 , b 2 ) .
(3) Establish the third layer of the AE network to determine the parameters of ELM. ELM not only has a faster learning speed than the traditional learning methods but it also has a good generalization performance. However, ELM needs more hidden nodes than conventional tuning-based learning algorithms due to the random set of input weights and hidden biases. Therefore, we establish the third layer of the AE network to determine the input weights and hidden biases for ELM. Similar to step (2), the output H 3 of the third hidden layer and network parameters θ 3 = ( W 3 , b 3 ) can be obtained. We can utilize W 3 as the input weights and b 3 as the hidden biases of ELM, then the hidden layer output matrix of ELM is H 3 .
(4) Establish the ELM network as a classifier. The input is H 2 , the input weights and hidden biases are θ 3 = ( W 3 , b 3 ) , and the hidden layer output matrix is H 3 . Then, as described in Section 3.1, the output weight vector β can be calculated according to Equation (19).

4. Experimental Results and Discussion

In this section, we will verify the effectiveness of the proposed algorithm. The experiments were performed on an Intel(R) Core(TM) 3.60 GHz CPU with 8 GB of RAM and the MATLAB R2013a environment.
In this section, we utilize measured radar HRRP data from three real airplanes that are measured by a C-band radar with a center frequency of 5.52 GHz and a bandwidth of 400 MHz to validate the effectiveness of the proposed method. The An-26 is a medium-sized propeller airplane, the Yark-42 is a large and medium-sized jet airplane, and the Citation business jet is a small-sized jet airplane. The three aircraft models are shown in Figure 6. The detailed size of each airplane and the parameters of the measured radar are listed in Table 1. In our experiments, each aircraft target has 26,000 HRRP samples and the measured HRRP is a 256-dimensional vector.
In order to verify the validity of the algorithm proposed in this paper, we compared it with the commonly used methods: PCA [3], MTL TSB-HMMS [6], ELM [26], SAE [21], and DDAEs [23]. The activation function of the hidden layer of ELM is sigmoid and G ( a , b , x ) = 1 / ( 1 + exp ( ( a x + b ) ) ) . The regularization coefficient C is 0.2. The number of hidden nodes of ELM is 1500. Due to the sample dimension being 256, we set the number of nodes in the visible layer of deep architecture to 256. It is well known that a more abstract feature representation can be obtained with an increase in the network depth. However, too many layers can make the network difficult to train effectively and brings in more parameters to learn. Through the analysis of the experimental data and task requirements, we found that three is a good choice for the number of hidden layers. Therefore, we set the number of hidden layers to be three and the number of nodes in the hidden layers as 1500-500-50, respectively. From Figure 7 we can see that the mean square error (MSE) of each layer reconstruction of the network model decreases with an increase of iterations. When the number of iterations is 25, the MSE is less than 0.003. Therefore, in order to speed up the training, we set the number of iterations in the network to 25.
Before network training, data pre-processing is needed to solve the amplitude-scale and time-shift sensitivities. According to the previous study [3,5,6,7], we usually use the energy normalization method and time-shift compensation algorithm to cope with the above issues.
Figure 8 shows the range profiles of pre-processed aircraft targets. In the non-cooperative situation, such as at the battle with time, the amount of data under the test is usually huge, but the training data is limited. This is because the radar system cannot be guaranteed to detect and track the non-cooperative targets for a long period of time, which will cause HRRP data to be lost or not observed. Therefore, it is very important to study the generalization performance of the model and obtain good recognition performance under the conditions of fewer training samples.
As is shown in Figure 9, as the number of training samples increases, the classification accuracy of different algorithms also increases. However, deep architecture algorithms (e.g., SAE, DDAEs, and the proposed method) are more accurate than shallow architecture algorithms (e.g., PCA, MTL TSB-HMMS, and ELM). The traditional recognition algorithms rely on the experience of the researchers and require a complete set of training samples to ensure excellent recognition performance. Because of the shallow architecture, these algorithms cannot effectively separate the intrinsic class information of the target from some external factors in the feature space. The depth structure algorithms lose the inherent class information of the target as little as possible while demodulating the coupling relationship between various factors layer by layer. More intuitively, the low-level features in the deep network are usually distributed and can be shared among different classes, while the high-level features are usually more abstract and more separable. Therefore, better generalization performance is a great advantage of deep networks. Due to the proposed method not only obtaining the deep feature representation of radar HRRP but also making better use of the target category information, the classification performance of the proposed method is better than that of SAE and DDAEs. In addition, when the training sample is smaller, the classification performance of the proposed method is better than the other algorithms, which shows that the proposed method has better generalization performance. When the number of training samples for each target is 3500, the classification accuracy of different algorithms is listed in Table 2. It can be seen from the table that when the number of training samples is 3500, the accuracy of the proposed algorithm reaches 95.01%, which is 0.22% higher than that of the DDAE algorithm, and 1.5% higher than that of the SAE. The accuracy of the shallow structure algorithms is not more than 90%. It can be concluded that the proposed method can obtain better classification performance when there is only a small amount of training samples available.
As shown in Table 3, the training time of SAE, DDAEs, and the proposed method are compared. The proposed method is almost five times faster than SAE in training time; that is because we need to add the softmax regression classifier to the top encoding layer of SAE, and the last step of training is to fine-tune the parameters of all layers to achieve the desired classification performance. This process will take a lot of time. The proposed method adds ELM with faster learning speed and less required tuning parameters to the top layer of SAE as a classifier. The proposed method does not need to fine-tune the parameters of all layers, thus reducing the network training steps and training time. SAE and DDAEs are similar in training time because their network structures are the same.
It can be seen from Figure 10 that the classification accuracy of ELM becomes much better as the hidden nodes increase. When the number of hidden nodes is 1500, the classification accuracy is 89.01%. When the number of hidden nodes increases to 4000, ELM reaches an accuracy of 90.01%. After that, the value is almost unchanged all the time because the ELM is in an over-fitting state. Therefore, we know that in order to get a better classification effect, ELM needs more hidden nodes, which will make the network structure more complex. As we know from Table 2, only 50 hidden nodes are required to obtain an accuracy of 95.01% when the proposed method uses regularized ELM for classification. Therefore, the proposed method can effectively reduce the hidden nodes of ELM and simplify the network structure.

5. Conclusions

In this paper, we have proposed a novel radar HRRP target recognition method based on SAE and regularized ELM. SAE, as an important component of the deep learning structure, can extract deep features and mine the essential information of radar HRRP, which has a beneficial effect on recognition. ELM is also useful for recognition because of its fast learning speed and good generalization performance. Experimental results show that the proposed method does not only reduce the network training time but also makes the ELM achieve high recognition accuracy under the condition of using fewer hidden nodes. In addition, when there is only a small amount of training samples available, the proposed method can also obtain good recognition performance. However, we also know that in real situations the training samples are usually obtained under the condition of high signal-to-noise ratio (SNR) via some cooperative measurement experiments, while the test samples are usually achieved in the non-cooperative circumstance where the high SNR cannot be guaranteed due to the severe measurement conditions. Thus, it is important to optimize the proposed method to match the noise level of the received test samples in the recognition stage. Stacked denoising sparse autoencoder (sDSAE) can effectively eliminate the influence of noise. Therefore, in the near future, we will consider combining sDSAE with ELM to solve this problem.

Acknowledgments

The work in this paper has been supported by the National Science Foundation of China (61422114 and 61501481) and the Natural Science Fund for Distinguished Young Scholars of Hunan Province under Grant No. 2015JJ1003.

Author Contributions

F.Z. proposed the SAE-ELM for radar target recognition and completed this manuscript. Y.L. revised this manuscript. K.H.; S.Z., and Z.Z. provided helpful advice.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Liu, Y.X.; Zhu, D.K.; Li, X.; Zhuang, Z.W. Micromotion characteristic acquisition based on wideband radar phase. IEEE Trans. Geosci. Remote Sens. 2014, 52, 3650–3657. [Google Scholar] [CrossRef]
  2. Vespe, M.; Baker, C.J.; Griffiths, H.D. Radar target classification using multiple perspectives. IET Radar Sonar Navig. 2007, 1, 300–307. [Google Scholar] [CrossRef]
  3. Du, L.; Liu, H.W.; Bao, Z.; Zhang, J. Radar automatic target recognition using complex high-resolution range profiles. IET Radar Sonar Navig. 2007, 1, 18–26. [Google Scholar] [CrossRef]
  4. Du, L.; Wang, P.H.; Liu, H.W.; Pan, M.; Chen, F.; Bao, Z. Bayesian spatiotemporal multitask learning for radar HRRP target recognition. IEEE Trans. Signal Process. 2011, 59, 3182–3196. [Google Scholar] [CrossRef]
  5. Liu, J.; Fang, N.; Xie, Y.J.; Wang, B.F. Scale-space theory-based multi-scale features for aircraft classification using HRRP. Electron. Lett. 2016, 52, 475–477. [Google Scholar] [CrossRef]
  6. Zhang, X.D.; Shi, Y.; Bao, Z. A new feature vector using selected bispectra for signal classification with application in radar target recognition. IEEE Trans. Signal Process. 2001, 49, 1875–1885. [Google Scholar] [CrossRef]
  7. Du, L.; Liu, H.W.; Wang, P.H.; Feng, B.; Pan, M.; Bao, Z. Noise robust radar HRRP target recognition based on multitask factor analysis with small training data size. IEEE Trans. Signal Process. 2012, 60, 3546–3559. [Google Scholar]
  8. López-Rodríguez, P.; Escot-Bocanegra, D.; Fernández-Recio, R.; Bravo, I. Non-cooperative target recognition by means of singular value decomposition applied to radar high resolution range profiles. Sensors 2015, 15, 422–439. [Google Scholar] [CrossRef] [PubMed]
  9. López-Rodríguez, P.; Escot-Bocanegra, D.; Fernández-Recio, R.; Bravo, I. Non-cooperative identification of civil aircraft using a generalised mutual subspace method. IET Radar Sonar Navig. 2016, 10, 186–191. [Google Scholar] [CrossRef]
  10. Slomka, J.S. Features for high resolution radar range profile based ship classification. In Proceedings of the Fifth International Symposium on Signal Processing and its Applications (ISSPA), Brisbane, Australia, 22–25 August 1999; pp. 329–332. [Google Scholar]
  11. Christopher, M.; Alireza, K. Maritime ATR using classifier combination and high resolution range profiles. IEEE Trans. Aerosp. Electron. Syst. 2011, 47, 2558–2573. [Google Scholar]
  12. Li, J.; Stoica, P. Efficient mixed-spectrum estimation with application to feature extraction. IEEE Trans. Signal Process. 1996, 42, 281–295. [Google Scholar]
  13. Pei, B.; Bao, Z. Multi-aspect radar target recognition method based on scattering centers and HMMs classifiers. IEEE Trans. Aerosp. Electron. Syst. 2005, 41, 1067–1074. [Google Scholar]
  14. Jiang, Y.; Han, Y.B.; Sheng, W.X. Target recognition of radar HRRP using manifold learning with feature weighting. In Proceedings of the 2016 IEEE International Workshop on Electromagnetics: Applications and Student Innovation Competition (iWEM), Nanjing, China, 16–18 May 2016; pp. 1–3. [Google Scholar]
  15. Li, L.; Liu, Z. Noise-robust HRRP target recognition method via sparse-low-rank representation. Electron. Lett. 2017, 53, 1602–1604. [Google Scholar] [CrossRef]
  16. Hinton, G.E.; Salakhutdinov, R.R. Reducing the dimensionality of data with neural networks. Science 2006, 313, 504–507. [Google Scholar] [CrossRef] [PubMed]
  17. Lecun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef] [PubMed]
  18. Pan, M.; Jiang, J.; Kong, Q.P.; Jiang, J.; Sheng, Q.H.; Zhou, T. Radar HRRP target recognition based on t-SNE segmentation and discriminant deep belief network. IEEE Geosci. Remote Sens. Lett. 2017, 14, 1609–1613. [Google Scholar] [CrossRef]
  19. Sun, W.J.; Shao, S.Y.; Zhao, R.; Yan, R.Q.; Zhang, X.W.; Chen, X.F. A sparse auto-encoder-based deep neural network approach for induction motor faults classification. Measurement 2016, 89, 171–178. [Google Scholar] [CrossRef]
  20. Zhao, F.X.; Liu, Y.X.; Huo, K. Radar target recognition based on stacked denoising sparse autoencoder. Chin. J. Radar 2017, 6, 149–156. [Google Scholar]
  21. Kang, M.; Ji, K.; Leng, X.; Xing, X.; Zou, H. Synthetic aperture radar target recognition with feature fusion based on a stacked autoencoder. Sensors 2017, 17, 192. [Google Scholar] [CrossRef] [PubMed]
  22. Protopapadakis, E.; Voulodimos, A.; Doulamis, A.; Doulamis, N.; Dres, D.; Bimpas, M. Stacked autoencoders for outlier detection in over-the-horizon radar signals. Comput. Intell. Neurosci. 2017. [Google Scholar] [CrossRef]
  23. Pan, M.; Jiang, J.; Li, Z.; Cao, J.; Zhou, T. Radar HRRP recognition based on discriminant deep autoencoders with small training data size. Electron. Lett. 2016, 52, 1725–1727. [Google Scholar]
  24. Feng, B.; Chen, B.; Liu, H.W. Radar HRRP target recognition with deep networks. Pattern Recognit. 2017, 61, 379–393. [Google Scholar] [CrossRef]
  25. Zhai, Y.; Chen, B.; Zhang, H.; Wang, Z.J. Robust variational auto-encoder for radar HRRP target recognition. In Proceedings of the International Conference on Intelligent Science and Big Data Engineering, Dalian, China, 22–23 September 2017; pp. 356–367. [Google Scholar]
  26. Huang, G.B.; Zhu, Q.Y.; Siew, C.K. Extreme learning machine: Theory and applications. Neurocomputing 2006, 70, 489–501. [Google Scholar] [CrossRef]
  27. Mou, R.; Chen, Q.; Huang, M. An improved BP neural network and its application. In Proceedings of the 2012 Fourth International Conference on Computational and Information Sciences, Chongqing, China, 10–12 August 2012; pp. 477–480. [Google Scholar]
  28. Tang, J.; Deng, C.; Huang, G.B. Extreme learning machine for multilayer perceptron. IEEE Trans. Neural Netw. Learn. Syst. 2016, 27, 809–821. [Google Scholar] [CrossRef] [PubMed]
  29. Huang, G.; Huang, G.B.; Song, S.J.; You, K. Trends in extreme learning machines: A review. Neural Netw. 2015, 61, 32–48. [Google Scholar] [CrossRef] [PubMed]
  30. Wang, H.; Fan, W.; Sun, F.W.; Qian, X.J. An adaptive ensemble model of extreme learning machine for time series prediction. In Proceedings of the 12th International Computer Conference on Wavelet Active Media Technology and Information Processing (ICCWAMTIP), Chengdu, China, 18–20 December 2015; pp. 80–85. [Google Scholar]
  31. Huang, G.B.; Zhou, H.M.; Ding, X.J.; Zhang, R. Extreme learning machine for regression and multiclass classification. IEEE Trans. Syst. Man Cybern. B Cybern. 2012, 42, 513–529. [Google Scholar] [CrossRef] [PubMed]
  32. Liang, N.Y.; Huang, G.B.; Saratchandran, P.; Sundararajan, N. A fast and accurate online sequential learning algorithm for feedforward networks. IEEE Trans. Neural Netw. 2006, 17, 1411–1423. [Google Scholar] [CrossRef] [PubMed]
  33. Feng, G.R.; Huang, G.B.; Lin, Q.P.; Gay, R. Error minimized extreme learning machine with growth of hidden nodes and incremental learning. IEEE Trans. Neural Netw. 2009, 20, 1352–1357. [Google Scholar] [CrossRef] [PubMed]
  34. Li, M.B.; Huang, G.B.; Saratchandran, P.; Sundararajan, N. Fully complex extreme learning machine. Neurocomputing 2005, 68, 306–314. [Google Scholar] [CrossRef]
  35. Huang, G.B.; Chen, L. Enhanced random search based incremental extreme learning machine. Neurocomputing 2008, 71, 3460–3468. [Google Scholar] [CrossRef]
  36. Pal, M.; Maxwell, A.E.; Warner, T.A. Kernel-based extreme learning machine for remote-sensing image classification. Remote Sens. Lett. 2013, 4, 853–862. [Google Scholar] [CrossRef]
  37. Liu, J.; Fang, N.; Xie, Y.J.; Wang, B.F. Radar target classification using support vector machine and subspace methods. IET Radar Sonar Navig. 2015, 9, 632–640. [Google Scholar] [CrossRef]
  38. Liu, N.; Wang, H. Evolutionary extreme learning machine and its application to image analysis. J. Signal Proc. Syst. Signal Image Video Technol. 2013, 73, 73–81. [Google Scholar] [CrossRef]
  39. Ding, S.F.; Zhang, N.; Zhang, J.; Xu, X.Z.; Shi, Z.Z. Unsupervised extreme learning machine with representational features. Int. J. Mach. Learn. Cybern. 2017, 8, 587–595. [Google Scholar] [CrossRef]
  40. Zeng, Y.; Xu, X.; Shen, D.; Fang, Y.; Xiao, Z. Traffic sign recognition using kernel extreme learning machines with deep perceptual features. IEEE Trans. Intell. Trans. Syst. 2017, 18, 1647–1653. [Google Scholar] [CrossRef]
  41. Yang, X.Y.; Pang, S.; Shen, W.; Lin, X.S.; Jiang, K.Y.; Wang, Y.H. Aero engine fault diagnosis using an optimized extreme learning machine. Int. J. Aerosp. Eng. 2016. [Google Scholar] [CrossRef]
  42. Zhao, F.X.; Liu, Y.X.; Huo, K.; Zhang, Z.S. Radar Target Classification Using an Evolutionary Extreme Learning Machine Based on Improved Quantum-Behaved Particle Swarm Optimization. Math. Probl. Eng. 2017. [Google Scholar] [CrossRef]
  43. Kumar, V.; Nandi, G.C.; Kala, R. Static hand gesture recognition using stacked denoising sparse autoencoders. In Proceedings of the Seventh International Conference on Contemporary Computing (IC3), Noida, India, 7–9 August 2014; pp. 99–104. [Google Scholar]
  44. Kullback, S.; Leibler, R.A. On information and sufficiency. Ann. Math. Stat. 1951, 22, 79–86. [Google Scholar] [CrossRef]
  45. Peng, C.; Yan, J.; Duan, S.K.; Wang, L.D.; Jia, P.F.; Zhang, S.L. Enhancing electronic nose performance based on a novel QPSO-KELM model. Sensors 2016, 16, 520. [Google Scholar] [CrossRef] [PubMed]
  46. Rao, C.R.; Mitra, S.K. Generalized Inverse of Matrices and Its Applications; Wiley: New York, NY, USA, 1971. [Google Scholar]
  47. Serre, D. Matrices: Theory and Applications; Springer: New York, NY, USA, 2002. [Google Scholar]
  48. Fletcher, R. Practical Methods of Optimization: Constrained Optimization; Wiley: New York, NY, USA, 1981. [Google Scholar]
  49. Yan, D.Q.; Chu, Y.H.; Zhang, H.Y.; Liu, D.S. Information discriminative extreme learning machine. Soft Comput. 2016, 1–13. [Google Scholar] [CrossRef]
Figure 1. Illustration of a high-resolution range profile (HRRP) sample from a plane target.
Figure 1. Illustration of a high-resolution range profile (HRRP) sample from a plane target.
Sensors 18 00173 g001
Figure 2. The structure of an autoencoder (AE).
Figure 2. The structure of an autoencoder (AE).
Sensors 18 00173 g002
Figure 3. The structure of an AE and a stacked autoencoder (SAE). (a) A three-layer AE; (b) An SAE composed of two autoencoders.
Figure 3. The structure of an AE and a stacked autoencoder (SAE). (a) A three-layer AE; (b) An SAE composed of two autoencoders.
Sensors 18 00173 g003
Figure 4. The stacked autoencoder-extreme learning machine (SAE-ELM) system architecture.
Figure 4. The stacked autoencoder-extreme learning machine (SAE-ELM) system architecture.
Sensors 18 00173 g004
Figure 5. Illustration of the structure of SAE-ELM.
Figure 5. Illustration of the structure of SAE-ELM.
Sensors 18 00173 g005
Figure 6. The Imagery targets. (a) An-26 airfreighter; (b) Yark-42; (c) Citation business jet.
Figure 6. The Imagery targets. (a) An-26 airfreighter; (b) Yark-42; (c) Citation business jet.
Sensors 18 00173 g006
Figure 7. Reconstruction error of SAE pre-training.
Figure 7. Reconstruction error of SAE pre-training.
Sensors 18 00173 g007
Figure 8. Range profiles of the three real airplanes. (a) An-26 airfreighter; (b) Yark-42; (c) Citation business jet.
Figure 8. Range profiles of the three real airplanes. (a) An-26 airfreighter; (b) Yark-42; (c) Citation business jet.
Sensors 18 00173 g008
Figure 9. Classification accuracy against different HRRP training samples.
Figure 9. Classification accuracy against different HRRP training samples.
Sensors 18 00173 g009
Figure 10. Accuracy of ELM in different hidden nodes.
Figure 10. Accuracy of ELM in different hidden nodes.
Sensors 18 00173 g010
Table 1. Parameters of the airplanes and radar in the inverse synthetic aperture radar (ISAR) experiment.
Table 1. Parameters of the airplanes and radar in the inverse synthetic aperture radar (ISAR) experiment.
Radar ParametersCenter Frequency5520 MHz
Bandwidth400 MHz
AirplaneLength (m)Width (m)Height (m)
An-2623.8029.209.83
Yark-4236.3834.889.83
Citation business jet14.4015.904.57
Table 2. The classification accuracy comparison of different methods.
Table 2. The classification accuracy comparison of different methods.
MethodClassification Accuracy (%)
PCA74.38
MTL TSB-HMMS86.87
ELM89.01
SAE93.51
DDAEs94.79
Proposed method95.01
Table 3. The training time of the different methods.
Table 3. The training time of the different methods.
MethodTraining Time (s)
SAE624.73
DDAEs625.14
Proposed method106.67

Share and Cite

MDPI and ACS Style

Zhao, F.; Liu, Y.; Huo, K.; Zhang, S.; Zhang, Z. Radar HRRP Target Recognition Based on Stacked Autoencoder and Extreme Learning Machine. Sensors 2018, 18, 173. https://doi.org/10.3390/s18010173

AMA Style

Zhao F, Liu Y, Huo K, Zhang S, Zhang Z. Radar HRRP Target Recognition Based on Stacked Autoencoder and Extreme Learning Machine. Sensors. 2018; 18(1):173. https://doi.org/10.3390/s18010173

Chicago/Turabian Style

Zhao, Feixiang, Yongxiang Liu, Kai Huo, Shuanghui Zhang, and Zhongshuai Zhang. 2018. "Radar HRRP Target Recognition Based on Stacked Autoencoder and Extreme Learning Machine" Sensors 18, no. 1: 173. https://doi.org/10.3390/s18010173

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop