Next Article in Journal
Genetic Programming Guided Mapping of Forest Canopy Height by Combining LiDAR Satellites with Sentinel-1/2, Terrain, and Climate Data
Next Article in Special Issue
Radio Frequency Interference Mitigation in Synthetic Aperture Radar Data Based on Instantaneous Spectrum Forward Consecutive Mean Excision
Previous Article in Journal
Double-Factor Tensor Cascaded-Rank Decomposition for Hyperspectral Image Denoising
Previous Article in Special Issue
GPU-Accelerated Signal Processing for Passive Bistatic Radar
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Radar Emitter Signal Intra-Pulse Modulation Open Set Recognition Based on Deep Neural Network

School of Electronic Engineering, Xidian University, Xi’an 710071, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2024, 16(1), 108; https://doi.org/10.3390/rs16010108
Submission received: 7 November 2023 / Revised: 15 December 2023 / Accepted: 24 December 2023 / Published: 26 December 2023
(This article belongs to the Special Issue Advanced Radar Signal Processing and Applications)

Abstract

:
Radar emitter signal intra-pulse modulation recognition is important for modern electronic reconnaissance systems to analyze target radar systems. In the actual environment, the intra-pulse modulations of the sampled radar emitter signals contain not only the known types in the library but also the unknown types. Therefore, the existing recognition methods, which are based on a closed set, cannot recognize the unknown samples. In order to solve this problem, in this paper, we proposed a method for radar emitter signal intra-pulse modulation open set recognition. The proposed method could classify the known modulations and identify the unknown modulation by using an original deep neural network-based recognition model trained on a closed set, estimating the signal-to-noise ratio, and calculating the reconstruction loss by an encoder–decoder model. For a given sample, the original deep neural network-based recognition model will label it as a certain known class temporarily. By estimating the SNR of the sample and calculating the reconstruction loss by inputting the sample to the corresponding encoder–decoder model related to the temporary predicted known class, whether the sample belongs to the predicted temporary known class or the unknown class will be confirmed. Experiments were conducted with five different openness conditions. The experimental results indicate that the proposed method has good performance on radar emitter signal intra-pulse modulation open set recognition.

1. Introduction

The modern electromagnetic environment has become more complex with the advancement of modern radar systems [1], and the number of types of intra-pulse modulation of radar emitter signals is increasing. Intra-pulse modulation recognition technologies are important for modern electronic reconnaissance systems, such as electronic support measure (ESM) systems, electronic intelligence (ELINT) systems, and radar warning receivers (RWRs) [2,3,4]. Accurately recognizing the intra-pulse modulation of radar emitter signals could increase the reliability of analyzing the function and information of the target radar.
Due to the rapid development of deep learning, recently, many researchers have combined deep learning methods and radar technologies to recognize intra-pulse modulations of radar emitter signals. In [5], Qu et al. preprocessed time-domain signals to obtain time-frequency images (TFIs) by using a Cohen class time-frequency distribution. Then, they designed a convolutional neural network (CNN) to recognize the intra-pulse modulations. In [6], Yu et al. used short-time Fourier transformation (STFT) to convert radar signals into TFIs, and then trained a CNN model to recognize their intra-pulse modulation. In [7], Yuan et al. proposed a 1-D selective kernel convolutional neural network to recognize eleven types of intra-pulse modulations based on the frequency-domain sequences of radar emitter signals. Wu et al. [8] combined a 1-D CNN and attention mechanism to recognize seven types of radar emitter signal intra-pulse modulations in the time domain. Si et al. [9] used a smooth pseudo-Wigner–Ville distribution transformation to covert dual-component radar signals in the time domain into TFIs and accomplished multi-label recognition tasks by using EfficientNet [10]. Similarly, Wan et al. [11] proposed a novel modulation recognition method of multi-component radar signals based on the deep convolutional neural network with a convolutional block attention module (DCNN-CBAM) to recognize multi-component radar signals. In [12], Cai et al. proposed a self-supervised learning framework with virtual adversarial training, which overcomes the limitation that the performance of radar signal intra-pulse modulation recognition mainly relies on a large number of labeled samples. Unlike most of the fully supervised situations, Yuan et al. [13] focused on a more common situation, where a very limited number of labeled samples and a large number of unlabeled samples are provided, and proposed a semi-supervised intra-pulse modulation recognition method.
However, there are still some problems in the above intra-pulse modulation recognition methods. The existing intra-pulse modulation recognition methods focus on the situation that the types of intra-pulse modulation in the testing dataset are the same types of intra-pulse modulation in the training dataset. Therefore, this type of intra-pulse modulation recognition is closed set recognition (CSR). Based on this assumption, these deep neural network-based methods perform well. But in the real environment, intra-pulse modulations are varied. Except for the known intra-pulse modulation types, unknown types emerge from time to time. These existing deep neural network-based CSR methods will naturally but mistakenly recognize an unknown type of intra-pulse modulation as one of the known types. Although the occurrence probability of this situation could be reduced by collecting as many types of intra-pulse modulations as possible, their collection for a training library is difficult as intra-pulse modulations are mainly used in the military field. In practical applications, it is important to not only classify the known classes but also label the unknown classes, which is called open set recognition (OSR). The OSR of radar emitter signal intra-pulse modulations is a tough challenge of great significance for the analysis of the information of a target radar emitter.
Recently, OSR methods have been shifted to the deep learning field. In [14], Hendrycks et al. presented a method using softmax distributions to detect misclassified and out-of-distribution samples, which provides the possibility of OSR in a deep neural network (DNN). The most representative research is that of Bendale et al. [15], who proposed the OpenMax approach based on the theory of the Weibull distribution, which replaced the softmax layer with an OpenMax layer to reassign the value of the activation vector. In the field of radar signal recognition, many studies have been conducted. For example, in [16], Sun et al. proposed a pseudo-signal generation and recognition neural network to identify specific emitters under an open set condition. Zhou et al. [17] presented two models for radar jamming OSR that can accurately classify known jamming and distinguish unknown jamming.
With increasingly complex electromagnetic environments, intra-pulse modulation recognition methods should be able to deal with an open set condition. These methods should not only classify the known intra-pulse modulations with high accuracy, but also have the ability to identify the unknown intra-pulse modulations correctly. Due to the great success of DNN-based methods in radar emitter signal intra-pulse modulation recognition, we proposed a DNN-based method to solve the OSR problem in intra-pulse modulation. This method does not change the existing well-trained recognition model in a closed set condition. Instead, in this method, taking the number of known intra-pulse modulations, we design the same number of encoder–decoder models to reconstruct the TFIs in the same class through compression and decompression. By estimating the signal-to-noise ratio (SNR) for each sample in a validation dataset, using the corresponding encoder–decoder model to calculate the reconstruction loss for the samples in the validation dataset, the thresholds with SNR for classifying known and unknown intra-pulse modulation could be set based on different recognition accuracies of the known classes. Combined the recognition results from the existing well-trained recognition model with the SNR of the testing samples, thresholds with SNR, and reconstruction loss values, the original recognition result could be revised.
The main contribution of this paper is that unlike most of the closed set situations, we focus on a more common open set situation, where the unknown samples are included for recognition during the testing session. Another contribution is that our method is based on the DNN with a closed set situation, which is trained only with known samples. Through estimating the SNR and calculating the reconstruction loss for testing the sample based on the proposed encoder–decoder model, the recognition result could be confirmed as a certain known class or as an unknown class.
This paper is organized as follows: In Section 2, the proposed method is described in detail. The dataset and experimental settings are shown in Section 3. The extensive experiments and corresponding analysis are described in Section 4. The ablation study is given in Section 5. Finally, the conclusion is presented in Section 6.

2. Materials and Methods

OSR of intra-pulse modulation refers to a case in which the samples used for the training model are in a certain number of classes. When the model is applied for a testing session, the types of intra-pulse modulation for the testing samples not only include the known classes from the training library, but also the unknown classes.
Let X t r a i n = ( x i , y i ) : i ( 1 , 2 , , K t r a i n ) be a known dataset in the library, where x i and y i are the known samples and the corresponding labels, and K t r a i n is the number of total samples. The label of y i belongs to { 1 , 2 , , N } , where N is the number of total known classes. For the testing dataset X t e s t = ( x j t e s t , y j t e s t ) : j ( 1 , 2 , , K t e s t ) , where x j t e s t and y j t e s t are the testing samples and the corresponding labels, and K t e s t is the number of total samples. In this case, the label of y j t e s t belongs to { 1 , 2 , , N , N + 1 } , where the category of { 1 , 2 , , N } refers to the known classes and the category of { N + 1 } refers to the unknown class. Therefore, a good model for intra-pulse modulation OSR should have good accuracy not only in classifying samples from known classes, but also in identifying samples from unknown classes.
In this paper, the method for intra-pulse modulation OSR mainly consists of 3 steps:
Step 1: Data preprocessing and original recognition model training.
(1)
Split the known samples dataset into a training dataset and validation dataset. Note that both the training dataset and validation dataset do not contain unknown samples.
(2)
Estimate the SNR for each sample in the validation dataset.
(3)
Convert the training samples and validation samples into TFIs by using STFT.
(4)
Train an original recognition model in the CSR way with TFIs.
Step 2: Encoder–decoder model training and threshold setting.
(1)
Select the samples from one known class and only use these TFIs to train the encoder–decoder model. Then, use the corresponding encoder–decoder model to calculate the reconstruction loss of the corresponding validation TFIs. As the SNR of the validation samples has been obtained, by setting the interval of SNR, the corresponding reconstruction loss in each interval of SNR could be counted. In this paper, the reconstruction loss is mean square error (MSE). Note that the number of encoder–decoder models is the same as the number of known classes. For example, if there are 11 known classes, then there will be 11 corresponding encoder–decoder models, where each model is only responsible for reconstructing one class.
(2)
Set the threshold for reconstruction loss in each interval of SNR. The details of the threshold setting are introduced as follows: For a known class, we could calculate the reconstruction loss of its validation samples from each SNR interval. Assuming that each SNR interval contains 400 samples, for example, then, we sort these 400 samples’ reconstruction losses from a low value to a high value for each interval. Then we set, for example, the top highest 1%, 2%, …, 25%, etc. reconstruction losses of these 400 samples as the threshold. Similarly, we could get the thresholds for the other known classes based on the above operation. Note that the ratio (1%, 2%, …, 25%, etc.) is the same for each SNR interval in each known class. After obtaining all of the thresholds for each SNR interval of all of the known classes, we could evaluate the recognition accuracy in the validation dataset through the corresponding thresholds controlled by the same ratio (like 5%). For example, if we set the ratio as 5%, that means the corresponding threshold will be related to the top 5% highest reconstruction losses in each SNR interval for each known class in the validation dataset. Based on the ratio, the recognition results will be revised (some samples will be labeled as unknown), leading to different mean recognition accuracies in the known class validation dataset. Therefore, by setting a suitable ratio, it could be ensured that the methods have 80%, 85%, 90%, and 95%, respectively, mean recognition accuracy (also the mean recognition accuracy for known classes) on the corresponding validation dataset. For example, when we set the ratios as 4% and 20%, the mean recognition accuracies on the validation dataset are 95% and 80%, respectively. Based on this threshold setting with 80%, 85%, 90%, and 95% mean recognition accuracy on the validation dataset, we test the methods with the corresponding testing dataset.
(3)
Perform (1) and (2) in Step 2 until all of the known classes are covered.
Step 3: Recognition of the testing sample.
(1)
Estimate the SNR for the testing samples and convert the testing samples to TFIs by using STFT. Now the testing samples contain the unknown class.
(2)
Use the well-trained recognition model in Step 1 to predict the class of the testing sample.
(3)
Based on the prediction result, use the corresponding encoder–decoder model to calculate the reconstruction loss of this testing sample. For example, if a testing sample is recognized as category { 2 } , then only the encoder–decoder model for category { 2 } will be used to calculate its reconstruction loss.
(4)
Based on the SNR and reconstruction loss of the testing sample and the threshold for reconstruction loss in each interval of the SNR for the used encoder–decoder model, the final recognition result for this sample will be obtained. Note that if the reconstruction loss is higher than the threshold, then the corresponding sample is seen as the unknown class. If it is not, the corresponding sample is seen as the known class, and its label is the predicted result from the original recognition model.
A flow chart of the proposed intra-pulse modulation OSR method is shown in Figure 1. Figure 2 shows a sketch map of the proposed method in actual deployment.

2.1. Data Preprocessing

In this paper, we chose to use short-time Fourier transformation (STFT) to convert the time domain signal into a time-frequency domain. STFT is based on Fourier transformation, which is faster than other time-frequency analysis tools. The mathematical expression of STFT [18,19] could be written as:
T F ( m , n ) = k = 0 N 1 s ( k ) h ( k m ) e j 2 π n N k
where T F is the result of the STFT. m and n are the time index and frequency index of T F , respectively. s ( k ) is the sampled sequence of the intra-pulse signal. h ( k ) is the window function, which is used to obtain the sub-signal of s ( k ) . In this paper, the length of the obtained sub-signal s ( k ) h ( k m ) is set to be 256, which is also the value of N . Hamming Window is chosen as the window function, with a length of 64 (a quarter of the length of the sub-signal). Compared with a traditional Fourier transformation, which provides only global frequency-domain features, the time shift and frequency shift of the window function provide STFT with localization characteristics. By using STFT, the radar emitter signals could be analyzed with time-frequency features.
As the result of STFT is a complex matrix, first we calculate the square of the modulus value of each element in T F , then we resize the resulting matrix in an image processing way, where the resized result is a 128 × 128 matrix. Finally, we normalize the value of each element in this matrix so that the value for each element of the resized matrix will range from 0 to 1. The process could be written as:
T F s q u a r e ( a , b ) = T F s q u a r e ( a , b ) × T F s q u a r e * ( a , b ) , a [ 1 , 2 , , m ] , b [ 1 , 2 , , 256 ]
imresize ( ) : T F s q u a r e m × 256 T F r e s i z e 128 × 128
T F I ( i , j ) = T F r e s i z e ( i , j ) min ( T F r e s i z e ) max ( T F r e s i z e ) min ( T F r e s i z e ) , i [ 1 , 2 , , 128 ] , j [ 1 , 2 , , 128 ]
where T F s q u a r e is the result calculated by squaring each element in T F and T F s q u a r e * ( a , b ) is the conjugation of T F s q u a r e ( a , b ) . The operation of imresize ( ) is based on nearest-neighbor interpolation. T F r e s i z e is the result of resizing T F s q u a r e as a 128 × 128 matrix. max ( ) and min ( ) are the functions used to obtain the max value and min value of the matrix, respectively. T F I is the final result used for training and testing the model.
The results of data preprocessing based on STFT are shown in Section 3. Through the data preprocessing based on STFT, the time-domain sampled radar emitter signals with different intra-pulse modulations can be converted into time-frequency images, which are used as the input for deep neural network models.

2.2. SNR Estimation

The signal-to-noise ratio (SNR) is important for intra-pulse modulation recognition. We assume that the noise is additional white Gaussian noise [20]. Traditionally, except for the intra-pulse part, the sampler also samples the blank part, which only includes the noise. Let the received sequence of the blank part be n o i s e b ( n ) . If the received sequence of pure intra-pulse signal is s ( n ) , then the actual received sequence of the intra-pulse signal polluted by noise could be written as:
x ( n ) = s ( n ) + n o i s e i n ( n )
where x ( n ) is the sampled intra-pulse sequence polluted by noise. n o i s e i n ( n ) refers to the noise sequence in the intra-pulse part. The power of x ( n ) could be calculated as follows:
P x = i = 0 N 1 ( x ( i ) ) 2 N
where P x is the power of x ( n ) and N is the length of the sequence in the intra-pulse part. As P x is the additional result of the power of s ( n ) and the power of n o i s e i n ( n ) , the following relationship can be written as:
P x = P s + P n o i s e i n
P s = i = 0 N 1 ( s ( i ) ) 2 N
P n o i s e i n = i = 0 N 1 ( n o i s e i n ( i ) ) 2 N
where P s and P n o i s e i n are the power of s ( n ) and n o i s e i n ( n ) , respectively.
Based on the definition of SNR, the actual value of SNR is calculated as:
S N R a c t u a l = 10 log 10 P s P n o i s e i n
As the power of noise is stable during a period of time, the power of n i n ( t ) is closely similar to that of n b ( t ) . Therefore, the estimated SNR could be written as:
S N R e s t i m a t e d = 10 log 10 P x P n o i s e b P n o i s e b
P n o i s e b = k = 0 M 1 ( n o i s e b ( k ) ) 2 M
where S N R c a l c u l a t e d is the estimated SNR of the signal. M is the length of the sequence in the blank part.

2.3. Original Recognition Model for CSR

In many studies, convolutional neural networks (CNNs) have been used for recognizing intra-pulse modulations. Unlike the recognition with ImageNet [21,22], which usually contains more than 1000 classes, there are fewer types of intra-pulse modulations. Therefore, we take a CNN model with a normal VGG-based structure [23] as the intra-pulse modulation recognition backbone. The structure of the original recognition model used in this paper is shown in Figure 3.
The input shape of the original recognition model is 128 × 128 . There are five convolutional blocks and each block contains two convolutional layers, a max-pooling layer where the stride and pooling size are 2, and a batch-normalization layer [24]. The kernel size in all convolutional blocks is 3 and the channels for convolutional block 1 to convolutional block 5 are 16, 32, 64, 128 and 256, respectively. The fully connection layer contains 256 nodes and the number of nodes of the output layer is the same as the number of known classes. Except for the output layer, which uses “softmax,” the activation function for the rest is “ReLU” [25]. At the stage of training this recognition model, the cross-entropy function is selected as the loss function.

2.4. Encoder–Decoder Model

In anomaly detection tasks, the auto-encoder models have higher reconstruction loss for the anomaly samples than the normal samples. And inspired by SegNet [26], we chose to use an encoder–decoder-based structure to compress and decompress the TFIs and revise the recognition result predicted by the original recognition model. The structure of the encoder–decoder model used in this paper is shown in Figure 4.
The input shape of the encoder–decoder model is 128 × 128 . The encoder–decoder model mainly contains two parts.
The first part is used to compress the TFIs. There are six compressed convolutional blocks in the compress-encoder part. The compressed convolutional block is the same as the convolutional block in the original recognition model in Section 3.3. The channels for compressed convolutional block 1 to compressed convolutional block 6 are 16, 32, 64, 128, 128, and 128, respectively.
The second part of the encoder–decoder model is the decompress-encoder part, which is used for up-sampling the features from the compress-encoder part and reconstructing the TFIs. There are six decompressed convolutional blocks in the decompress-encoder part. Each block includes an up-sampling layer where the rate is 2, two convolutional layers, and a batch-normalization layer. The kernel size in all decompress convolutional blocks is 3. And the channels for decompressed convolutional block 1 to decompressed convolutional block 6 are 128, 128, 128, 64, 32, and 16, respectively. The activation function used in this part is “ReLU”.
The output layer includes a convolutional layer with “Sigmoid” as the activation function, where the channel and kernel size are 1. Therefore, the output shape of the encoder–decoder model is same as the input shape. During the training session, MSE was used as the loss function.

3. Dataset and Experimental Settings

This section provides a simulation dataset of eleven different intra-pulse modulations of radar emitter signals. Then, a description of the experiments with different levels of openness is provided in detail. In addition, the two most representative methods based on OpenMax and the maximum logit score (MLS) were used as the baseline methods for intra-pulse modulation of OSR.
All programs related to the experiments were conducted with a computer with Intel 10,900 K, 128 GB RAM, and RTX 4090 GPU hardware capabilities; “MATLAB 2021a” software; and the “Keras” and “Python” programming languages.

3.1. Intra-Pulse Modulation Radar Emitter Signal Datasets

In the actual environment, the carrier frequency of a radar emitter signal could range from 300 MHz to 300 GHz. The receiver, where its local oscillators will down mix the frequency, could output the signal sequences with a lower frequency. These low-frequency signal sequences are finally aimed for analyzing and preprocessing later. Although the pulse width could range from several microseconds to a hundred microseconds, a short pulse width is more common. Therefore, in this paper, we focus on a certain range of pulse width, from 5 μs to 20 μs. Besides, assuming the bandwidth for the receiver is 100 MHz, we set a 200 MHz sampling frequency with single-channel sampling.
As a result, in this paper, we simulated eleven types of intra-pulse modulation signals, which include single-carrier frequency (SCF) signals, linear frequency modulation (LFM) signals, sinusoidal frequency modulation (SFM) signals, quadratic frequency modulation (EQFM) signals, dual linear frequency modulation (DLFM) signals, multiple linear frequency modulation (MLFM) signals, binary frequency shift keying (BFSK) signals, quadrature frequency shift keying (QFSK) signals, binary phase shift keying (BPSK) signals, Frank phase-coded (Frank) signals, and composite modulation with LFM and BPSK (LFM-BPSK) signals. The parameters of the eleven types of intra-pulse modulations are shown in Table 1.
The SNR for each type of intra-pulse modulation ranges from −10.5 dB to 10.5 dB. That is, there are 21 intervals of the SNR split by 1 dB. The mid value of each corresponding interval is [ 10   dB , 9   dB , , 9   dB , 10   dB ] . For example, if the SNR = 2 dB, this means that the actual SNR in this interval belongs to [ 1.5   dB , 2.5   dB ) .
In each interval of the SNR, the number of samples for each type of intra-pulse modulation is 1500. Where 600 samples are training samples, 400 samples are validation samples and 500 samples are testing samples. Figure 5 shows the heat maps of the TFIs of eleven types of intra-pulse modulation signals when the SNR is 10 dB.

3.2. Description of the Experiments with Different Openness

In this paper, we conducted several experiments with different levels of openness to show the performance of the proposed OSR method for intra-pulse modulation. Specifically, we take five main conditions into consideration. The details of these five conditions are shown as follows:
Condition 1: Ten out of eleven types of intra-pulse modulation are known classes, and the other type is an unknown class. Therefore, there are 11 different situations in Condition 1.
Condition 2: Nine out of eleven types of intra-pulse modulation are known classes, and the other two types are unknown classes. Therefore, there are 55 different situations in Condition 2.
Condition 3: Eight out of eleven types of intra-pulse modulation are known classes, and the other three types are unknown classes. Therefore, there are 165 different situations in Condition 3.
Condition 4: Seven out of eleven types of intra-pulse modulation are known classes, and the other four types are unknown classes. Therefore, there are 330 different situations in Condition 4.
Condition 5: Six out of eleven types of intra-pulse modulation are known classes, and the other five types are unknown classes. Therefore, there are 462 different situations in Condition 5.

3.3. Baseline Methods

OpenMax has been widely used for OSR tasks. It combines deep learning with extreme theory to estimate the probability of unknown classes. In [17], the authors used STFT and OpenMax to recognize different jamming signals. Although the DNN model and the source signal are different, the OpenMax-based recognition method in [17] is still instructive. Therefore, we chose to reuse its OpenMax method accompanied by the recognition model in Section 2.3 as the baseline method. In addition, the maximum logit score (MLS) method was proven competitive with or to outperform other OSR methods in [27]. Therefore, in this paper, it was also chosen as a baseline method.

4. Experiments

4.1. Training Details and Computational Cost

The original recognition model for intra-pulse modulation CSR was trained with 20 epochs and the batch size was 64. The weights for the testing sections were chosen to have the highest overall recognition accuracy on the corresponding validation dataset.
For the encoder–decoder model, the training epoch was 300 and the batch size was 40. The optimization algorithm was all set as adaptive moment estimation (ADAM) [28] with a 0.001 learning rate. The weights, which have the lowest overall reconstruction loss on the according validation dataset, were chosen as in the testing section.
The average data preprocessing time per sample based on the method in Section 2.1 and the dataset in Section 3.1 was around 14.4 ms. The floating point operations (FLOPs), parameters, and time usage for the training and testing sessions based on the CSR model and encoder–decoder model are shown in Table 2 and Table 3, respectively.

4.2. Experimental Results of the Proposed Method and Baseline Methods

Our proposed method and the baseline method were trained based on the dataset in Section 3.1 with different conditions as in Section 3.2. First of all, we provided well-trained the average recognition accuracy of the original recognition model on the corresponding known testing dataset in a closed set situation. The results are shown in Table 4, which indicates that the original recognition model has good performance on intra-pulse modulation CSR.
Next, we conducted the experiments for the intra-pulse modulation OSR. In real actual radar systems, the average recognition accuracy on the known classes should be kept higher than or at least close to a certain value. Based on this reason, the average accuracy threshold on known classes was set to be 80%, 85%, 90%, and 95%, respectively.
Therefore, for each situation of the different conditions in Section 3.2, we set different thresholds for the reconstruction loss for our method and the score thresholds for the OpenMax-and MLS-based methods, so that the average accuracy of these methods on the corresponding validation dataset could adapt to the requirements (equal to or close to the average accuracy thresholds as much as possible).
The setting of the threshold for the OpenMax-based method and MLS-based method was based on the revised softmax score and the maximum logit score, respectively. Both methods follow the principle that if the score of the given testing sample is lower than the threshold, then the sample will be seen as unknown.
The setting of the threshold for our method is based on the principle in Section 2, where we control the ratio to ensure that the mean recognition accuracy in corresponding validation datasets is close to 80%, 85%, 90%, and 95%, respectively. Figure 6 shows the thresholds of reconstruction loss in the corresponding validation dataset with different SNRs based on these 11 encoder–decoder models when selecting 85%, 90%, and 95% samples as the known, respectively.
Then, we evaluated the above models by testing the dataset. Note that the threshold settings are all based on the corresponding validation dataset.
Figure 6 shows that a higher ratio for selecting samples as the known leads to a higher threshold for reconstruction loss. In addition, it can be seen that when the value of the SNR rises, the reconstruction loss of the corresponding validation samples based on the encoder–decoder model generally decreases.
Table 5, Table 6, Table 7 and Table 8 give the testing results of the average recognition accuracy of the methods for known classes and unknown classes, where the average accuracy threshold for the validation dataset is 80%, 85%, 90%, and 95%, respectively. Note that if the recognition accuracy for one of situations in one of conditions cannot reach the average accuracy threshold (like 95%), for OpenMax, the revised softmax-based threshold will be 0 and for our method, the threshold of the SNR will be positive infinity.
As the results show, both of the methods have good performance in recognizing the known intra-pulse modulations. But there is a significant difference for these three methods in the average recognition accuracies in the unknown class. When the average accuracy threshold is 95%, the average recognition accuracy in the unknown class of the OpenMax-based method is at most 0.28, while our proposed OSR method and the MLS-based method perform better, where the average recognition accuracy in the unknown class is over 0.50. As the value of the average accuracy threshold decreases, the average recognition accuracy in the unknown class for the three methods increases. Besides, although the MLS-based method performs better in Condition 1 to Condition 3, when the openness rises, our method is better at recognizing the unknown samples.
The average recognition accuracy with five conditions and different SNRs is shown in Figure 7, Figure 8, Figure 9, Figure 10 and Figure 11.
From the figures, it can be seen that our OSR method and the MLS-based method are more stable than the OpenMax-based method. In contrast, the known classes accuracy of the OpenMax-based method first rises with the increment of SNR, then it drops a lot when the SNR keeps increasing. Although the accuracy for known classes based on the OpenMax-based method is higher when SNR is higher, in the situation of a lower SNR, the OpenMax-based method performs extremely poorly compared with our OSR method and the MLS-based method. In addition, the trend of the accuracy in the unknown class based on the OpenMax-based method is decreasing as the SNR increases, while the trend based on our OSR method and the MLS-based method is generally increasing. In addition, although the MLS-based method is better at lower SNR, when the SNR goes up, the average recognition accuracy in the unknown class is less competitive than our method, where the average recognition accuracy based our method in the unknown class is equal or more than 80% at 5 dB.
As a result, based on extensive experimental results with five different openness conditions, it is indicated that our proposed method, which better meets the actual requirement where the recognition of known classes is taken as the first consideration for actual radar systems, is effective for recognizing intra-pulse modulations of radar emitter signals in an open set. Figure 12, Figure 13, Figure 14, Figure 15 and Figure 16 provide some confusion matrices based on our method for one of the situations in the testing dataset when the SNR is 0 dB, 5 dB, and 10 dB in Condition 1 to Condition 5, respectively.

5. Discussion

We further analyzed the influence of the structure of the reconstruction model on recognition accuracy. The ablation experiments are based on the same dataset and conditions as in Section 3.1 and Section 3.2, and the training setting is the same as in Section 4.1. In addition, we provide some scenarios where our method could potentially be applied.

5.1. Ablation Study on the Structure of the Reconstruction Model on Recognition Accuracy

In this section, the original recognition models are same to those in Section 4.2. We changed the number of compressed convolutional blocks and decompressed convolutional blocks. Specifically, three ablation models are included. These three models are based the same encoder–decoder structure in Section 2.4. The differences are the block numbers. The details are as follows:
Ablation Model 1 (AM1): 4th to 6th compressed convolutional blocks and decompressed convolutional blocks are deleted.
Ablation Model 2 (AM2): 5th to 6th compressed convolutional blocks and decompressed convolutional blocks are deleted.
Ablation Model 3 (AM3): 6th compressed convolutional block and decompressed convolutional block are deleted.
In addition, the parameters of the reserved convolutional blocks are the same as the blocks in Section 2.4.
Except for these three models, we also applied the widely used reconstruction models named U-Net [29] and UNet3+ [30] as two extra ablation models.
Based on these reconstruction models, we conducted the intra-pulse modulation OSR experiments. The average recognition results are shown in Table 9, Table 10, Table 11 and Table 12. In addition, we summarized the average reconstruction loss on the corresponding validation dataset in Table 13.
From Table 9, Table 10, Table 11, Table 12 and Table 13, it could be found that the deeper structure of the proposed encoder–decoder model provides better performance in recognizing the unknown intra-pulse modulation. In addition, the experimental results of AM1 to AM3 and ours show that the structure of the encoder–decoder will obtain more reconstruction loss when the depth increases. In addition, the average reconstruction loss of the U-Net model and UNet3+ model is quite low, compared among the Ablation Model 1 to Ablation Model 3 and the model in Section 2.4, which means these two models reconstruct the TFIs in high quality.
However, although these two models with complex structures have lower reconstruction loss in the validation dataset, they do not provide good performance in recognizing the unknown intra-pulse modulation. The reason for this is that the merging connection from the encoder part to the decoder part keeps many shallower features, where the decoder part receives many features without deep compression. If the model receives an unknown sample, due to the merging connection, the output will still have low reconstruction loss.
The encoder–decoder structure in Section 2.4, which does not have an extra connection branch, extremely compresses the feature used for the decoder part. Therefore, the output reconstructed TFIs have higher reconstruction loss. If this model receives an unknown sample, then a very slight change from the output of its encoder part will cause a higher difference in the output result, reflected in the higher reconstruction loss compared with the unknown samples.
In terms of the intra-pulse modulation OSR, our proposed structure of an encoder–decoder model without an extra merging connection is more suitable and effective for recognizing the unknown class.

5.2. Application Scenarios

DNN-based models have been widely used for radar emitter signal intra-pulse modulation recognition. In real applications, except for the known classes, radar systems are often facing unknown intra-pulse modulations. It is essential for the systems not only to classify the known classes in the library correctly, but also to identify the unknown samples. Currently, there are few pieces of literature focusing on OSR in radar emitter signal intra-pulse modulation. Through the result of Table 5, Table 6, Table 7, Table 8, Table 9, Table 10, Table 11 and Table 12 in Section 4.2 and Section 5.1, we could reach a conclusion that a deeper encoder–decoder structure without an extra merging connection could increase the performance in recognizing unknown intra-pulse modulation.
At the same time, our method could be applied to other radar signal recognition tasks. For example, in radar jamming recognition tasks [17,31], the library cannot include all of the jamming types from the target jamming emitter. Therefore, it is important to accurately classify the known jamming types and recognize the unknown jamming. Besides, automatic target recognition based on synthetic aperture radar images [32] and high-resolution range profile recognition [33] is also facing the problem that unknown targets often emerge and cannot be neglected. Correctly recognizing the unknown targets will provide better intelligence for the systems. Although the source data for different radar signal recognition tasks are different, the framework of our method, like threshold setting and an encoder–decoder model for revising the recognition result, still could provide guidance for designing corresponding recognition methods.

6. Conclusions

In this paper, we focused on an open set situation instead of a closed set situation and proposed a method for recognizing radar emitter signal intra-pulse modulations. Through the extensive experiments, including five different openness conditions, we found that our proposed method, which combines data preprocessing based on STFT, SNR estimation, a recognition model for CSR, and an encoder–decoder model, is effective in recognizing an unknown intra-pulse modulation with good performance in classifying known intra-pulse modulations. Additionally, according to an ablation study of the effect of the structure of the reconstruction model on recognition accuracy, we could conclude that a deeper encoder–decoder structure without extra merging connections leads to better performance in recognizing unknown samples. Our proposed method in this paper could provide guidance for radar emitter signal intra-pulse modulation OSR tasks.
As the current method only recognizes unknown samples from different types as one unknown class, in future work, we will focus on the task of classifying these unknown samples into specific classes.

Author Contributions

Conceptualization, S.Y., P.L. and B.W.; methodology, S.Y.; software, S.Y.; validation, S.Y.; formal analysis, S.Y.; investigation, S.Y.; resources, S.Y.; data curation, S.Y.; writing—original draft preparation, S.Y.; writing—review and editing, S.Y.; visualization, S.Y.; supervision, P.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

Data are contained within the article.

Acknowledgments

The authors would like to thank the reviewers for their valuable comments, which improved the quality of paper.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Zhao, G. Principle of Radar Countermeasure, 2nd ed.; Xidian University Press: Xi’an, China, 2012. [Google Scholar]
  2. Richards, M.A. Fundamentals of Radar Signal Processing, 2nd ed.; McGraw-Hill Education: New York, NY, USA, 2005. [Google Scholar]
  3. Barton, D.K. Radar System Analysis and Modeling; Artech: London, UK, 2004. [Google Scholar]
  4. Wiley, R.G.; Ebrary, I. ELINT: The Interception and Analysis of Radar Signals; Artech: London, UK, 2006. [Google Scholar]
  5. Qu, Z.; Mao, X.; Deng, Z. Radar Signal Intra-Pulse Modulation Recognition Based on Convolutional Neural Network. IEEE Access 2018, 6, 43874–43884. [Google Scholar] [CrossRef]
  6. Yu, Z.; Tang, J.; Wang, Z. GCPS: A CNN Performance Evaluation Criterion for Radar Signal Intrapulse Modulation Recognition. IEEE Commun. Lett. 2021, 25, 2290–2294. [Google Scholar] [CrossRef]
  7. Yuan, S.; Wu, B.; Li, P. Intra-Pulse Modulation Classification of Radar Emitter Signals Based on a 1-D Selective Kernel Convolutional Neural Network. Remote Sens. 2021, 13, 2799. [Google Scholar] [CrossRef]
  8. Wu, B.; Yuan, S.; Li, P.; Jing, Z.; Huang, S.; Zhao, Y. Radar Emitter Signal Recognition Based on One-Dimensional Convolutional Neural Network with Attention Mechanism. Sensors 2020, 20, 6350. [Google Scholar] [CrossRef]
  9. Si, W.; Wan, C.; Deng, Z. Intra-Pulse Modulation Recognition of Dual-Component Radar Signals Based on Deep Convolutional Neural Network. IEEE Commun. Lett. 2021, 25, 3305–3309. [Google Scholar] [CrossRef]
  10. Tan, M.; Le, Q. Efficientnet: Rethinking model scaling for convolutional neural networks. In Proceedings of the International Conference on Machine Learning, Long Beach, CA, USA, 9–15 June 2019; pp. 6105–6114. [Google Scholar]
  11. Wan, C.; Si, W.; Deng, Z. Research on modulation recognition method of multi-component radar signals based on deep convolution neural network. IET Radar Sonar Navig. 2023, 17, 1313–1326. [Google Scholar] [CrossRef]
  12. Cai, J.; He, M.; Cao, X.; Gan, F. Semi-Supervised Radar Intra-Pulse Signal Modulation Classification With Virtual Adversarial Training. IEEE Internet Things J. 2023. [Google Scholar] [CrossRef]
  13. Yuan, S.; Li, P.; Wu, B.; Li, X.; Wang, J. Semi-Supervised Classification for Intra-Pulse Modulation of Radar Emitter Signals Using Convolutional Neural Network. Remote Sens. 2022, 14, 2059. [Google Scholar] [CrossRef]
  14. Hendrycks, D.; Gimpel, K. A baseline for detecting misclassified and out-of-distribution examples in neural networks. arXiv 2016, arXiv:1610.02136. [Google Scholar]
  15. Bendale, A.; Boult, T.E. Towards Open Set Deep Networks. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 1563–1572. [Google Scholar] [CrossRef]
  16. Sun, C.; Du, Y.; Qiao, X.; Wu, H.; Zhang, T. Research on the Enhancement Method of Specific Emitter Open Set Recognition. Electronics 2023, 12, 4399. [Google Scholar] [CrossRef]
  17. Zhou, Y.; Shang, S.; Song, X.; Zhang, S.; You, T.; Zhang, L. Intelligent Radar Jamming Recognition in Open Set Environment Based on Deep Learning Networks. Remote Sens. 2022, 14, 6220. [Google Scholar] [CrossRef]
  18. Auger, F.; Flandrin, P. Improving the readability of time-frequency and time-scale representations by the reassignment method. IEEE Trans. Signal Process. 1995, 43, 1068–1089. [Google Scholar] [CrossRef]
  19. Wang, X.; Huang, G.; Zhou, Z.; Tian, W.; Yao, J.; Gao, J. Radar Emitter Recognition Based on the Energy Cumulant of Short Time Fourier Transform and Reinforced Deep Belief Network. Sensors 2018, 18, 3103. [Google Scholar] [CrossRef] [PubMed]
  20. Hoang, L.M.; Kim, M.; Kong, S.-H. Automatic Recognition of General LPI Radar Waveform Using SSD and Supplementary Classifier. IEEE Trans. Signal Process. 2019, 67, 3516–3530. [Google Scholar] [CrossRef]
  21. Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; Rabinovich, A. Going deeper with convolutions. In Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015; pp. 1–9. [Google Scholar] [CrossRef]
  22. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar] [CrossRef]
  23. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  24. Ioffe, S.; Szegedy, C. Batch normalization: Accelerating deep network training by reducing internal covariate shift. arXiv 2015, arXiv:1502.03167. [Google Scholar]
  25. Nair, V.; Hinton, G.E. Rectified linear units improve restricted boltzmann machines. In Proceedings of the 27th International Conference on Machine Learning (ICML-10), Haifa, Israel, 21–24 June 2010. [Google Scholar]
  26. Badrinarayanan, V.; Kendall, A.; Cipolla, R. SegNet: A Deep Convolutional Encoder-Decoder Architecture for Image Segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 2481–2495. [Google Scholar] [CrossRef] [PubMed]
  27. Vaze, S.; Han, K.; Vedaldi, A.; Zisserman, A. Open-set recognition: A good closed-set classifier is all you need. arXiv 2021, arXiv:2110.06207. [Google Scholar]
  28. Kingma, D.P.; Ba, J. Adam: A Method for Stochastic Optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
  29. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Munich, Germany, 5–9 October 2015; Springer: Cham, Switzerland, 2015; pp. 234–241. [Google Scholar]
  30. Huang, H.; Lin, L.; Tong, R.; Hu, H.; Zhang, Q.; Iwamoto, Y.; Han, X.; Chen, Y.-W.; Wu, J. UNet 3+: A Full-Scale Connected UNet for Medical Image Segmentation. In Proceedings of the ICASSP 2020—2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Barcelona, Spain, 4–8 May 2020; pp. 1055–1059. [Google Scholar] [CrossRef]
  31. Lv, Q.; Quan, Y.; Feng, W.; Sha, M.; Dong, S.; Xing, M. Radar Deception Jamming Recognition Based on Weighted Ensemble CNN With Transfer Learning. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5107511. [Google Scholar] [CrossRef]
  32. Zhou, X.; Bai, X.; Xue, R. ISAR Target Recognition Based on Capsule Net. In Proceedings of the 2021 CIE International Conference on Radar (Radar), Haikou, China, 15–19 December 2021; pp. 39–42. [Google Scholar] [CrossRef]
  33. Guo, D.; Chen, B.; Chen, W.; Wang, C.; Liu, H.; Zhou, M. Variational Temporal Deep Generative Model for Radar HRRP Target Recognition. IEEE Trans. Signal Process. 2020, 68, 5795–5809. [Google Scholar] [CrossRef]
Figure 1. A flow chart of the proposed intra-pulse modulation OSR method.
Figure 1. A flow chart of the proposed intra-pulse modulation OSR method.
Remotesensing 16 00108 g001
Figure 2. A sketch map of the proposed method in actual deployment. For a given sample, the well-trained original recognition model for CSR will label it a certain known class temporarily. Through estimating the SNR of the sample and calculating the reconstruction loss by inputting the sample to the corresponding encoder–decoder model related to the temporary predicted known class, whether the sample belongs to the predicted temporary known class or the unknown class will be confirmed.
Figure 2. A sketch map of the proposed method in actual deployment. For a given sample, the well-trained original recognition model for CSR will label it a certain known class temporarily. Through estimating the SNR of the sample and calculating the reconstruction loss by inputting the sample to the corresponding encoder–decoder model related to the temporary predicted known class, whether the sample belongs to the predicted temporary known class or the unknown class will be confirmed.
Remotesensing 16 00108 g002
Figure 3. The structure of the original recognition model used in this paper.
Figure 3. The structure of the original recognition model used in this paper.
Remotesensing 16 00108 g003
Figure 4. The structure of the encoder–decoder model used in this paper.
Figure 4. The structure of the encoder–decoder model used in this paper.
Remotesensing 16 00108 g004
Figure 5. The heat maps of the TFIs of eleven types of intra-pulse modulation signals when the SNR is 10 dB.
Figure 5. The heat maps of the TFIs of eleven types of intra-pulse modulation signals when the SNR is 10 dB.
Remotesensing 16 00108 g005
Figure 6. The thresholds of reconstruction loss in the corresponding validation datasets with different SNRs based on these 11 encoder–decoder models when selecting 85%, 90%, and 95% samples as the known, respectively, which equals to setting the top highest 15%, 10%, and 5% reconstruction losses from the corresponding validation samples as the thresholds in each interval of the SNR in one known class. The marker point refers to the accuracy for the 1 dB interval of the SNR.
Figure 6. The thresholds of reconstruction loss in the corresponding validation datasets with different SNRs based on these 11 encoder–decoder models when selecting 85%, 90%, and 95% samples as the known, respectively, which equals to setting the top highest 15%, 10%, and 5% reconstruction losses from the corresponding validation samples as the thresholds in each interval of the SNR in one known class. The marker point refers to the accuracy for the 1 dB interval of the SNR.
Remotesensing 16 00108 g006
Figure 7. The average recognition accuracy of our proposed method, the OpenMax-based method, and the MLS-based method on the according test dataset for different SNRs in Condition 1. The marker point refers to the accuracy for the 1 dB interval of the SNR. The results are based on the average results of 11 situations in Condition 1. “Threshold of 80%, 85%, 90%, and 95%” means the average accuracy in the corresponding validation dataset with known classes is 80%, 85%, 90%, and 95%, respectively.
Figure 7. The average recognition accuracy of our proposed method, the OpenMax-based method, and the MLS-based method on the according test dataset for different SNRs in Condition 1. The marker point refers to the accuracy for the 1 dB interval of the SNR. The results are based on the average results of 11 situations in Condition 1. “Threshold of 80%, 85%, 90%, and 95%” means the average accuracy in the corresponding validation dataset with known classes is 80%, 85%, 90%, and 95%, respectively.
Remotesensing 16 00108 g007
Figure 8. The average recognition accuracy of our proposed method, the OpenMax-based method, and the MLS-based method in the corresponding test dataset for different SNRs in Condition 2. The marker point refers to the accuracy for the 1 dB interval of the SNR. The results are based on the average results of 55 situations in Condition 2. “Threshold of 80%, 85%, 90%, and 95%” means the average accuracy in the corresponding validation dataset with known classes is 80%, 85%, 90%, and 95%, respectively.
Figure 8. The average recognition accuracy of our proposed method, the OpenMax-based method, and the MLS-based method in the corresponding test dataset for different SNRs in Condition 2. The marker point refers to the accuracy for the 1 dB interval of the SNR. The results are based on the average results of 55 situations in Condition 2. “Threshold of 80%, 85%, 90%, and 95%” means the average accuracy in the corresponding validation dataset with known classes is 80%, 85%, 90%, and 95%, respectively.
Remotesensing 16 00108 g008
Figure 9. The average recognition accuracy of our proposed method, the OpenMax-based method, and the MLS-based method in the corresponding test dataset for different SNRs in Condition 3. The marker point refers to the accuracy for the 1 dB interval of the SNR. The results are based on the average results of 165 situations in Condition 3. “Threshold of 80%, 85%, 90%, and 95%” means the average accuracy in the corresponding validation dataset with known classes is 80%, 85%, 90%, and 95%, respectively.
Figure 9. The average recognition accuracy of our proposed method, the OpenMax-based method, and the MLS-based method in the corresponding test dataset for different SNRs in Condition 3. The marker point refers to the accuracy for the 1 dB interval of the SNR. The results are based on the average results of 165 situations in Condition 3. “Threshold of 80%, 85%, 90%, and 95%” means the average accuracy in the corresponding validation dataset with known classes is 80%, 85%, 90%, and 95%, respectively.
Remotesensing 16 00108 g009
Figure 10. The average recognition accuracy of our proposed method, the OpenMax-based method, and the MLS-based method in the corresponding test dataset for different SNRs in Condition 4. The marker point refers to the accuracy for the 1 dB interval of the SNR. The results are based on the average results of 330 situations in Condition 4. “Threshold of 80%, 85%, 90%, and 95%” means the average accuracy in the corresponding validation dataset with known classes is 80%, 85%, 90%, and 95%, respectively.
Figure 10. The average recognition accuracy of our proposed method, the OpenMax-based method, and the MLS-based method in the corresponding test dataset for different SNRs in Condition 4. The marker point refers to the accuracy for the 1 dB interval of the SNR. The results are based on the average results of 330 situations in Condition 4. “Threshold of 80%, 85%, 90%, and 95%” means the average accuracy in the corresponding validation dataset with known classes is 80%, 85%, 90%, and 95%, respectively.
Remotesensing 16 00108 g010
Figure 11. The average recognition accuracy of our proposed method, the OpenMax-based method, and the MLS-based method in the corresponding test dataset for different SNRs in Condition 5. The marker point refers to the accuracy for the 1 dB interval of the SNR. The results are based on the average results of 462 situations in Condition 5. “Threshold of 80%, 85%, 90%, and 95%” means the average accuracy in the corresponding validation dataset with known classes is 80%, 85%, 90%, and 95%, respectively.
Figure 11. The average recognition accuracy of our proposed method, the OpenMax-based method, and the MLS-based method in the corresponding test dataset for different SNRs in Condition 5. The marker point refers to the accuracy for the 1 dB interval of the SNR. The results are based on the average results of 462 situations in Condition 5. “Threshold of 80%, 85%, 90%, and 95%” means the average accuracy in the corresponding validation dataset with known classes is 80%, 85%, 90%, and 95%, respectively.
Remotesensing 16 00108 g011
Figure 12. The confusion matrix based on our method for one of the situations in the testing dataset when the SNR is 0 dB, 5 dB, and 10 dB (from left to right) in Condition 1. In this situation, LFM-BPSK is selected as the unknown class. Darker background colors represent more quantities.
Figure 12. The confusion matrix based on our method for one of the situations in the testing dataset when the SNR is 0 dB, 5 dB, and 10 dB (from left to right) in Condition 1. In this situation, LFM-BPSK is selected as the unknown class. Darker background colors represent more quantities.
Remotesensing 16 00108 g012
Figure 13. The confusion matrix based on our method for one of the situations in the testing dataset when the SNR is 0 dB, 5 dB, and 10 dB (from left to right) in Condition 2. In this situation, SCF and LFM-BPSK are selected as the unknown classes. Darker background colors represent more quantities.
Figure 13. The confusion matrix based on our method for one of the situations in the testing dataset when the SNR is 0 dB, 5 dB, and 10 dB (from left to right) in Condition 2. In this situation, SCF and LFM-BPSK are selected as the unknown classes. Darker background colors represent more quantities.
Remotesensing 16 00108 g013
Figure 14. The confusion matrix based on our method for one of the situations in the testing dataset when the SNR is 0 dB, 5 dB, and 10 dB (from left to right) in Condition 3. In this situation, SCF, LFM and SFM are selected as the unknown classes. Darker background colors represent more quantities.
Figure 14. The confusion matrix based on our method for one of the situations in the testing dataset when the SNR is 0 dB, 5 dB, and 10 dB (from left to right) in Condition 3. In this situation, SCF, LFM and SFM are selected as the unknown classes. Darker background colors represent more quantities.
Remotesensing 16 00108 g014
Figure 15. The confusion matrix based on our method for one of the situations in the testing dataset when the SNR is 0 dB, 5 dB, and 10 dB (from left to right) in Condition 4. In this situation, LFM, SFM, EQFM, and DLFM are selected as the unknown classes. Darker background colors represent more quantities.
Figure 15. The confusion matrix based on our method for one of the situations in the testing dataset when the SNR is 0 dB, 5 dB, and 10 dB (from left to right) in Condition 4. In this situation, LFM, SFM, EQFM, and DLFM are selected as the unknown classes. Darker background colors represent more quantities.
Remotesensing 16 00108 g015
Figure 16. The confusion matrix based on our method for one of the situations in the testing dataset when the SNR is 0 dB, 5 dB, and 10 dB (from left to right) in Condition 5. In this situation, BFSK, QFSK, BPSK, FRANK and LFM-BPSK are selected as the unknown classes. Darker background colors represent more quantities.
Figure 16. The confusion matrix based on our method for one of the situations in the testing dataset when the SNR is 0 dB, 5 dB, and 10 dB (from left to right) in Condition 5. In this situation, BFSK, QFSK, BPSK, FRANK and LFM-BPSK are selected as the unknown classes. Darker background colors represent more quantities.
Remotesensing 16 00108 g016
Table 1. The parameters of eleven types of intra-pulse modulations.
Table 1. The parameters of eleven types of intra-pulse modulations.
TypeFrequencyParametersDetails
SCF10 MHz–90 MHzNoneNone
LFM10 MHz–90 MHzBandwidth: 10 MHz–70 MHz1. Both up LFM and down LFM are included
2. Both max value and min value of the instantaneous frequency for LFM range from 10 MHz to 70 MHz
SFM10 MHz–90MHzBandwidth: 10 MHz–70 MHzBoth max value and min value of the instantaneous frequency for SFM range from 10 MHz to 70 MHz
EQFM10 MHz–90 MHzBandwidth: 10 MHz–70 MHz1. The instantaneous frequency increases first and then decreases, or decreases first and then increases
2. Both max value and min value of the instantaneous frequency for EQFM range from 10 MHz to 70 MHz
DLFM10 MHz–90 MHzBandwidth: 10 MHz–70 MHz1. The instantaneous frequency increases first and then decreases, or decreases first and then increases
2. Both max value and min value of the instantaneous frequency for DLFM range from 10 MHz to 70 MHz
MLFM10 MHz–90 MHz
10 MHz–90 MHz
Bandwidth: 10 MHz–70 MHz
Bandwidth: 10 MHz–70 MHz
Segment ratio: 20–80%
1. Up LFM and down LFM are included in each of the two parts
2. Both max value and min value of the instantaneous frequency for each part of the MLFM range from 10 MHz to 70 MHz
3. The distance of the instantaneous frequency in the end of the first part and the instantaneous frequency in the start of the last part is more than 10 MHz
BFSK10 MHz–90 MHz
10 MHz–90 MHz
5, 7, 11, 13-bit Barker codeThe distance of two sub-carrier frequency is more than 10 MHz
QFSK10 MHz–90 MHz
10 MHz–90 MHz
10 MHz–90 MHz
10 MHz–90 MHz
16-bit Frank codeThe distance of each two sub-carrier frequency is more than 10 MHz
BPSK10 MHz–90 MHz5, 7, 11, 13-bit Barker codeNone
FRANK10 MHz–90 MHzPhase number: 6, 7, 8None
LFM-BPSK10 MHz–90 MHzBandwidth: 10 MHz–70 MHz
5, 7, 11, 13-bit Barker code
1. Both up LFM and down LFM are included
2. Both max value and min value of the instantaneous frequency for LFM-BPSK range from 10 MHz to 70 MHz
Table 2. The FLOPs, parameters, and time usage for the training and testing sessions based on the CSR model.
Table 2. The FLOPs, parameters, and time usage for the training and testing sessions based on the CSR model.
FLOPs4.5 M
Parameters2.2 M
Training Time/Iteration
(Batch Size: 64)
13.7 ms
Forward Time
(Single Sample)
2.3 ms
Table 3. The FLOPs, parameters, and time usage for training and testing sessions based on the encoder–decoder model.
Table 3. The FLOPs, parameters, and time usage for training and testing sessions based on the encoder–decoder model.
FLOPs3.8 M
Parameters1.9 M
Training Time/Iteration
(Batch Size: 40)
23 ms
Forward Time
(Single Sample)
4.9 ms
Table 4. The average recognition accuracy on the corresponding known testing dataset with different SNRs and conditions based on the original recognition models in the situation of a closed set. The number of known classes is 10, 9, 8, 7, and 6 for Condition 1 to Condition 5, respectively. The point of the SNR refers to the mean accuracy for the 1 dB interval of the SNR. The results for each condition are based on the average results of all situations in the corresponding condition. For Condition 1 to Condition 5, the number of situations is 11, 55, 165, 330, and 462, respectively.
Table 4. The average recognition accuracy on the corresponding known testing dataset with different SNRs and conditions based on the original recognition models in the situation of a closed set. The number of known classes is 10, 9, 8, 7, and 6 for Condition 1 to Condition 5, respectively. The point of the SNR refers to the mean accuracy for the 1 dB interval of the SNR. The results for each condition are based on the average results of all situations in the corresponding condition. For Condition 1 to Condition 5, the number of situations is 11, 55, 165, 330, and 462, respectively.
SNR/dBCondition 1Condition 2Condition 3Condition 4Condition 5
−100.87450.88530.89710.90810.9196
−90.93560.94090.94690.95290.9594
−80.96720.96980.97300.97610.9794
−70.98810.98890.98990.99110.9925
−60.99560.99550.99590.99640.9969
−50.99800.99810.99840.99840.9988
−40.99930.99950.99960.99950.9997
−30.99990.99980.99980.99990.9999
−20.99990.99990.99991.00001.0000
−11.00000.99990.99991.00001.0000
01.00001.00001.00001.00001.0000
11.00001.00001.00001.00001.0000
20.99980.99980.99990.99990.9999
31.00001.00001.00001.00001.0000
40.99991.00001.00001.00001.0000
51.00001.00001.00001.00001.0000
61.00001.00001.00001.00001.0000
71.00001.00001.00001.00001.0000
81.00000.99991.00001.00001.0000
91.00001.00001.00001.00001.0000
101.00001.00001.00001.00001.0000
Average Accuracy0.98850.98940.99050.99150.9927
Table 5. Under the average accuracy threshold of 80%, the average recognition accuracy of our proposed method, the OpenMax-based method and MLS-based method on the corresponding test dataset with 5 different conditions. The results for each condition are based on the average results of all situations in the corresponding condition. For Condition 1 to Condition 5, the number of situations is 11, 55, 165, 330, and 462, respectively.
Table 5. Under the average accuracy threshold of 80%, the average recognition accuracy of our proposed method, the OpenMax-based method and MLS-based method on the corresponding test dataset with 5 different conditions. The results for each condition are based on the average results of all situations in the corresponding condition. For Condition 1 to Condition 5, the number of situations is 11, 55, 165, 330, and 462, respectively.
Method
(Average Accuracy Threshold of 80%)
Condition 1Condition 2Condition 3Condition 4Condition 5
KnownUnknownKnownUnknownKnownUnknownKnownUnknownKnownUnknown
Ours0.78690.67620.78760.67840.78840.68540.78910.68960.79000.6977
OpenMax0.80630.62070.80580.60950.80790.59600.81000.58380.81240.5811
MLS0.80770.71830.79830.76570.79920.72080.79960.68790.80030.6684
Table 6. Under the average accuracy threshold of 85%, the average recognition accuracy of our proposed method, the OpenMax-based method and MLS-based method on the corresponding test dataset with 5 different conditions. The results for each condition are based on the average results of all situations in the corresponding condition. For Condition 1 to Condition 5, the number of situations is 11, 55, 165, 330, and 462, respectively.
Table 6. Under the average accuracy threshold of 85%, the average recognition accuracy of our proposed method, the OpenMax-based method and MLS-based method on the corresponding test dataset with 5 different conditions. The results for each condition are based on the average results of all situations in the corresponding condition. For Condition 1 to Condition 5, the number of situations is 11, 55, 165, 330, and 462, respectively.
Method
(Average Accuracy Threshold of 85%)
Condition 1Condition 2Condition 3Condition 4Condition 5
KnownUnknownKnownUnknownKnownUnknownKnownUnknownKnownUnknown
Ours0.84680.63370.83770.64440.83860.65240.83940.65750.84030.6665
OpenMax0.85170.53570.85150.53080.85190.52790.85270.52480.85320.5265
MLS0.85470.68420.85480.72310.85570.67970.85620.64550.85690.6267
Table 7. Under the average accuracy threshold of 90%, the average recognition accuracy of our proposed method, the OpenMax-based method and MLS-based method on the corresponding test dataset with 5 different conditions. The results for each condition are based on the average results of all situations in the corresponding condition. For Condition 1 to Condition 5, the number of situations is 11, 55, 165, 330, and 462, respectively.
Table 7. Under the average accuracy threshold of 90%, the average recognition accuracy of our proposed method, the OpenMax-based method and MLS-based method on the corresponding test dataset with 5 different conditions. The results for each condition are based on the average results of all situations in the corresponding condition. For Condition 1 to Condition 5, the number of situations is 11, 55, 165, 330, and 462, respectively.
Method
(Average Accuracy Threshold of 90%)
Condition 1Condition 2Condition 3Condition 4Condition 5
KnownUnknownKnownUnknownKnownUnknownKnownUnknownKnownUnknown
Ours0.90000.58910.89050.60310.89150.61210.89240.61820.89340.6283
OpenMax0.90140.40140.90170.41320.90180.42580.90140.43330.90120.4353
MLS0.90590.63800.90600.67040.90710.63040.90770.59440.89870.5880
Table 8. Under the average accuracy threshold of 95%, the average recognition accuracy of our proposed method, the OpenMax-based method and MLS-based method on the corresponding test dataset with 5 different conditions. The results for each condition are based on the average results of all situations in the corresponding condition. For Condition 1 to Condition 5, the number of situations is 11, 55, 165, 330, and 462, respectively.
Table 8. Under the average accuracy threshold of 95%, the average recognition accuracy of our proposed method, the OpenMax-based method and MLS-based method on the corresponding test dataset with 5 different conditions. The results for each condition are based on the average results of all situations in the corresponding condition. For Condition 1 to Condition 5, the number of situations is 11, 55, 165, 330, and 462, respectively.
Method
(Average Accuracy Threshold of 95%)
Condition 1Condition 2Condition 3Condition 4Condition 5
KnownUnknownKnownUnknownKnownUnknownKnownUnknownKnownUnknown
Ours0.95300.52480.95390.53240.94490.55760.94590.56520.94700.5767
OpenMax0.95330.16670.95110.21770.95220.24070.95130.26690.94940.2793
MLS0.95150.57500.95190.59510.95300.56000.95390.52420.95480.5086
Table 9. Under the average accuracy threshold of 80%, the average recognition accuracy in the corresponding test dataset with 5 different conditions based on the ablation models and our proposed model. The results for each condition are based on the average results of all situations in the corresponding condition. For Condition 1 to Condition 5, the number of situations is 11, 55, 165, 330, and 462, respectively.
Table 9. Under the average accuracy threshold of 80%, the average recognition accuracy in the corresponding test dataset with 5 different conditions based on the ablation models and our proposed model. The results for each condition are based on the average results of all situations in the corresponding condition. For Condition 1 to Condition 5, the number of situations is 11, 55, 165, 330, and 462, respectively.
Method
(Average
Accuracy Threshold of 80%)
Condition 1Condition 2Condition 3Condition 4Condition 5
KnownUnknownKnownUnknownKnownUnknownKnownUnknownKnownUnknown
Ours0.78690.67620.78760.67840.78840.68540.78910.68960.79000.6977
AM10.80990.36960.80060.38170.80150.38180.80230.38150.80330.3790
AM20.80060.44270.80140.45040.80220.45150.80310.45200.80400.4516
AM30.79960.59930.80030.60270.80110.60910.80190.61080.80280.6134
U-Net0.80240.33330.80310.32730.80390.33200.80470.33450.80570.3371
UNet3+0.79980.36050.80050.36450.80140.36960.80220.37170.80320.3742
Table 10. Under the average accuracy threshold of 85%, the average recognition accuracy in the corresponding test dataset with 5 different conditions based on the ablation models and our proposed model. The results for each condition are based on the average results of all situations in the corresponding condition. For Condition 1 to Condition 5, the number of situations is 11, 55, 165, 330, and 462, respectively.
Table 10. Under the average accuracy threshold of 85%, the average recognition accuracy in the corresponding test dataset with 5 different conditions based on the ablation models and our proposed model. The results for each condition are based on the average results of all situations in the corresponding condition. For Condition 1 to Condition 5, the number of situations is 11, 55, 165, 330, and 462, respectively.
Method
(Average
Accuracy Threshold of 85%)
Condition 1Condition 2Condition 3Condition 4Condition 5
KnownUnknownKnownUnknownKnownUnknownKnownUnknownKnownUnknown
Ours0.84680.63370.83770.64440.83860.65240.83940.65750.84030.6665
AM10.84900.34210.84980.34680.85080.34730.85170.34720.85260.3446
AM20.84960.40650.85040.41410.85130.41580.85220.41680.85320.4168
AM30.84890.56200.84960.56640.85050.57360.85140.57600.85230.5793
U-Net0.85000.27680.85080.27200.85170.27620.85260.27910.85360.2816
UNet3+0.84940.30420.85010.30820.85110.31350.85190.31620.85290.3189
Table 11. Under the average accuracy threshold of 90%, the average recognition accuracy in the corresponding test dataset with 5 different conditions based on the ablation models and our proposed model. The results for each condition are based on the average results of all situations in the corresponding condition. For Condition 1 to Condition 5, the number of situations is 11, 55, 165, 330, and 462, respectively.
Table 11. Under the average accuracy threshold of 90%, the average recognition accuracy in the corresponding test dataset with 5 different conditions based on the ablation models and our proposed model. The results for each condition are based on the average results of all situations in the corresponding condition. For Condition 1 to Condition 5, the number of situations is 11, 55, 165, 330, and 462, respectively.
Method
(Average
Accuracy Threshold of 90%)
Condition 1Condition 2Condition 3Condition 4Condition 5
KnownUnknownKnownUnknownKnownUnknownKnownUnknownKnownUnknown
Ours0.90000.58910.89050.60310.89150.61210.89240.61820.89340.6283
AM10.90760.29370.89920.30660.90010.30740.90110.30770.90210.3053
AM20.90770.36160.89910.37650.90010.37880.90100.38010.90210.3803
AM30.89890.51610.89980.52160.90070.52990.90170.53320.90270.5374
U-Net0.89890.21330.89970.20860.90070.21240.90160.21540.90260.2176
UNet3+0.89870.24370.89950.24790.90050.25350.90140.25670.90250.2601
Table 12. Under the average accuracy threshold of 95%, the average recognition accuracy in the corresponding test dataset with 5 different conditions based on the ablation models and our proposed model. The results for each condition are based on the average results of all situations in the corresponding condition. For Condition 1 to Condition 5, the number of situations is 11, 55, 165, 330, and 462, respectively.
Table 12. Under the average accuracy threshold of 95%, the average recognition accuracy in the corresponding test dataset with 5 different conditions based on the ablation models and our proposed model. The results for each condition are based on the average results of all situations in the corresponding condition. For Condition 1 to Condition 5, the number of situations is 11, 55, 165, 330, and 462, respectively.
Method
(Average
Accuracy Threshold of 95%)
Condition 1Condition 2Condition 3Condition 4Condition 5
KnownUnknownKnownUnknownKnownUnknownKnownUnknownKnownUnknown
Ours0.95300.52480.95390.53240.94490.55760.94590.56520.94700.5767
AM10.95600.24570.94730.26150.94840.26280.94940.26330.95050.2612
AM20.95630.30890.94730.32760.94840.33040.94940.33220.95050.3326
AM30.95760.44280.94920.46500.95020.47440.95120.47870.95230.4837
U-Net0.95630.12230.94720.13710.94830.14000.94920.14270.95030.1442
UNet3+0.95650.15470.95740.15820.94870.18090.94970.18480.95080.1890
Table 13. The average reconstruction loss in the corresponding validation dataset based on the ablation models and our proposed model. As there are 11 types of intra-pulse modulations, the results for each reconstruction model are based on the average results of 11 sub-reconstruction encoder–decoder models. As the original loss is based on MSE, which is too small, we report it by magnifying itself 16,384 times (equal to 128 × 128).
Table 13. The average reconstruction loss in the corresponding validation dataset based on the ablation models and our proposed model. As there are 11 types of intra-pulse modulations, the results for each reconstruction model are based on the average results of 11 sub-reconstruction encoder–decoder models. As the original loss is based on MSE, which is too small, we report it by magnifying itself 16,384 times (equal to 128 × 128).
Reconstruction ModelOursAM1AM2AM3U-NetUNet3+
Average Reconstruction Loss
(The original value based on MSE is magnified 16,384 times (equals to 128 × 128))
36.77774.501510.347024.07210.02940.0333
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Yuan, S.; Li, P.; Wu, B. Radar Emitter Signal Intra-Pulse Modulation Open Set Recognition Based on Deep Neural Network. Remote Sens. 2024, 16, 108. https://doi.org/10.3390/rs16010108

AMA Style

Yuan S, Li P, Wu B. Radar Emitter Signal Intra-Pulse Modulation Open Set Recognition Based on Deep Neural Network. Remote Sensing. 2024; 16(1):108. https://doi.org/10.3390/rs16010108

Chicago/Turabian Style

Yuan, Shibo, Peng Li, and Bin Wu. 2024. "Radar Emitter Signal Intra-Pulse Modulation Open Set Recognition Based on Deep Neural Network" Remote Sensing 16, no. 1: 108. https://doi.org/10.3390/rs16010108

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop