Next Article in Journal
A Comparison of Gravimetric, Isostatic, and Spectral Decomposition Methods for a Possible Enhancement of the Mantle Signature in the Long-Wavelength Geoidal Geometry
Next Article in Special Issue
Multi-Label Radar Compound Jamming Signal Recognition Using Complex-Valued CNN with Jamming Class Representation Fusion
Previous Article in Journal
Remote Sensing Monitoring of Drought in Southwest China Using Random Forest and eXtreme Gradient Boosting Methods
Previous Article in Special Issue
Decentralized Approach for Translational Motion Estimation with Multistatic Inverse Synthetic Aperture Radar Systems
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Radar Emitter Structure Inversion Method Based on Metric and Deep Learning

1
Key Laboratory of Advanced Marine Communication and Information Technology, Ministry of Industry and Information Technology, Harbin Engineering University, Harbin 150001, China
2
Harbin Institute of Technology, Harbin 150001, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2023, 15(19), 4844; https://doi.org/10.3390/rs15194844
Submission received: 29 August 2023 / Revised: 22 September 2023 / Accepted: 4 October 2023 / Published: 6 October 2023
(This article belongs to the Special Issue Advanced Radar Signal Processing and Applications)

Abstract

:
With the rapid development of modern military countermeasure technology, deep distinguish hostile radar is essential in electronic warfare. However, traditional radio frequency (RF) feature extraction methods can easily be interfered by signal information and fail due to the lack of research on RF feature extraction techniques for complex situations. Therefore, in this paper, first, the generation mechanism of RF structure information is discussed, and the influence of different signal information introduced by different operating parameters on RF structure feature extraction is analyzed. Then, an autoencoder (AE) network and an autoencoder metric (AEM) network are designed, introducing metric learning ideas, so that the extracted deep RF structure features have good stability and divisibility. Finally, radar emitter structure (RES) inversion is realized using the centroid-matching method. The experimental results demonstrate that this method exhibits good inversion performance under variable operating parameters (modulation type, frequency, bandwidth, input power). RES inversion including unknown operating parameters is realized for the first time, and it is shown that metric learning has the advantage of separability of RF feature extraction, which can provide an idea in emitter and RF feature extraction.

1. Introduction

A radar emitter is the core component of a radar system. It produces a radio frequency (RF) signal with sufficient power and transmits it through an antenna. To avoid being intercepted by electronic reconnaissance equipment, hostile radar emitter structures (RES) have become increasingly diverse. In this paper, the RES can contain many different radar emitter individuals (REI), and they are considered as a class of structures. Thus, in complex electronic reconnaissance situations, the hostile radar cannot be deeply recognized, through only the radar modulation type identification [1,2] and REI identification [3,4,5,6,7,8,9,10,11]. Therefore, this paper studies methods for inverting RES from RF signals emitted by radar emitters to provide more information at a new level on electronic reconnaissance systems. There are essential differences between REI identification and RES inversion. REI identification is the identification of differences between individuals in different radar emitters, while RES inversion focuses on the differences in radar emitter structures. Thus, REI identification and RES inversion are studies of radar emitters at different levels, and they have essential differences. At this moment, we are only interested in RES and not REI. Simultaneously, if the same class of RES contains different REIs, it will introduce certain challenges to RES inversion.
The non-ideal characteristics generated by each REI as “RF individual information” and the overall non-ideal characteristics generated by each class of RES are called “RF structure information” in this paper. In addition, each class of RES has unique inherent characteristics. Different classes of RES have different physical mechanisms and different RF structure information. At the same time, the radar-emitted RF signal will carry these unique RF structure features; therefore, RES inversion can be achieved using RF structure features.
Meanwhile, with the emergence of multifunctional radars, radar-emitted RF signals are becoming increasingly complex. Signals with different modulation types and operating parameters are used when performing different tasks. Therefore, different operating parameters can cause signal differences between RF signals, which are called “signal information” in this paper. Therefore, RF signals are mainly influenced by two factors: signal information and RF information. The complex situation is that the parameters’ change randomly introduces great challenges to the electronic reconnaissance system.
The performance of RES inversion depends on the stability and separability of extracted features. At present, feature extraction methods are mainly divided into the time domain [12], frequency domain [13], time–frequency domain [14], and transform domain [15,16,17,18,19,20,21]. The shortcoming of these methods is their reliance on professional knowledge and experience for manual feature extraction and design. In recent years, deep learning technology [22,23] has been widely used in various fields, including speech [24] and image [25] processing. It has also been widely used in radar emitter identification fields [26,27,28,29,30,31], integrating artificial intelligence technology with radar identification. Specially in the study of REI identification [26,29,32], in general, various neural networks can effectively carry out corresponding tasks under the condition of having good training data and corresponding labels (i.e., closed-set problem). Meanwhile, as an unsupervised network, the autoencoder (AE) [33] has been proven to have excellent performance in feature representation. Different improvements have been made for different situations, such as sparse autoencoders (SAE) [34,35,36], stacked sparse autoencoders (sSAE) [37,38], denoising autoencoders (DAE) [39,40], and compression autoencoders (CAE) [41]. However, unsupervised learning does not always produce the desired results in the case of unknown operating parameters (i.e., open set problem). If labels are used to enhance the separability of features, combining supervised and unsupervised training can improve the performance of the neural network.
However, whether using traditional feature extraction methods or neural networks, these methods mainly focus on the homo-mode operating parameters (constant operating parameters), in which conditions are too ideal. What needs to be further discussed is whether these methods have good generalization performance under the hetero-mode operating parameters (variable operating parameters). Thus, the main problem of these methods involves the extraction of the RF features of the RF signal directly. However, these methods do not pay attention to the interference introduced by signal information. Does the extracted RF feature contain a large amount of signal information?
Therefore, the problem to be solved in this paper is as follows: how can we reduce the interference of signal information in complex situations so that the extracted RF structure features have good stability and separability? Moreover, how can these RF structure features be used to achieve RES inversion while retaining the RES inversion algorithm’s good robustness and generalization performance under unknown operating parameters? As far as we know, there are no relevant papers that have examined this issue. It is worth noting that this complexity includes the complexity of RES modeling (we invert the twelve radar emitters into three classes of structures) and the complexity of the variable operating parameters (modulation type, frequency, bandwidth, and input power).
In summary, it is challenging to realize RES inversion in complex situations, and the traditional method can easily fail due the to interference of signal information due to the lack of influence of signal information under variable parameters on extraction of RF features. In the field of radar signal recognition, the closed-set problem of signal modulation type recognition refers to the recognition of known modulation type signals, while the open-set problem refers to the recognition of unknown modulation type signals [42], and its open-set problem is more challenging. Researchers have introduced metric learning into signal modulation type recognition and demonstrated the advantages of triplet loss functions in feature extraction [28]. This paper conducts an RES inversion study on closed-set (known operating parameters) and open-set (unknown operating parameters) problems using hetero-mode operating parameters, which combines metric learning and deep learning techniques to study how effectively to reduce the interference caused by signal information. We attempt to find methods that can render the neural network more attentive toward the common RF structure information of RF signals among these hetero-mode operating parameters in the learning process; we also attempt to render the neural network insensitive to changes in terms of operating parameters, and we provide good stability and separability for extracted RF structure features, which is the key to RES inversion. Figure 1 shows the overall flowchart of the method proposed. First, RF signals are obtained using RES modeling. Then, the AEM network is constrained in the feature domain using the deep signal features extracted from the AE network in the data domain, and the RF signal is reconstructed as unsupervised to weaken the signal information in the feature domain of the RF signal under hetero-mode operating parameters; RF structure information is highlighted. Distinct from traditional algorithms, the RF structure features automatically extracted via deep learning are called “deep RF structure features”, and the signal features are called “deep signal features” in this paper.
The main contributions of this paper are summarized as follows:
  • Stability: A dual neural network (AE and AEM network) is designed based on the idea of unsupervised learning. The use of deep learning techniques to reduce the interference caused by signal information is essential for extracting RF features that are not sensitive to parameter variations. This idea is instructive for RF feature extraction techniques under variable operating parameters.
  • Separability: Based on the idea of supervised learning, a metric scorer is designed in the AEM network using the metric learning approach; using this, the network shortens the distance of RF signals in the feature domain between the hetero-mode operating parameters of the same class of RES during the learning process, while widening the distance between RF signals in different classes of RES to improve the separability of the deep RF structure features learned from RF signals.
  • Unsupervised and supervised joint optimization methods are designed to optimize the objective function of AEM networks.
  • Different from the traditional REI method used with constant operating parameters, the RES inversion method using variable operating parameters (modulation type, frequency, bandwidth, and input power) is proposed for the first time, and centroid-matching method is used for RES inversion from radar-transmitted RF signals in the test stage. Moreover, the proposed method maintains good generalization performance under unknown operating parameters.
The remainder of this paper is organized as follows. Section 2 introduces the problem’s definition. Section 3 presents and verifies RES modeling. Section 4 presents the proposed RES inversion method. Section 5 details the experimental results and analysis of the proposed method. Finally, the conclusions are drawn in Section 6.

2. Problem Definition

A radar emitter is the core component of a radar system. Radar emitters are typically classified as pulse-modulated emitters and continuous wave emitters. Pulse-modulated emitters are usually categorized into single-stage oscillating emitters and main vibration amplification emitters. Most radars, particularly those that are highly stable and those with high-performance tracking, telemetry, and control (TT&C), as well as phased array radars, have adopted a main vibration amplification radar emitter. Recently, with the rapid development of gallium arsenide (GaAs) and gallium nitride (GaN), studies on all-solid-state radar emitters in the C-band and X-band have approached the practical stage. All all-solid-state radar emitters are generally classified into two categories: a high-power solid-state emitter with a centralized synthesis output structure and a distributed synthetic phased radar emitter. Therefore, there is a variety of RES due to these different structure design parameters.
Generally, the radar emitter source is mainly composed of a signal source and an RF link module [43,44]. The signal source generates a source signal of the desired frequency. The RF mixer and RF amplifier are the core modules of the RF link. The RF mixer is responsible for moving the intermediate frequency signal spectrum to the RF area, and the RF amplifier is responsible for amplifying the input RF signal. Due to the defects of the underlying physical components, the signal source has phase truncation errors, which makes RF signals appear as “burrs” (spurious information). RF modules, such as RF mixers and RF amplifiers, have nonlinear distortion and phase noise, which will produce high harmonics and intermodulation distortion. The phase noise level of the RF mixer can introduce significant differences to the RF signal in both time and time–frequency domains. The different third-order intercept points (IP3) of the RF amplifier will cause different amplitudes in the RF signal.
Because different classes of RES are composed of different structure design parameters, such as RES1 having only one stage of amplification while RES2 is designed for two-stage amplification, there are structure differences between different classes of RES. At the same time, under the same class of RES, different individuals have the same structure design parameters and use the same components, but because there are subtle differences between each component; this subtle difference is called “RF individual information” in this paper, and it is also called an “RF fingerprint” [45]. This results in individual differences among different individuals under the same RES. Therefore, there are essential differences between REI identification and RES inversion in this paper.
As shown in Figure 2, red, yellow, and green represent three classes of RES, and numbers 1, 2, and 3 correspond to three different operating parameters, while the number 4 corresponds to unknown operating parameters. The RF structure information carried by RF signals and transmitted by radar emitters with different structures is noticeably different under ideal conditions. All types of neural networks can be used for classification and recognition if using the homo-mode operating parameter—that is, 1, 2, and 3 are the same operating parameters—and the training data and corresponding labels are good.
With the emergence of multi-functional and phased array radars, operating parameters in relation to different tasks can be changed. For instance, multi-functional radars support various modes, such as tracking and searching, single-target tracking, multi-target tracking, and others. The performance of the neural network will be seriously affected either by RES inversion or by REI identification under hetero-mode operating parameters. Next, the influence of the change in operating parameters upon RES inversion is analyzed.
Question 1: Training data question: Although a large amount of training data can be obtained under certain known operating parameters (e.g., the frequency is 25 MHz and 26 MHz), it is not possible to collect all training data within all changing ranges of operating parameters (e.g., the frequency is 27 MHz). Additionally, samples with different operating parameters are often imbalanced (for example, there are more samples at 25 MHz than at 26 MHz) due to the limitations of the working scene, parameter selection and data collection conditions, and other factors. Therefore, a complete database cannot be easily established using electronic reconnaissance systems.
Question 2: The influence of hetero-mode operating parameters: With the improvement in the circuit integration process, the difference in nonlinear characteristics between components is decreasing, but radar operating parameters are constantly changing, and the difference gradually increases. As a result, if the same class of RES transmits RF signals with different operating parameters, the signal information carried by the RES may cover the RF structure information, or it may completely exceed the RF structure information. If different classes of RES transmit RF signals with different operating parameters, the signal information carried by the different classes of RES may exceed the RF structure information that is carried. As a result, the neural network might pay more attention to each signal information, which results in the performance of the neural network being significantly affected by the change in operating parameters. When there are different REIs under the same class of RES, difficulties in classification greatly increases under hetero-mode operating parameters.
Question 3: The influence of unknown operating parameters: When receiving an RF signal with unknown operating parameters transmitted by a known RES, the neural network maps the training data domain to the feature domain completely according to the known operating parameters and determines the classification boundary. When the operating parameters of the received RF signal are not within the operating parameters range of the training data, can this neural network be used to map the RF signal with unknown operating parameters to the feature domain of known operating parameters? Will the neural network make false judgments?
Question 4: The influence of noise: When noise levels are high in the real environment, both the RF signal and structure information may be covered at the same time. It is unknown whether noise will cause the neural network to be unable to extract meaningful features, thereby making the performance of the neural network more sensitive to noise and less robust.
Therefore, the problem to be solved in this paper is to extract the good stability and separability of features that represent RF structure information under this hetero-mode operating parameter; then, this feature is used to invert the RES and make the neural network insensitive to changes in the operating parameters.

3. RES Modeling

The importance of RES modeling lies in the construction of a dataset that corresponds to the actual situation and the generation of the required RF signal using the structure feature level modeling method. This is a forward modeling process that provides a basis for subsequent inversion methods. Having a high-quality dataset is the premise and basis for training neural networks. However, due to the confidentiality of military systems, it is very difficult to obtain RF signals under such hetero-mode operating parameters. Therefore, simulation data [46,47] are mostly used to conduct experiments on electronic reconnaissance systems. Hence, RES modeling is the premise of obtaining a simulation dataset, which is of great importance. Large amounts of data can be obtained via the modeled RES.
The current methods for constructing simulation data are mainly classified into two categories: single-device modeling and mathematical modeling. In single-device modeling, power amplifiers are mostly studied [48,49,50,51], and behavioral modeling, like Saleh [49], polynomial [50], and Volterra series [51] models, is usually used to simulate nonlinear features and memory effect. In mathematical modeling, there are mainly two types of models: a time-domain envelope model that can reflect the transient and steady-state features of the radar emitter [52], and the phase noise model [53] with the frequency shift mode [54] relative to frequency stability that can reflect the spectrum purity of the radar-emitted signal. Their main shortcomings are that they can only reflect the features of a certain module in the RES model but cannot fully reflect its overall features.
Therefore, to provide basic conditions for theoretical analyses and the practical verification of this study, this paper builds three classes of RES models by selecting different structure design parameters and using a simulation platform with measured parameters according to the working principle of radars. The S2P file is used for simulation, and the S2P file can be approximately considered as the reflection of a field measurement component in software. Therefore, the simulation data are closer to real data. In addition, under the same RES model, four REI models are built by slightly adjusting the parameters of each module. As a result, there are structural differences among the RES models of different classes and individual differences among different REI models of the same class when building RES in complex situations.

3.1. RES Model

As shown in Figure 3, a signal source model and three RF link models (PA1, PA2, and PA3 models) are built. Three RES models (RES1, RES2, and RES3) of different classes are developed by combining the signal source and RF link models. Each RES model includes two parts, the signal source model and RF link model, which are introduced in detail in the following sections.

3.1.1. Signal Source Model

Considering that direct digital synthesizer (DDS) technology has become the mainstream of signal sources in modern radar emitters, in the actual circuit, the DDS model increases the number of phase accumulator bits in order to improve frequency resolutions. In order to reduce the ROM’s storage capacity, the higher A bits of the N bits are selected to address the amplitude information in the ROM, and the B = N A bits are truncated, which introduces a phase truncation error [55].
Therefore, the DDS model [55], widely used in the radar application field, was selected as the signal source to generate continuous wave (CW), linear frequency modulation (LFM), binary phase shift keying (BPSK), and quadrature phase shift keying (QPSK) signals using spurious information from the four DDS models of different modulation types, which were built according to the principle of phase truncation error: DDS_LFM model, DDS_BPSK model, and DDS_QPSK model.
DDS_CW model: According to the CW signal simulation principle, the frequency of the DDS_CW signal generated by the DDS_CW model is as follows:
f CW = ( K CW / 2 N ) f c ,
where N represents the bits number of the phase accumulator, f c denotes the frequency of the reference clock source, and K CW is the frequency control word.
DDS_ LFM model: According to the LFM signal modulation principle, the DDS_LFM model is developed to generate the DDS_LFM signal. To perform the function of linear frequency modulation, the result of the frequency accumulator exhibits a linearly increasing characteristic, while the result of the phase accumulator exhibits the characteristic of a quadratic function. The initial frequency f LFM and bandwidth B LFM of the DDS_LFM signal are expressed as follows:
f LFM = ( K LFM / 2 N ) f c
B LFM = ( K 1 LFM / 2 N ) f c 2 T ,
where K LFM represents the frequency control word, K 1 LFM denotes the chirp rate control word (i.e., frequency step word), and T denotes the pulse width.
DDS_BPSK model: The DDS_BPSK model is developed according to the BPSK signal modulation principle. The DDS_CW signal is used as a carrier signal, and the DDS_BPSK signal is generated using a 7-bit Barker code ( 0001101 ) .
DDS_QPSK model: The DDS_QPSK model is developed according to the QPSK signal modulation principle. The DDS_CW signal is used as a carrier signal, and the DDS_QPSK signal is generated using the phase encoding ( 0213030303120 ) .

3.1.2. RF Link Model

Practically, the signal generated by the signal source should undergo multi-stage amplification to achieve the transmission power required for the transmission distance. The parameter simulation of each component is performed using the S2P file. The latter is composed of the S parameter ( S 11 , S 12 , S 21 , S 22 ) matrix, and the value of the S parameter is obtained by the manufacturer’s field measurement of the component. Therefore, an S2P file can be approximately considered as the reflection of a field measurement component in the software.
The RF link model is built using different amplification stages and module parameters. The RF mixer module is set as a linear device, and the RF amplifier module is set as a nonlinear device, which includes the following: the PA1 model is a first-stage amplifier, which uses the simulation amplifier; the PA2 model is a two-stage amplifier, which uses the HMC637ALP5E GaAs power amplifier and simulation amplifier; and the PA3 model is a three-stage amplifier, which uses the HMC637ALP5E GaAs power amplifier, the HMC8500PM5E GaN power amplifier, and the simulation amplifier.
The individual differences of each class of RES models are simulated by slightly adjusting the phase noise level and frequency offset of the RF mixer and the IP3 value of the RF amplifier. The specific parameters of each module are shown in Table 1. The reference impedance is 50 ohms, the noise figure is 8 dB, and the operating frequency of these devices is 2.1 GHz.

3.2. RF Signal

Based on the above modeling, different operating parameters, namely, the modulation type (MT), frequency (F), bandwidth (B), and input power (P), are selected to generate source signals with four DDS signal source models, and complex signals are obtained using Hilbert transformations as input signals to the RF link model. They are then amplified by PA1, PA2, and PA3 models and output as RF signals of three classes of RES models.
As a result, RF signals generated by modeling are composed of different structure design parameters and different operating parameters. Different operating parameters cause differences in signal information between RF signals, and different structure design parameters cause differences in RF structure information between RF signals, as shown in Figure 4. The modeling dataset for complex situations is generated. This includes three classes of RES models and four REI models under the same class of RES model, i.e., a total of twelve radar emitters.

4. RES Inversion Method

As observed in the above modeling, different structure design parameters and operating parameters can produce various RF signals that carry different RF structure information and signal information, or they map the parameter domain to the data domain. If the data domain can be mapped to the feature domain so that the neural network can learn the relevant features representing the RF structure information and signal information, it will be conducive to the RES inversion.
Then, by comparing the channel system [56], the RES can be regarded as an RF structure system, R F S ( · ) , which carries RF structure information. Different classes of RES correspond to different RF structure systems because they are composed of different structure design parameters and carry different RF structure information. This is a complex nonlinear and unknown system; the specific parameters of the system cannot be obtained via simple calculations.
The ideal RF signal x without RF structure information, that is, R F S ( x ) = x , can be considered as the baseband signal s with different operating parameters after passing through the modulation system M T ( · ) :
x = M T ( s ) .
Therefore, the RF signal y carrying RF structure information can be considered as the ideal RF signal obtained via the RF structure system:
y = R F S ( x ) = R F S ( M T ( s ) ) .
Based on the above analysis, a dual neural network is designed in this section. An AE network aims to train a feature encoder that is capable of extracting deep signal features in an unsupervised manner using ideal RF signals. As for the AEM network, the feature encoder of the AE network is used to extract the deep signal features, constrain the AEM network in the feature domain, and reconstruct the RF signal in an unsupervised manner in order to learn the potential RF structure features that can stably characterize the RES from the data domain. The purpose of this step is to map the data domain to the feature domain and back to the data domain via the powerful nonlinear processing ability of the neural network, thereby mapping the RF structure information to the deep RF structure features one by one. Then, a metric scorer is designed using metric learning, and the AEM network’s structure and objective function are optimized end-to-end, resulting in deep RF structure features for each class of RES that are close to the centroid of the deep RF structure features of the same class of RES and far from the centroid of the deep RF structure features of other classes of RES. During the learning process, the network shortens the distance between the RF signals of hetero-mode operating parameters in the same class of RES in the feature domain and widens the distance between RF signals in the feature domain of different classes of RES. This dual neural network is described in detail in the sequel. To illustrate the RES inversion method in the simplest form, consider a deep neural network as an example, where each module contains two hidden layers. In specific cases, one can choose more complex neural networks, such as convolutional neural network (CNN) or long short-term memory (LSTM) [57].

4.1. AE Network

As shown in Figure 5, network 1 is an AE network composed of a feature encoder f e and a reconstruction decoder f d .
First, the input to the AE network is an ideal RF signal without RF structure information x .
Then, the hidden layer dimension is set to be smaller than the input dimension. The feature encoder automatically learns the compression feature of the RF signal, which is h s i g —the deep signal feature. The RF signal x ^ is then reconstructed from this feature unsupervised via the reconstruction decoder, thereby realizing the mapping from the data domain to the feature domain and then back to the data domain.
The coding and decoding process of the AE network can be described as follows:
h s i g = f e ( x , θ e )
x ^ = f d ( h s i g , θ d ) ,
where θ e represents the feature encoder parameters, and θ d represents the reconstruction decoder parameters.
This reconstruction method can ensure that the loss of learned feature information is minimal. To train the AE network, it is necessary to define an objective function; that is, the reconstruction error:
J A E ( { θ e , θ d } ) = 1 I i = 1 I L ( x , x ^ ) ,
where I is the number of input signal samples, and L ( x , x ^ ) represents the loss function. Common forms such as the mean square error (MSE) and cross entropy (CE) can be selected. This paper adopts MSE:
L ( x , x ^ ) = | | x x ^ | | 2 .
Therefore, the reconstruction objective function based on the MSE loss function is expressed as follows:
J A E ( { θ e , θ d } ) = 1 I i = 1 I | | x f d ( f e ( x , θ e ) , θ d ) | | 2 .
The exact solution of the input RF signal obtained via the AE network is as follows:
x ^ = f d ( f e ( x , θ e ) , θ d ) .
Finally, the optimization method is used to train the AE network continuously, and the trained feature encoder is saved.

4.2. AEM Network

As shown in Figure 6, the AEM network is composed of a feature encoder g e , a reconstruction decoder g d , and a metric scorer g s . In the following sections, this network is described in detail, including the input signal, network structure, and objective function.

4.2.1. Feature Encoder

First, the feature encoder of the AEM network has the same design as the feature encoder of AE, but the input is an RF signal y carrying RF structure information.
Then, g e automatically learns the deep RF structure feature h s t r of the RF signal. The mapping between the data domain and feature domain is realized:
h s t r = g e ( y , ϑ e ) ,
where ϑ e represents the feature encoder parameter.
After this steap, h s t r is fed into the two branches of the reconstruction decoder and the metric scorer. The RF signal is reconstructed in the feature domain unsupervised, aiming at obtaining stable extracted deep RF structure features. In addition, the similarity score is computed using the supervised metric method in the feature domain, aiming at obtaining the good separability of the extracted deep RF structure features.

4.2.2. Reconstruction Decoder

The RF signal y is input into the feature encoder of the AEM network and the feature encoder of the trained AE network to obtain deep RF structure features and deep signal features. In the reconstruction decoder, h s t r and h s i g are spliced in parallel as the input of the reconstruction decoder:
h r d = [ h s t r ; h s i g ] .
The reconstruction decoder reconstructs the RF signal y ^ from h r d in an unsupervised manner, thereby realizing the mapping between the feature domain and data domain. The decoding process can be described as follows:
y ^ = g d ( h r d , ϑ d ) ,
where ϑ d represents the reconstruction decoder parameter.
In addition, the reconstruction objective function is designed based on the MSE loss function:
J r d ( { ϑ e , ϑ d } ) = 1 I i = 1 I L ( y , y ^ ) = 1 I i = 1 I | | y g d ( g e ( y , ϑ e ) , ϑ d ) | | 2 .

4.2.3. Metric Scorer

Metric learning is one of the core problems of pattern recognition. To measure the similarity between samples, metric learning can also be considered as similarity. Its purpose is to minimize the distance between the same class of samples and increase the distance between different classes of samples. Therefore, the metric learning method selected in this paper aims to improve the feature domain separability of the extracted deep RF structure features.
Drawing on the design concept of the generalized end-to-end loss function [58], the input of the metric scorer is designed as samples of a batch, including N different classes of RES and M samples under each class, as shown in Figure 7.
First, ( N × M ) RF signals y j i ( 1 j N , 1 i M ) are acquired to build a batch:
[ ( y 11 , y 12 , , y 1 M ) , ( y 21 , y 22 , , y 2 M ) , , ( y N 1 , y N 2 , , y N M ) ] ,
which are then fed into the feature encoder to extract their ( N × M ) deep RF structure features. h s t r _ j i represents the deep RF structure features of the i-th RF signal of the j-th class of RES.
Then, the centroids of the deep RF structure features of the j-th class of RES are calculated:
c k = E m [ h s t r _ j m ] = 1 M m = 1 M h s t r _ j m , ( 1 j , k N , 1 i M ) ,
where h s t r _ j m = [ h s r t _ j 1 , h s r t _ j 2 , , h s r t _ j M ] .
In addition, it is noted that the deletion of a term when calculating the centroid of the real RES can stabilize the training process and help avoid trivial solutions [58]. Therefore, when k = j , the centroid of deep RF structure features is calculated using the following formula:
c j ( i ) = 1 M 1 m = 1 , m i M h s t r _ j m .
In addition, the deep RF structure features h s t r _ j i of each RF signal and the centroids c k of all classes are combined in series to form ( N × M × N ) RF feature pairs as the input of the metric scorer.
h m s _ j i k = [ h s t r _ j i , c j ( j ) ] i f k = j ; [ h s t r _ j i , c k ] o t h e r w i s e .
The similarity score of each RF feature pair is calculated using a metric scorer, and the mapping between the feature domain and metric domain is realized:
s j i k = g s ( h m s _ j i k , ϑ s ) ,
where ϑ s represents the metric scorer parameter.
In the metric scorer, for each class of RES, each deep RF structure feature in M samples should have a higher similarity relative to the centroid of the deep RF structure feature in the same class, and it is far from the centroid of deep RF structure features in other classes. As shown in Figure 7, the expected effect of the metric scorer is that it introduces a larger value with respect to the colored part of the similarity score matrix, while the gray part has a smaller value.
When designing the objective function, first, the ( N × M × N ) similarity score is mapped into an ( N × M , N ) similarity score, and the vector s j i , k of each row of the matrix is as follows:
s j i , k = [ ( s 11 , k , s 12 , k , , s 1 M , k ) , ( s 21 , k , s 22 , k , , s 2 M , k ) , , ( s N 1 , k , s N 2 , k , , s N M , k ) ] .
The SoftMax function is applied to the similarity score of each column so that the score is a scalar in the range of 0 1 ; then, the similarity score is regressed to the CE loss function, and the score objective function of a sample of the metric scorer is as follows:
L m s = log exp ( s j i , j ) k = 1 N exp ( s j i , k ) = s j i , j + log k = 1 N exp ( s j i , k ) ,
where the first term is the loss function of positive samples, and the second term is the loss function of negative samples.
That is, for k = 1 , , N , k = j makes the output of the metric scorer equal to 1, indicating that the features and feature centroid in the RF feature pair belong to the same class of RES (positive samples; that is, the same or different operating parameters of the same class of RES). Otherwise, the output is equal to 0, indicating that the features and feature centroid of the RF feature pair do not belong to the same class of RES (negative samples). The objective is to make the network reduce the distance between the RF signals of the hetero-mode operating parameters under the same class of RES in the feature domain during the learning process and increase the distance between the RF signals of different classes of RES in the feature domain, thereby improving the feature domain separability of the deep RF structure features.
Finally, the score objective function of a batch of the metric scorer is as follows:
J m s ( { ϑ s } ) = j = 1 , i = 1 N   ×   M L m s .

4.2.4. Objective Function

In the AEM network, the feature encoder, reconstruction decoder, and metric scorer are regarded as a single neural network, and the three subnetworks are jointly trained in end-to-end training to optimize them relative to one direction. As a result, the inconsistency in the process of separate network training can be eliminated, and the best network can be obtained. Therefore, the overall objective function of the AEM network is as follows:
J A E M ( { ϑ e , ϑ d , ϑ s } ) = λ 1 J r d ( { ϑ e , ϑ d } ) + λ 2 J m s ( { ϑ s } ) ,
where λ 1 and λ 2 represent the proportion weight of the reconstruction objective function and score objective function, respectively. In this paper, the two objective functions are regarded as the same important component, namely, λ 1 , λ 2 = 0.5 . Finally, the appropriate optimization method is selected to optimize the network.

4.3. RES Inversion Method

The flow chart of the RES inversion method is shown in Figure 8, and the overall process of RES modeling and RES conversion can be observed in the figure, which shows the training process of the AE network and AEM network, as well as the testing process of the AEM network. The specific steps are summarized as follows.
  • Building datasets:
    Modeling training dataset: yhe three classes of RES modeled above are used to generate the training dataset under hetero-mode operating parameters, and corresponding labels are added;
    Ideal training dataset: yhe ideal training dataset is generated based on the operating parameters of the modeling training dataset;
    Test datasets: using the three classes of RES models, closed-set test datasets and open-set test datasets are generated.
  • Data preprocessing:
    In-phase/quadrature (I/Q) data are extracted from complex RF signals, and they are concatenated for normalization. The sampling rate is set to f s = 1 GHz , and the pulse width is T = 10 6 s ; thus, the number of complex RF signal sampling points is 1000. Therefore, the number of input nodes is 1000 × 2 = 2000 .
  • Training stage: The specific training process of the network is described as follows:
    (a)
    AE network: The ideal training dataset is used as input to the AE network. Both the setup feature encoder and the reconstruction decoder have two hidden layers. The number of output nodes of the feature encoder is set to 50; thus, the number of nodes of each layer of the feature encoder is 2000 - 600 - 150 - 50 . The number of nodes in each layer of the reconstruction decoder is 50 - 150 - 600 - 2000 . The objective function J A E ( { θ e , θ d } ) is used to train the AE network. The adaptive moment estimation (Adam) optimizer is used for network training. The learning rate is set to 0.0001, and the batch size is set to 60. The network parameters are selected and determined empirically after many experiments; the parameters θ e , θ d of the feature encoder and the reconstruction decoder are obtained, and the trained feature encoder is saved.
    (b)
    AEM network: The modeling training dataset and corresponding labels are used as inputs to the AEM network. Each of the setup feature encoder, the reconstruction decoder, and the metric scorer contain two hidden layers. The number of feature encoder output nodes is set to 50. The number of nodes in each layer of the feature encoder is 2000 - 600 - 150 - 50 , while the number of nodes is 50 - 150 - 600 - 2000 for each layer of the reconstruction decoder, and 100 - 50 - 20 - 1 is observed for each layer of the metric scorer. The objective function J A E M ( { ϑ e , ϑ d , ϑ s } ) is used along with the Adam optimizer to train the AEM network. The learning rate was set to 0.0001 and the batch size was set to 3 × 20 , i.e., a batch contains three classes of RES with 20 samples per class. The network parameters are selected and determined empirically after many experiments. The parameters ϑ e , ϑ d , ϑ s of the feature encoder, the reconstruction decoder, and the metric scorer were obtained, and the trained feature encoder and metric scorer were saved.
    (c)
    Centroid database: In the last epoch, the centroid of each class of deep RF structure features calculated by each batch is saved and averaged to generate a centroid database, which is used as the centroid of deep RF structure features in the test phase.
  • Test stage:
    The network is tested using the test dataset. First, the test data are input into the feature encoder of the AEM network to extract the deep RF structure features. For each class in the centroid database, class centroids are then combined together and input into the metric scorer to calculate the similarity score. The maximum matching score corresponds to the class of RES, and the RES inversion is therefore realized.
Figure 8. Flowchart of the RES inversion method.
Figure 8. Flowchart of the RES inversion method.
Remotesensing 15 04844 g008

5. Experiments

5.1. Simulation and Analysis of the Modelled RF Signal

First, the spurious information introduced by DDS signal source models with different modulation types is verified using experiments. The frequency of the reference clock source is f c = 2 GHz , the number of bits in the phase accumulator is N = 15 , the number of bits for phase truncation is B = 6 , the pulse width is T = 10 6 s , and the sampling rate is f s = 1 GHz . Thus, K CW = 409.6 , K LFM = 409.6 , and K 1 LFM = 0.0819 . According to the parameters set in the model, f CW = f BPSK = f LFM = f QPSK = 20 MHz and B LFM = 10 MHz can be obtained. The frequency variation range of the LFM is 20∼30 MHz.
Figure 9 shows four modulation types of source signals generated by the ideal signal source and four DDS source models. It can be observed from the spectrum that the phase truncation error of the DDS model has a certain influence on the spectrum of the four modulation type signals. In addition, the presence of spurious information can be clearly observed in the spectrum.
The RF structure information of the different classes of RES models is then verified using experiments. Figure 10 illustrates the spectrum diagram of the RF signals produced by different classes of RES models. It is clear that the three classes of RES models exhibit distinct differences in the spectrum purity of RF signals of the four modulation types. Each RF signal carries different RF structure information, with the RF signal generated by the RES1 model having the smallest amount of RF structure information. This is because the amplifier of the RES1 model is relatively ideal and only exhibits one stage of amplification. The measured S parameter is used for the amplifier modeling of RES2 and RES3 models, and the RF structure information is obvious. The structure differences caused by the usage mode, working principle, and material composition of each module are also verified. The generated RF signal also carries various levels of unique RF structure information. In addition, there are obvious differences in the spectrum purity of RF signals with various modulation types under the same RES model, and each RF signal carries different signal information.
Finally, the individual information introduced by different REI models under the same class of RES models is validated experimentally. As shown in Figure 11, as observed from the spectrum of RF signals (DDS_CW) generated by four different REI models under the RES2 model, the RF signals of four different REI models under the same RES2 model carry different individual information, and there is less individual information than RF structure information, reflecting the subtle differences among individuals.

5.2. Hetero-Mode Operating Parameter Dataset

In the modeling training dataset, there are RF signals of three modulation types relative to the DDS_CW, DDS_LFM, and DDS_BPSK models. More precisely, the DDS_CW signal is set to different frequencies, the DDS_LFM signal is set to different initial frequencies and bandwidths, the DDS_BPSK signal is set to different carrier frequencies, and the RF link model is set to different input powers. The operating parameters of the training dataset are shown in Table 2. The number of RF signal samples of each modulation type of each RES model is 2000. Gaussian noise with SNR = 10 dB is then added to each RF signal, and the simulation sample size is 2000 × 3 × 3 = 18,000 with respect to the hetero-mode operating parameters of the modeling training dataset as well as the corresponding simulation of the ideal training dataset. Each sample in the modeling training dataset is then supplemented with a corresponding label in the one-hot encoding mode, and the label is a 1 × 3 vector. To obtain the unbalanced dataset of hetero-mode operating parameters, these parameters will be randomly selected. For instance, there will be 6 × 6 × 3 = 108 combinations of operating parameters under the DDS_LFM model.
The test dataset is divided into a closed-set test dataset and open-set test dataset. The operating parameters of the closed-set test dataset and the training dataset are the same, but these data and their operating parameters have not participated in the model’s training, i.e., the network has not perceived these data. The operating parameters of the open-set test dataset are different from those of the training dataset; that is, the network has never seen these operating parameters nor participated in network training. As shown in Table 3, the test dataset is divided into a closed set and five different open-set test datasets, which are composed of F change; B change; P change; MT change; and MT, F, and P change. The number of RF signal samples for each modulation type under each class of RES models is 40. Gaussian noise with different SNR = ( 10 , 5 , 0 , 5 , 10 , 15 ) dB is also added. For instance, there are 40 × 3 × 3 × 6 = 2160 samples in the closed-set test dataset.

5.3. Feature Extraction Performance and Analysis of the AEM Network

First, to verify whether the AEM network can extract the RF structure information carried by RF signals under hetero-mode operating parameters, the centroids of the deep RF structure features of three classes of RES saved by the AEM network after training are visualized and analyzed in the form of a heat distribution map [59]. Figure 12 shows a visualization of the 50-dimensional deep RF structure feature centroid combination of three classes of RES extracted from the AEM network. It can be observed that the centroids have obvious differences. The feasibility and uniqueness of extracting deep RF structure features from the AEM network are verified. Therefore, these deep RF structure features can be used for RES inversion.
Sample data with SNR = 10 dB were selected in the closed-set test dataset to further verify whether the deep RF structure features extracted by the AEM network in this paper have good separability under hetero-mode operating parameters. The deep RF structure features extracted by the AEM network have reduced dimensionality using t-SNE [28] technology, and the distribution of deep RF structure features is visualized in two-dimensional space. Figure 13a–c show the visualization of the deep RF structure feature distribution of three classes of RES extracted by the AEM network using CW, BPSK, and LFM modulation types, respectively. It is clear that these RF structure features have obvious separability relative to different modulation types, and the separability effect is the best under the LFM modulation type.
Finally, the stability of the deep RF structure features extracted using the AEM network was further verified under hetero-mode operating parameters. The ideal test dataset without RF structure information under the operating parameters corresponding to the closed-set test dataset ( SNR = 10 dB ) was simulated and input into the AEM network to verify its separability and inversion rate. This is a mathematical paradox; if the AEM network cannot clearly distinguish between ideal test data without RF structure information, this indirectly proves that it is extracting deep RF structure features rather than deep signal features. Figure 14a is a visualization of a two-dimensional space that uses t-SNE technology to reduce the dimension of extracted features. It can be observed that using the ideal test dataset, the features extracted by the AEM network are not separable. Figure 14b shows the inversion confusion matrix, where it can be observed that the AEM network no longer has the inversion performance. Therefore, the feasibility and stability of deep RF structure features extracted by the AEM network were indirectly proven.

5.4. Comparison and Analysis of the Generalization Performance of Different Networks

Since RES inversion is a new problem, there are no comparable methods in the published literature. Therefore, in this section, the generalization performance of the AEM network is evaluated via experimental comparisons with some relevant baseline neural networks. Due to space limitations, only some of the experimental results from the test dataset are shown. The baseline network1 is a deep neural network (DNN): it consists of a feature encoder with two hidden layers and a SoftMax classifier (i.e., a feature encoder and a classifier without a reconstruction decoder); the baseline network2 is a CNN: it includes three convolutional and pooling layers, two DNN layers, and a SoftMax classifier. These two networks are used to demonstrate the role of including a reconstruction decoder in this AEM network. Another network used for comparisons is the SAE network: it consists of three stacked AE networks and a SoftMax classifier (i.e., a feature encoder, a reconstruction decoder, and a classifier). The network is trained by initializing and fine-tuning layer by layer. This network is used to demonstrate the function of the metric scorer in the AEM network. Except for the differences in the network’s structure design, the other conditions are the same as the proposed AEM network.
In the closed-set test dataset, samples with SNR = 10 dB are selected to verify the performance of the comparison network and the proposed AEM network. Figure 15a–d visualizes the four networks using t-SNE technology to reduce the extracted deep RF structure features in two-dimensional space. It can be observed that the features extracted by the four networks are distinguishable. However, since the DNN, CNN, and SAE networks use the SoftMax classifier, the spacing between the same class is large, and the boundary of different classes is blurred. The AEM network makes the distance between the same class smaller, the distance between different classes larger, and the boundary clearer using metric learning. Therefore, features extracted by the AEM under hetero-mode operating parameters have higher separability, and samples mixed with three different modulation types of the same class of RES are grouped together. Figure 15e–h illustrates the inversion confusion matrix of the four networks. As observed, all four networks exhibit good inversion performance.
Due to the lack of reconstruction performance in DNN and CNN networks, Figure 16a,b only shows the time-domain reconstructions of SAE and AEM networks. It can be observed that the SAE network has poor reconstruction performance. The AEM network has good reconstruction performance, and it can constrain the features to make them more stable.
In the open-set test dataset4, samples with SNR = 10 dB are selected to verify the performance of the comparison network and the proposed AEM network under unknown operating parameters (MT). Figure 17a–d visualizes the four networks using t-SNE technology to reduce the extracted deep RF structure features in two-dimensional space. It can be observed that in DNN, CNN, and SAE networks, the distance between elements of the same class is large, the boundary between different classes is fuzzy, and even aliasing occurs between classes. The AEM network can still make the spacing within the same class smaller, the spacing between different classes larger, and the boundary clearer under unknown operating parameters. Figure 17e–h depicts the inversion confusion matrix of the four networks, with the total inversion rates being 77%, 72%, 82%, and 89%, respectively. It can be observed in the figure that the inversion rates of the DNN, CNN, and SAE networks are low, while the proposed AEM network can also obtain satisfactory inversion performance under unknown operating parameters. In addition, the inversion rates of RES1 and RES3 can still reach 100%, but the inversion rate of RES2 is lower than that of RES1 and RES3, and part of RES2 is inverted into RES1. This could be because RES1 and RES2 models are more similar—i.e., their structure difference is small—while the RES3 model is significantly different due to its larger RF structure information, which further explains the consistency with previous RES modeling and spectrum analyses.

5.5. Robust Performance and Analysis of the AEM Network

Next, the robustness of the proposed network is verified under different SNRs, and experiments are conducted separately relative to all the above test datasets. The inversion rates under different SNRs are shown in Table 4. It is clear that the proposed network exhibits higher inversion rates under both closed- and open-set test datasets. Particularly under different unknown operating parameters, the inversion performance of the network is still relatively stable as the SNR decreases gradually. When the SNR is −10 dB, the inversion rate can still reach more than 80%, which proves the anti-noise performance of the proposed AEM network. In addition, in the training stage, the network only learns the data in the environment. In the environment of low SNR without training, the deep RF structure features extracted by the AEM are still unaffected by the unknown low-SNR environment, which further demonstrates that AEM has high robustness and generalization performances.

6. Discussion

The results of the above comparison experiment demonstrate that although DNN, CNN, and SAE networks have good inversion performances using the closed-set test dataset, their inversion rate is greatly reduced relative to the open-set test dataset because the features extracted from the known operating parameters of the network do not have good stability and separability, which greatly affects the network via changes in the operating parameters. With unknown operating parameters, the generalization performance of the network is reduced. The proposed AEM network can extract features with the essence of RF structure information by means of the joint optimization of metric learning and reconstruction learning to obtain better inversion performance, and such features have good stability and separability. Furthermore, the network exhibits good generalization performance relative to unknown operating parameters.

7. Conclusions

In this paper, an RES inversion method based on a dual neural network is proposed in complex situations. Using the method of joint unsupervised and supervised training and the powerful nonlinear processing ability of neural networks, the extracted deep RF structure features have good stability and separability, and the RES is then inverted via RF signals in the feature domain. Combined with deep learning technology, RES inversion is studied from a new perspective. The inversion rate and visualization feature map experiments show that the proposed method exhibits high inversion performance in complex situations and high robustness performance under low-SNR conditions. Even under unknown operating parameters, the network also has strong generalization performances. Via RES inversion, the hostile radar can be deeply recognized, which helps in identifying its performance, threat level, and combat capability and efficiently supports cognitive reconnaissance of the electronic countermeasure system. In the future, we will further study RF feature extraction technologies and RES inversion methods using real data in complex situations.

Author Contributions

All authors contributed to the conception and design of the experiments and the interpretation of simulation results. W.Z. conceived the idea, prepared the manuscript, and conducted numerical and experimental validations. L.L. and Y.J. substantially revised the manuscript, and Y.S. and Y.Y. contributed additional revisions of the text. All authors read and approved the final manuscript.

Funding

This work is supported in part by the National Natural Science Foundation of China, under Grant 62071137, and in part by the Key Laboratory of Advanced Marine Communication and Information Technology.

Data Availability Statement

The data presented in this study are all simulated using the proposed model.

Acknowledgments

The authors would like to thank the editors and the reviewers for their comments on this article.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Zhang, M.; Diao, M.; Guo, L. Convolutional neural networks for automatic cognitive radio waveform recognition. IEEE Access 2017, 5, 11074–11082. [Google Scholar] [CrossRef]
  2. Wei, S.; Qu, Q.; Su, H.; Wang, M.; Shi, J.; Hao, X. Intra-pulse modulation radar signal recognition based on CLDN network. IET Radar Sonar Navig. 2020, 14, 803–810. [Google Scholar] [CrossRef]
  3. Talbot, K.I.; Duley, P.R.; Hyatt, M.H. Specific emitter identification and verification. Technol. Rev. 2003, 113, 133. [Google Scholar]
  4. Gok, G.; Alp, Y.K.; Arikan, O. A new method for specific emitter identification with results on real radar measurements. IEEE Trans. Inf. Forensics Secur. 2020, 15, 3335–3346. [Google Scholar] [CrossRef]
  5. Ru, X.H.; Liu, Z.; Huang, Z.T.; Jiang, W.L. Evaluation of unintentional modulation for pulse compression signals based on spectrum asymmetry. IET Radar Sonar Navig. 2017, 11, 656–663. [Google Scholar] [CrossRef]
  6. Huang, G.; Yuan, Y.; Wang, X.; Huang, Z. Specific emitter identification based on nonlinear dynamical characteristics. Can. J. Electr. Comput. Eng. 2016, 39, 34–41. [Google Scholar] [CrossRef]
  7. Dudczyk, J. A method of feature selection in the aspect of specific identification of radar signals. Bull. Pol. Acad. Sci. Tech. Sci. 2017, 65, 113–119. [Google Scholar] [CrossRef]
  8. Ye, W.; Peng, C. Recognition algorithm of emitter signals based on PCA+ CNN. In Proceedings of the 2018 IEEE 3rd Advanced Information Technology, Electronic and Automation Control Conference (IAEAC), Chongqing, China, 12–14 October 2018; pp. 2410–2414. [Google Scholar]
  9. McGinthy, J.M.; Wong, L.J.; Michaels, A.J. Groundwork for neural network-based specific emitter identification authentication for IoT. IEEE Internet Things J. 2019, 6, 6429–6440. [Google Scholar] [CrossRef]
  10. Zhu, Z.; Ji, H.; Li, L. Deep multi-modal subspace interactive mutual network for specific emitter identification. In IEEE Transactions on Aerospace and Electronic Systems; IEEE: New York, NY, USA, 2023; pp. 4289–4300. [Google Scholar]
  11. Tan, K.; Yan, W.; Zhang, L.; Ling, Q.; Xu, C. Semi-Supervised Specific Emitter Identification Based on Bispectrum Feature Extraction CGAN in Multiple Communication Scenarios. IEEE Trans. Aerosp. Electron. Syst. 2023, 59, 292–310. [Google Scholar] [CrossRef]
  12. Amar, A.; Leshem, A.; van der Veen, A.J. A low complexity blind estimator of narrowband polynomial phase signals. IEEE Trans. Signal Process. 2010, 58, 4674–4683. [Google Scholar] [CrossRef]
  13. Li, J.; Ying, Y. Radar signal recognition algorithm based on entropy theory. In Proceedings of the 2014 2nd International Conference on Systems and Informatics (ICSAI 2014), Shanghai, China, 15–17 November 2014; pp. 718–723. [Google Scholar]
  14. Zhang, M.; Liu, L.; Diao, M. LPI radar waveform recognition based on time-frequency distribution. Sensors 2016, 16, 1682. [Google Scholar] [CrossRef] [PubMed]
  15. Bertoncini, C.; Rudd, K.; Nousain, B.; Hinders, M. Wavelet fingerprinting of radio-frequency identification (RFID) tags. IEEE Trans. Ind. Electron. 2011, 59, 4843–4850. [Google Scholar] [CrossRef]
  16. Dragomiretskiy, K.; Zosso, D. Variational mode decomposition. IEEE Trans. Signal Process. 2013, 62, 531–544. [Google Scholar] [CrossRef]
  17. Yuan, Y.; Huang, Z.; Wu, H.; Wang, X. Specific emitter identification based on Hilbert-Huang transform-based time-frequency-energy distribution features. IET Commun. 2014, 8, 2404–2412. [Google Scholar] [CrossRef]
  18. Zhang, J.; Wang, F.; Dobre, O.A.; Zhong, Z. Specific emitter identification via Hilbert–Huang transform in single-hop and relaying scenarios. IEEE Trans. Inf. Forensics Secur. 2016, 11, 1192–1205. [Google Scholar] [CrossRef]
  19. Bihl, T.J.; Bauer, K.W.; Temple, M.A. Feature selection for RF fingerprinting with multiple discriminant analysis and using ZigBee device emissions. IEEE Trans. Inf. Forensics Secur. 2016, 11, 1862–1874. [Google Scholar] [CrossRef]
  20. Satija, U.; Trivedi, N.; Biswal, G.; Ramkumar, B. Specific emitter identification based on variational mode decomposition and spectral features in single hop and relaying scenarios. IEEE Trans. Inf. Forensics Secur. 2018, 14, 581–591. [Google Scholar] [CrossRef]
  21. Ding, X.; He, Q.; Shao, Y.; Huang, W. Transient feature extraction based on time–frequency manifold image synthesis for machinery fault diagnosis. IEEE Trans. Instrum. Meas. 2019, 68, 4242–4252. [Google Scholar] [CrossRef]
  22. Heigold, G.; Moreno, I.; Bengio, S.; Shazeer, N. End-to-end text-dependent speaker verification. In Proceedings of the 2016 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Shanghai, China, 20–25 March 2016; pp. 5115–5119. [Google Scholar]
  23. Xu, G.; Zhang, B.; Yu, H.; Chen, J.; Xing, M.; Hong, W. Sparse synthetic aperture radar imaging from compressed sensing and machine learning: Theories, applications, and trends. IEEE Geosci. Remote Sens. Mag. 2022, 10, 32–69. [Google Scholar] [CrossRef]
  24. Li, J. Recent advances in end-to-end automatic speech recognition. APSIPA Trans. Signal Inf. Process. 2022, 11, e8. [Google Scholar] [CrossRef]
  25. Li, Y. Research and application of deep learning in image recognition. In Proceedings of the 2022 IEEE 2nd International Conference on Power, Electronics and Computer Applications (ICPECA), Shenyang, China, 21–23 January 2022; pp. 994–999. [Google Scholar]
  26. Wu, X.; Shi, Y.; Meng, W.; Ma, X.; Fang, N. Specific emitter identification for satellite communication using probabilistic neural networks. Int. J. Satell. Commun. Netw. 2019, 37, 283–291. [Google Scholar] [CrossRef]
  27. Xiao, Y.; Liu, W.; Gao, L. Radar signal recognition based on transfer learning and feature fusion. Mob. Netw. Appl. 2020, 25, 1563–1571. [Google Scholar] [CrossRef]
  28. Liu, L.; Li, X. Radar signal recognition based on triplet convolutional neural network. EURASIP J. Adv. Signal Process. 2021, 2021, 112. [Google Scholar] [CrossRef]
  29. Ding, L.; Wang, S.; Wang, F.; Zhang, W. Specific emitter identification via convolutional neural networks. IEEE Commun. Lett. 2018, 22, 2591–2594. [Google Scholar] [CrossRef]
  30. Xie, J.; Li, Z.; Zhou, Z.; Liu, S. A novel bearing fault classification method based on XGBoost: The fusion of deep learning-based features and empirical features. IEEE Trans. Instrum. Meas. 2020, 70, 1–9. [Google Scholar] [CrossRef]
  31. Shi, G.; He, Y.; Zhang, C. Feature Extraction and Classification of Cataluminescence Images Based on Sparse Coding Convolutional Neural Networks. IEEE Trans. Instrum. Meas. 2021, 70, 1–11. [Google Scholar] [CrossRef]
  32. Zhu, M.; Feng, Z.; Stanković, L.; Ding, L.; Fan, J.; Zhou, X. A probe-feature for specific emitter identification using axiom-based grad-CAM. Signal Process. 2022, 201, 108685. [Google Scholar] [CrossRef]
  33. LeCun, Y.; Boser, B.; Denker, J.S.; Henderson, D.; Howard, R.E.; Hubbard, W.; Jackel, L.D. Backpropagation applied to handwritten zip code recognition. Neural Comput. 1989, 1, 541–551. [Google Scholar] [CrossRef]
  34. Deng, J.; Zhang, Z.; Marchi, E.; Schuller, B. Sparse autoencoder-based feature transfer learning for speech emotion recognition. In Proceedings of the 2013 Humaine Association Conference on Affective Computing and Intelligent Interaction, Geneva, Switzerland, 2–5 September 2013; pp. 511–516. [Google Scholar]
  35. Ng, A. Sparse autoencoder. CS294A Lect. Notes 2011, 72, 1–19. [Google Scholar]
  36. Su, S.Z.; Liu, Z.H.; Xu, S.P.; Li, S.Z.; Ji, R. Sparse auto-encoder based feature learning for human body detection in depth image. Signal Process. 2015, 112, 43–52. [Google Scholar] [CrossRef]
  37. Tao, C.; Pan, H.; Li, Y.; Zou, Z. Unsupervised spectral–spatial feature learning with stacked sparse autoencoder for hyperspectral imagery classification. IEEE Geosci. Remote Sens. Lett. 2015, 12, 2438–2442. [Google Scholar]
  38. Zhang, L.; Ma, W.; Zhang, D. Stacked sparse autoencoder in PolSAR data classification using local spatial information. IEEE Geosci. Remote Sens. Lett. 2016, 13, 1359–1363. [Google Scholar] [CrossRef]
  39. Lu, X.; Matsuda, S.; Hori, C.; Kashioka, H. Speech restoration based on deep learning autoencoder with layer-wised pretraining. In Proceedings of the Thirteenth Annual Conference of the International Speech Communication Association, Portland, OR, USA, 9–13 September 2012. [Google Scholar]
  40. Lu, X.; Tsao, Y.; Matsuda, S.; Hori, C. Speech enhancement based on deep denoising autoencoder. Interspeech 2013, 2013, 436–440. [Google Scholar]
  41. Rifai, S.; Vincent, P.; Muller, X.; Glorot, X.; Bengio, Y. Contractive auto-encoders: Explicit invariance during feature extraction. In Proceedings of the 28th International Conference on Machine Learning, Bellevue, WA, USA, 2–28 June 2011; pp. 833–840. [Google Scholar]
  42. Liu, L.; Li, X. Unknown radar waveform recognition system via triplet convolution network and support vector machine. Digit. Signal Process. 2022, 123, 103439. [Google Scholar] [CrossRef]
  43. Man, P.; Ding, C.; Ren, W.; Xu, G. A Specific Emitter Identification Algorithm under Zero Sample Condition Based on Metric Learning. Remote Sens. 2021, 13, 4919. [Google Scholar] [CrossRef]
  44. Man, P.; Ding, C.; Ren, W. A Nonlinear Fingerprint-Level Radar Simulation Modeling Method for Specific Emitter Identification. Electronics 2021, 10, 1030. [Google Scholar] [CrossRef]
  45. Hall, J.; Barbeau, M.; Kranakis, E. Detection of transient in radio frequency fingerprinting using signal phase. Wirel. Opt. Commun. 2003, 9, 13–18. [Google Scholar]
  46. Cao, R.; Cao, J.; Mei, J.P.; Yin, C.; Huang, X. Radar emitter identification with bispectrum and hierarchical extreme learning machine. Multimed. Tools Appl. 2019, 78, 28953–28970. [Google Scholar] [CrossRef]
  47. Lundén, J.; Koivunen, V. Automatic radar waveform recognition. IEEE J. Sel. Top. Signal Process. 2007, 1, 124–136. [Google Scholar] [CrossRef]
  48. Xie, Y.; Wang, S.; Zhang, E.; Zhao, Z. Specific Emitter Identification based on nonlinear complexity of signal. In Proceedings of the 2016 IEEE International Conference on Signal Processing, Communications and Computing (ICSPCC), Hong Kong, China, 5–8 August 2016; pp. 1–6. [Google Scholar]
  49. Saleh, A.A. Frequency-independent and frequency-dependent nonlinear models of TWT amplifiers. IEEE Trans. Commun. 1981, 29, 1715–1720. [Google Scholar] [CrossRef]
  50. Mkadem, F.; Boumaiza, S. Physically inspired neural network model for RF power amplifier behavioral modeling and digital predistortion. IEEE Trans. Microw. Theory Tech. 2011, 59, 913–923. [Google Scholar] [CrossRef]
  51. Fehri, B.; Boumaiza, S. Baseband equivalent Volterra series for behavioral modeling and digital predistortion of power amplifiers driven with wideband carrier aggregated signals. IEEE Trans. Microw. Theory Tech. 2014, 62, 2594–2603. [Google Scholar] [CrossRef]
  52. Wu, L.; Zhao, Y.; Feng, M.; Abdalla, F.Y.; Ullah, H. Specific emitter identification using IMF-DNA with a joint feature selection algorithm. Electronics 2019, 8, 934. [Google Scholar] [CrossRef]
  53. Kim, Y.W.; Du Yu, J. Phase noise model of single loop frequency synthesizer. IEEE Trans. Broadcast. 2008, 54, 112–119. [Google Scholar]
  54. Zhao, Y.; Wui, L.; Zhang, J.; Li, Y. Specific emitter identification using geometric features of frequency drift curve. Bull. Pol. Acad. Sciences. Tech. Sci. 2018, 66, 99–108. [Google Scholar]
  55. Yin, X.; Li, X.; Li, Y. Type Recognition of Frequency Synthesizer Based on Convolutional Neural Networks. In Proceedings of the 2020 4th International Conference on Digital Signal Processing, Portland, OR, USA, 9–13 September 2020; pp. 244–249. [Google Scholar]
  56. Abdallah, A.; Celik, A.; Mansour, M.M.; Eltawil, A.M. Deep learning based frequency-selective channel estimation for hybrid mmwave MIMO systems. IEEE Trans. Wirel. Commun. 2021, 21, 3804–3821. [Google Scholar] [CrossRef]
  57. Shashidhar, R.; Patilkulkarni, S.; Puneeth, S. Combining audio and visual speech recognition using LSTM and deep convolutional neural network. Int. J. Inf. Technol. 2022, 14, 3425–3436. [Google Scholar] [CrossRef]
  58. Wan, L.; Wang, Q.; Papir, A.; Moreno, I.L. Generalized end-to-end loss for speaker verification. In Proceedings of the 2018 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Calgary, AB, Canada, 15–20 April 2018; pp. 4879–4883. [Google Scholar]
  59. Lukacs, M.W.; Zeqolari, A.J.; Collins, P.J.; Temple, M.A. “RF-DNA” Fingerprinting for Antenna Classification. IEEE Antennas Wirel. Propag. Lett. 2015, 14, 1455–1458. [Google Scholar] [CrossRef]
Figure 1. Framework of the RES inversion method.
Figure 1. Framework of the RES inversion method.
Remotesensing 15 04844 g001
Figure 2. Influence of changes in the operating parameters of RES inversion. (a) Influence of hetero-mode operating parameters. (b) Influence of unknown operating parameter.
Figure 2. Influence of changes in the operating parameters of RES inversion. (a) Influence of hetero-mode operating parameters. (b) Influence of unknown operating parameter.
Remotesensing 15 04844 g002
Figure 3. Different RES models.
Figure 3. Different RES models.
Remotesensing 15 04844 g003
Figure 4. RF signal under hetero-mode operating parameters.
Figure 4. RF signal under hetero-mode operating parameters.
Remotesensing 15 04844 g004
Figure 5. AE network.
Figure 5. AE network.
Remotesensing 15 04844 g005
Figure 6. AEM network.
Figure 6. AEM network.
Remotesensing 15 04844 g006
Figure 7. Metric scorer.
Figure 7. Metric scorer.
Remotesensing 15 04844 g007
Figure 9. Spectrum diagram of the source signals of four modulation types. (a) CW; (b) BPSK; (c) LFM; (d) QPSK; (e) DDS_CW; (f) DDS_BPSK; (g) DDS_LFM; (h) DDS_QPSK.
Figure 9. Spectrum diagram of the source signals of four modulation types. (a) CW; (b) BPSK; (c) LFM; (d) QPSK; (e) DDS_CW; (f) DDS_BPSK; (g) DDS_LFM; (h) DDS_QPSK.
Remotesensing 15 04844 g009
Figure 10. Spectrum diagram of RF signals with different classes of RES models: (a) RES1_DDS_CW; (b) RES1_DDS_BPSK; (c) RES1_DDS_LFM; (d) RES1_DDS_QPSK; (e) RES2_DDS_CW; (f) RES2_DDS_BPSK; (g) RES2_DDS_LFM; (h) RES2_DDS_QPSK; (i) RES3_DDS_CW; (j) RES3_DDS_BPSK; (k) RES3_DDS_LFM; (l) RES3_DDS_QPSK.
Figure 10. Spectrum diagram of RF signals with different classes of RES models: (a) RES1_DDS_CW; (b) RES1_DDS_BPSK; (c) RES1_DDS_LFM; (d) RES1_DDS_QPSK; (e) RES2_DDS_CW; (f) RES2_DDS_BPSK; (g) RES2_DDS_LFM; (h) RES2_DDS_QPSK; (i) RES3_DDS_CW; (j) RES3_DDS_BPSK; (k) RES3_DDS_LFM; (l) RES3_DDS_QPSK.
Remotesensing 15 04844 g010
Figure 11. Spectrum of DDS_CW signals of four different individuals in the RES2 model: (a) RES2_DDS_CW_REI1; (b) RES2_DDS_CW_REI2; (c) RES2_DDS_CW_REI3; (d) RES2_DDS_CW_REI4.
Figure 11. Spectrum of DDS_CW signals of four different individuals in the RES2 model: (a) RES2_DDS_CW_REI1; (b) RES2_DDS_CW_REI2; (c) RES2_DDS_CW_REI3; (d) RES2_DDS_CW_REI4.
Remotesensing 15 04844 g011
Figure 12. Visualization of 50-dimensional deep RF structure feature centroids of three classes of RES using the AEM network.
Figure 12. Visualization of 50-dimensional deep RF structure feature centroids of three classes of RES using the AEM network.
Remotesensing 15 04844 g012
Figure 13. Visualization of deep RF structure features relative to different modulation types: (a) CW; (b) BPSK; (c) LFM.
Figure 13. Visualization of deep RF structure features relative to different modulation types: (a) CW; (b) BPSK; (c) LFM.
Remotesensing 15 04844 g013
Figure 14. Proof by contradiction: (a) visualization; (b) inversion confusion matrix.
Figure 14. Proof by contradiction: (a) visualization; (b) inversion confusion matrix.
Remotesensing 15 04844 g014
Figure 15. Experimental diagram of different networks in closed-set test dataset: (a) visualization of DNN network; (b) visualization of CNN network; (c) visualization of SAE network; (d) visualization of AEM network; (e) inversion of DNN network; (f) inversion of CNN network; (g) inversion of SAE network; (h) inversion of AEM network.
Figure 15. Experimental diagram of different networks in closed-set test dataset: (a) visualization of DNN network; (b) visualization of CNN network; (c) visualization of SAE network; (d) visualization of AEM network; (e) inversion of DNN network; (f) inversion of CNN network; (g) inversion of SAE network; (h) inversion of AEM network.
Remotesensing 15 04844 g015
Figure 16. Reconstruction experimental diagram of different networks in closed-set test dataset: (a) reconstruction of SAE network; (b) reconstruction of AEM network.
Figure 16. Reconstruction experimental diagram of different networks in closed-set test dataset: (a) reconstruction of SAE network; (b) reconstruction of AEM network.
Remotesensing 15 04844 g016
Figure 17. Experimental diagram of different networks in open-set test dataset4. (a) visualization of DNN network; (b) visualization of CNN network; (c) visualization of SAE network; (d) visualization of AEM network; (e) inversion of DNN network; (f) inversion of CNN network; (g) inversion of SAE network; (h) inversion of AEM network.
Figure 17. Experimental diagram of different networks in open-set test dataset4. (a) visualization of DNN network; (b) visualization of CNN network; (c) visualization of SAE network; (d) visualization of AEM network; (e) inversion of DNN network; (f) inversion of CNN network; (g) inversion of SAE network; (h) inversion of AEM network.
Remotesensing 15 04844 g017
Table 1. Design parameter values of different modules.
Table 1. Design parameter values of different modules.
ModuleDesign ParameterPA1PA2PA3
Mixer (Phase Noise)Frequency Offset (Hz)FO1 = [100 5000 30,000 50,000] FO2 = [100 1000 10,000 100,000]
Phase Noise Leve (dBc/Hz)PNL1 = [−70 −93 −107 −132] PNL2 = [−70 −120 −140 −150]
S Parameter[0, 0, 1, 0][0, 0, 1, 0][0, 0, 1, 0]
One-Stage AmplifierS Parameter[0, 0, sqrt (10,000), 0][0, 0, sqrt (1000) 0][0, 0, sqrt (100), 0]
IP3 (dBm)47∼5036∼3942∼45
Two-Stage AmplifierS ParameterHMC637ALP5EHMC637ALP5E
IP3 (dBm)4444
Three-Stage AmplifierS ParameterHMC8500PM5E
IP3 (dBm)42
Table 2. Operating parameters of the training dataset.
Table 2. Operating parameters of the training dataset.
ModelOperating ParameterRanges
DDS_CW ModelFrequency (MHz)25∼30
DDS_LFM ModelInitial Frequency (MHz)25∼30
Bandwidt (MHz)10∼15
DDS_BPSK ModelCarrier Frequency (MHz)25∼30
RF Link ModelInput Power (mW)0.1, 0.5, 1
Table 3. Operating parameters of the test dataset.
Table 3. Operating parameters of the test dataset.
Test DatasetF (MHz)B (MHz)P (mW)MT
Closed Set25∼3010∼150.1, 0.5, 1CW, LFM, BPSK
Open Set1 (F)31∼320.1, 0.5, 1CW
Open Set2 (B)25∼3016∼170.1, 0.5, 1LFM
Open Set3 (P)25∼300.3, 0.7BPSK
Open Set4 (MT)25∼300.1, 0.5, 1QPSK
Open Set5 (MT, F, and P)31∼320.3, 0.7QPSK
Table 4. Inversion rate of the AEM network under different SNRs.
Table 4. Inversion rate of the AEM network under different SNRs.
SNR (dB)1050−5−10−15
Closed Set100%100%100%97%86%68%
Open Set189%89%89%89%89%85%
Open Set289%89%90%89%88%68%
Open Set3100%100%100%99%85%54%
Open Set488%88%92%92%84%62%
Open Set586%86%88%84%70%52%
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Liu, L.; Zhang, W.; Jiang, Y.; Yang, Y.; Song, Y. Radar Emitter Structure Inversion Method Based on Metric and Deep Learning. Remote Sens. 2023, 15, 4844. https://doi.org/10.3390/rs15194844

AMA Style

Liu L, Zhang W, Jiang Y, Yang Y, Song Y. Radar Emitter Structure Inversion Method Based on Metric and Deep Learning. Remote Sensing. 2023; 15(19):4844. https://doi.org/10.3390/rs15194844

Chicago/Turabian Style

Liu, Lutao, Wei Zhang, Yilin Jiang, Yaozu Yang, and Yu Song. 2023. "Radar Emitter Structure Inversion Method Based on Metric and Deep Learning" Remote Sensing 15, no. 19: 4844. https://doi.org/10.3390/rs15194844

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop