Next Article in Journal
Comparison of Mean Values and Entropy in Accelerometry Time Series from Two Microtechnology Sensors Recorded at 100 vs. 1000 Hz During Cumulative Tackles in Young Elite Rugby League Players
Previous Article in Journal
Reinforcement-Learning-Based Fixed-Time Prescribed Performance Consensus Control for Stochastic Nonlinear MASs with Sensor Faults
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Robust Automatic Modulation Classification via a Lightweight Temporal Hybrid Neural Network

1
Aerospace Information Research Institute, Chinese Academy of Sciences, Beijing 100094, China
2
School of Electronic, Electrical and Communication Engineering, University of Chinese Academy of Sciences, Beijing 100049, China
*
Author to whom correspondence should be addressed.
Sensors 2024, 24(24), 7908; https://doi.org/10.3390/s24247908
Submission received: 26 September 2024 / Revised: 2 November 2024 / Accepted: 9 December 2024 / Published: 11 December 2024
(This article belongs to the Section Communications)

Abstract

:
In the rapidly developing field of wireless communications, the precise classification of modulated signals is essential for optimizing spectrum utilization and improving communication quality. However, existing networks face challenges in robustness against signals containing phase shift keying and computational efficiency. This paper introduces TCN-GRU, a lightweight model that combines the advantages of multiscale feature extraction of the temporal convolutional network (TCN) and global sequence modeling of gated recurrent unit (GRU). Compared to the state-of-the-art MCLDNN, TCN-GRU reduces parameters by 37.6%, achieving an accuracy of 0.6156 and 0.6466 on the RadioML2016.10a and RadioML2016.10b, respectively (versus MCLDNN’s 0.6101 and 0.6462). Furthermore, TCN-GRU demonstrates superior ability in distinguishing challenging modulations such as QAM16 and QAM64, and it improves classification accuracy by about 10.5% compared to MCLDNN. These results suggest that TCN-GRU is a robust and efficient solution for enhancing AMC in complex and noisy environments.

1. Introduction

Radio signals are a fundamental component of human communication. As communication technologies advance, channel environments have become increasingly complex, making accurate identification of various modulation methods within dense channels challenging due to noise and multipath fading. Automatic modulation classification (AMC) technology emerges as a critical solution in this context, serving as a vital intermediary between energy detection and the full demodulation process in decoding systems. AMC enhances spectrum efficiency, and it finds extensive application across diverse sectors. In commercial settings, AMC performs tasks like interference identification and frequency management, streamlining communication networks for better efficiency and reliability. In military contexts, AMC is indispensable for interception, jamming, and surveillance operations, offering strategic communication and signal intelligence advantages. The widespread applicability of AMC underscores its importance in advancing communication technology and supporting critical operational needs in various domains.
Originally, modulated signals were identified manually. Subsequently, AMC emerged, initially divided into likelihood-based (LB) and feature-based (FB) methods [1]. LB methods, grounded in mathematical rigor, offer optimal classification under Bayesian criteria but have high computational complexity and limited noise resistance [2]. FB methods use pattern recognition to map signal characteristics into feature spaces [3]. It utilizes prior knowledge for classifying signal modulation techniques, primarily through feature extraction including instantaneous aspects such as amplitude, frequency, and phase [4], higher-order statistics [5,6], transform domain characteristics involving Fourier and wavelet transform [7,8], as well as constellation [9]. Common classifiers of FB methods include support vector machines [10], decision trees [11], hidden Markov models [12], and k-nearest neighbor [13]. While effective, FB methods rely on manually extracted features, limiting their adaptability to complex, dynamic signals [14].
In complex electromagnetic environments, digital signals are vulnerable to noise interference. Elevated noise levels can obscure the information carried by the original signal, impacting the accuracy of modulation identification. The signal-to-noise ratio (SNR) is a critical metric to assess the extent of noise interference. It is defined by Equation (1):
SNR = 10 log P s P n
where P s denotes the signal power and P n signifies the power of the noise. A higher SNR indicates lesser noise content within the signal, while a lower SNR suggests increased noise levels. SNR falls below 0 dB indicating high noise interference. Accurate identification of modulation types is crucial for guiding demodulation in such cases.
Deep learning has demonstrated exceptional performance in fields such as image processing, speech recognition, and text analysis in recent years. This success has naturally extended to the field of AMC. Deep learning is data-driven. It adapts iteratively to input characteristics from large datasets, resulting in highly accurate, stable predictive models. Compared to LB and FB methods, it offers superior adaptability and excels at extracting complex features, particularly for the end-to-end classification of modulated signals at low SNR [15]. Currently, various deep network architectures including convolutional neural networks (CNNs) [16], recurrent neural networks (RNNs) [17], and a range of hybrid models are being used in AMC, each contributing uniquely to the advancement of AMC methodologies.
Networks based on CNN have shown impressive performance in image processing. O’Shea et al. employed a CNN model with two convolutional layers and two fully connected layers to classify 11 modulation formats of in-phase and quadrature (I/Q) sampled signals using the RadioML2016.10a dataset [18]. However, the relatively simple design of the network limited its ability to extract temporal features, resulting in an accuracy of only 57.0%. In contrast, Liu et al. [19] used the residual network (ResNet) [20] to analyze RadioML2016.10a and RadioML2016.10b. The accuracy on RadioML2016.10a was comparable to that of CNN, and the overall accuracy on RadioML2016.10b was able to reach 61.2% (compared to 59.4% for CNN), although it still fell short of the performance achieved by the hybrid network.
Fundamentally, AMC can be viewed as a time series modeling process. Therefore, many researchers have begun to study the effectiveness of RNN. Long short-term memory (LSTM) [21], known for its proficiency in modeling both forward and backward temporal dependencies, has been a focus in this area. Rajendran et al. [22] implemented a two-layer LSTM network to learn the amplitude and phase (A/P) characteristics. This method has an accuracy of nearly 87% for signals above 0 dB, but for signals below 0 dB, its classification ability is lower than that of the hybrid model [19,23] (about 2–4%). In addition, LSTM needs to be trained on a larger scale of data to achieve good results [15]. Gated recurrent unit (GRU) [24] has become a noteworthy alternative to LSTM. It provides similar modeling capabilities but with fewer parameters and faster training and prediction speed. Hong et al. [25] demonstrated the effectiveness of GRU in identifying I/Q signals in the RadioML2016.10a dataset. With RadioML2016.10a, GRU improves the classification accuracy of signals above 0 dB by about 4% compared to CNN and is similar to LSTM. It has an advantage in speed, with a training speed increase of about 14% compared to LSTM, which enhances its potential as a valuable tool in AMC.
The increasing complexity of wireless communication environments has posed significant challenges to accurately identifying signal modulation modes, particularly under low SNR conditions. Traditional methods such as LB and FB approaches, while effective in certain scenarios, suffer from drawbacks including high computational complexity, limited adaptability to noise interference, and inadequate feature extraction capabilities. Deep learning models, including CNNs and RNNs, have made strides in AMC, yet they often struggle to capture the full range of temporal dependencies or the intricate local–global feature relationships inherent in modulated signals.
Hybrid models, by integrating the strengths of CNNs and RNNs, offer superior results in modulation signal identification. This synergistic approach unlocks new potential in the realm of AMC. West et al. [26] integrating three one-dimensional convolutional layers with a single-layer LSTM. This model is faster than GRU and improved the classification accuracy by 1–2% over LSTM below 0 dB. Liu et al. [19] refined West’s model structure by incorporating a two-dimensional convolutional layer, denoted as convolutional long short-term deep neural network (CLDNN), enhancing its classification accuracy. Notably, the CLDNN model demonstrates superior performance to traditional CNN-type and RNN-type models such as ResNet, DenseNet [27], GRU, and LSTM, and its classification accuracy is more than 4–6% higher than previous models, being within −8–0 dB. This superiority underscores the effectiveness of combining CNN and RNN architectures in distinguishing features of different modulated signals. Advancing this approach, Xu et al. [23] proposed the multichannel CLDNN (MCLDNN). It utilizes three input channels to extract complementary features from the I/Q signal, with accuracies of approximately 61.0% and 64.6% for RadioML2016.10a and RadioML2016.10b, respectively (60.5% and 62.7% for CLDNN [19]).
Despite the significant advancements made by the aforementioned hybrid models in AMC compared to LB, FB, CNNs, and RNNs, they also have some limitations. Firstly, there is still room for improvement in the classification accuracy of hybrid networks, and some hybrid models [19,23,28] continue to face limitations when identifying signals with complex modulations, such as QAM and PSK. Both QAM and PSK use phase as the information carrier. As the signal’s SNR decreases, the phase’s change points are more easily masked by noise. The existing hybrid models generally use ordinary convolution to initially extract signal features without considering the multiscale features of QAM and PSK, which makes the classification accuracy relatively low. Additionally, these hybrid models often have complex network structures and numerous parameters, which require substantial computational resources and time during training and inference. To address these limitations, there is a need for a more robust network architecture that can effectively extract both the local fragmentary features and the overarching structural features of modulated signals, while also being computationally efficient.
In this paper, we introduce a novel hybrid neural network model, termed TCN-GRU. TCN-GRU was developed to overcome the challenges mentioned above by integrating the strengths of temporal convolutional networks (TCN) for multiscale feature synthesis and GRU for temporal sequence modeling, providing a more comprehensive solution to the limitations of previous methods. Then, TCN-GRU employs global average pooling (GAP) to fuse and map these features across the time scale. Finally, the classification result is obtained through a fully connected layer. The main contributions of this paper are as follows:
  • The proposed TCN-GRU enables multiscale feature extraction and fusion, which allows our model to improve the classification accuracy of QAM and PSK, compared to other models;
  • The proposed TCN-GRU is valid on RadioML2016.10a and RadioML2016.10b. Compared with some high-performing networks, our model demonstrates superior performance;
  • The proposed TCN-GRU has fewer parameters than the existing high-precision model, and is structurally more efficient, making it highly practical.

2. Materials and Methods

In this section, we introduce the publicly accessible modulation signal datasets used to validate our model’s performance and their preprocessing methods. We also explain the architecture of the proposed TCN-GRU model and outline the evaluation metrics to assess its effectiveness and classification ability.

2.1. Modulation Signal Datasets

This study use the RadioML2016.10a and RadioML2016.10b datasets from DeepSig [29] to train and evaluate various deep network models. These datasets include numerous signal samples across multiple modulation methods and simulate real-world electromagnetic conditions by incorporating additive white Gaussian noise, multipath fading, center frequency offset, and sample rate offset. The details of these datasets are provided in Table 1.
RadioML2016.10a and RadioML2016.10b consist of I/Q signals with 128 sample points. However, they can also be represented in A/P format. Previous research [22] has shown that data in A/P format can provide more intuitive physical features, thereby improving the model classification accuracy. The conversion from I/Q to A/P representation is described by Equations (2) and (3) as follows:
X A = X I 2 + X Q 2
X P = 2 π arctan X I X Q
where X I ,   X Q represent the I and Q channel signal sequences, respectively, and X A ,   X P denote the amplitude sequence and the normalized phase sequence, respectively.
In our experiments, RadioML2016.10a and RadioML2016.10b were randomly divided in an 8:1:1 ratio using a seed of 114,514. Of these, 80% of the signals (176,000 for RadioML2016.10a and 960,000 for RadioML2016.10b) is used for training, 10% (22,000 and 120,000) is used for validation—where validation loss determined model saving—and the remaining 10% is used for testing. The signal quantities for each modulation type and SNR level are uniformly distributed across all sections.
Using these two datasets to validate the model helps evaluate its convergence and robustness. RadioML2016.10a provides a baseline test covering a wide range of modulation types, and a smaller number of samples can test the convergence of the model. RadioML2016.10b has a larger number of data, which can evaluate the robustness of the model. Using two datasets to evaluate the model can more comprehensively reflect the classification ability of the model.

2.2. Deep Learining Networks

2.2.1. Temporal Convolutional Network

TCN is a stable and efficient model for time series prediction [30]. Similar to RNNs, TCN is stackable, handles arbitrary input lengths, and maintains equivalent lengths for input and output. To manage varying input lengths, TCN uses dilated convolutions [31,32], designed to capture temporal relationships by ensuring the output at time t depends solely on inputs up to the past-time point t 1 . For equal input–output lengths, TCNs use a 1D fully convolutional network architecture [33] with zero padding. TCN integrates these functionalities with residual blocks (TCNRB), facilitating deep stacking of layers while preserving the receptive field size of dilated convolutions. The architecture of TCNRB is depicted in Figure 1.
TCN has a flexible receptive field size, enabling it to handle time series of various lengths through dilated convolutions and residual structures. Dilated convolutions expand the receptive field exponentially via the dilation factor, significantly increasing the receptive field without adding network depth. This enables TCN to capture features at multiple scales, reducing overall depth while effectively processing time series of varying lengths. For one-dimensional time series x R n and filter f : { 0 , 1 , , k 1 } R , the operation process of dilated convolution is shown in Equation (4):
F ( s ) = i = 0 k 1 f ( i ) · x s d · i
where d represents the dilation factor, k denotes the kernel size, and s d · i stands for the timestamp of the past. The size of the receptive field R is calculated with Equation (5):
R = 1 + 2 · ( k 1 ) · N · i = 0 N d i
where k represents the size of the dilated convolution kernel in TCNRB, N represents the number of TCNRB layers, and d i represents the dilation coefficient of the i-th layer residual block, with d typically sey to 2.

2.2.2. Gate Recurrent Unit

GRU [24] simplifies the LSTM architecture. It combines the forget gate and input gate into a single update gate, streamlining the network unit. The GRU cell is illustrated in Figure 2, with forward propagation governed by Equations (6)–(9). This design makes GRU an efficient alternative to LSTM for sequence modeling tasks.
At each time step t, the input signal x t and the previously hidden state output h t 1 are fed into the GRU’s reset gate r t and update gate z t . These gates utilize the Sigmoid activation function, which outputs values between [ 0 , 1 ] . A value of 1 indicates full retention of new information, while 0 implies complete disregard. The reset gate determines how much past information is forgotten, while the update gate decides how much is retained. Next, the Hadamard product (denoted as ⊙) of r t and h t 1 captures relevant past information to create the candidate hidden state h ^ t , processed by the tanh function to mitigate gradient issues. Finally, the update gate controls the mix of memory from h t 1 and the new information from h t , as well as the amount of new information from h ^ t in the current output, allowing the GRU to balance past and new information effectively. These are calculated as follows:
r t = σ W r · [ h t 1 , x t ] + b r
z t = σ W z · [ h t 1 , x t ] + b z
h ^ t = tanh W h · [ r t h t 1 , x t ] + b z
h t = ( 1 z t ) h t 1 + z t h ^ t

2.2.3. TCN-GRU

Modulation signals exhibit diverse characteristics, and some related modulation modes may have highly similar time-domain features. A single network often faces limited classification accuracy for complex or similar modulated signals. Hybrid networks combine different network types in series and leverage the strengths of multiple networks. This synergy allows for more effective feature extraction and capturing the diverse characteristics of modulated signals in various forms. As a result, hybrid networks can overcome some of the feature recognition limitations inherent in single-network models. Our proposed TCN-GRU model combines two temporal feature extraction networks, i.e., TCN and GRU, resulting in a simpler structure with fewer parameters than many hybrid networks. In addition, TCN-GRU offers stronger feature extraction capabilities than CNN-type, RNN-type, and some hybrid models, making it effective for AMC.
Our TCN-GRU model consists of four TCNRBs and two GRU layers. Each TCNRB uses a convolution kernel size of 1 × 7 and contains a hidden size of 64. The GRU layers also have a hidden size of 64.
The process of the TCN-GRU in classifying modulated signals is shown in Figure 3. The 2 × 128 I/Q signal from RadioML2016 is first transformed into a 2 × 128 complex signal in A/P form. These A/P signals then are fed into the TCN-GRU network. The TCNRBs progressively increase the receptive field, capturing local features at shallow layers and broader structural features at deeper layers. These feature maps are then processed by the GRU layers, which extract deeper, overarching temporal features of the signal. We employ GAP to fuse multiscale and temporal sequence features. It can also reduce feature map dimensions and model parameters, summarize time-dimension information, and speed up training. Finally, a fully connected layer is used to classify the extracted features, producing predicted probabilities for various modulated signals. This structured approach enables TCN-GRU to efficiently and accurately identify modulation types in complex signals. For optimization, we employ cross-entropy loss to quantify the deviation between predictions and the ground truth, and the Adam optimization method [34] is used for updating network parameters.

2.3. Performance Metrics

To evaluate the models’ capability in identifying various modulated signals, four evaluation metrics are employed: Accuracy, Precision, Recall, and F1-Score. These metrics provide a comprehensive understanding of the models’ performance, reflecting different aspects of classification accuracy and reliability. The relationships between prediction and ground truth are defined as follows:
  • True Positive (TP): Positive samples correctly predicted as positive;
  • True Negative (TN): Negative samples correctly predicted as negative;
  • False Positive (FP): Negative samples incorrectly predicted as positive;
  • False Negative (FN): Positive samples incorrectly predicted as negative.
For an N-class classification problem, the Accuracy metric represents the ratio of correctly predicted samples to the total number of samples in the test set, measuring the probability of accurate predictions. For the i-th class, the calculation method of Accuracy is detailed in Equation (10):
Accuracy = i = 1 N ( TP i + TN i ) i = 1 N ( TP i + TN i + FP i + FN i )
Precision indicates the proportion of actual positive samples among all samples predicted as positive, making it vital when the consequence of FP is significant. Conversely, Recall reflects the proportion of positive samples correctly predicted. It becomes crucial when the cost of FN is high. The calculation methods for Precision and Recall are detailed in Equations (11) and (12):
Precision = 1 N i = 1 N TP i TP i + FP i
Recall = 1 N i = 1 N TP i TP i + FN i
Precision and Recall measure the model’s ability to differentiate between negative and positive samples. Typically, a model with high Precision tends to have a lower Recall and vice versa, reflecting an inherent trade-off between these metrics. F-Score is the weighted harmonic average of Precision and Recall, providing a balanced measure of the model’s robustness by considering its performance on both positive and negative samples. Equation (13) presents the formula for calculating the F-Score:
F - Score = ( 1 + β 2 ) · Precision · Recall β 2 ( Precision + Recall )
When β = 1 , the F-Score becomes the F1-Score, indicating that Precision and Recall are given equal weight. In the context of an N-class classification problem, the calculation of the F1-Score is presented in Equation (14):
F 1 - Score = 1 N · i = 1 N 2 · Precision i · Recall i Precision i + Recall i

3. Experiments and Results

In the following experiments, we first assess the proposed models’ performance under various parameter configurations. After identifying the optimal parameters, we conduct comparative and ablation studies using RadioML2016.10a and RadioML2016.10b. The models are evaluated across three dimensions: overall performance indices, overall accuracy at each SNR level, and accuracy for each type of modulated signal. All experiments are conducted on a graphics workstation with an Intel Xeon W-2245 CPU and an NVIDIA Quadro RTX5000 GPU, providing robust computational support for training and analyses.

3.1. Determination of Network Structure

TCN and GRU are stackable neural networks. Their performance is influenced by factors such as the number of layers, hidden size, and the convolutional kernel size. To find the optimal configuration with the highest accuracy, we designed a series of experiments using RadioML2016.10a to identify the most effective architecture.
  • Layer Optimization: We evaluated the valid F1-Score of networks with varying numbers of TCN and GRU layers, keeping hidden size and convolution kernel size constant. This determined the optimal layer count for both TCN and GRU.
  • Hidden size Optimization: With the TCN and GRU layers fixed, we assessed the valid accuracy for different hidden sizes in TCN and GRU, maintaining a constant kernel size to identify the ideal count for each network.
  • Convolution Kernel Size Optimization: Based on the optimal number of layers and hidden size, we evaluated the valid accuracy for networks with varying TCN convolution kernel sizes finalizing the network structure.
The comparative results are shown in Figure 4. The results indicate that a network with a four-layer TCN (kernel size of 7, 64 hidden size) combined with a two-layer GRU (64 hidden size) enhances the model’s capability to extract signal A/P features, resulting in higher accuracy.

3.2. Comparison of Performance Between Models

We conduct a performance comparison to evaluate TCN-GRU against some established methods in AMC, including CNN-type models like CNN2 [18] and ResNet [19], RNN-type models such as LSTM2 [22] and GRU2 [25], and hybrid models like CLDNN [19], MCLDNN [23], and CGDNet [35]. Notably, MCLDNN is the current state-of-the-art (SOTA) method in this field.
We also conduct ablation experiments to validate the efficacy of TCN-GRU. The control groups include GRU2-64: wop (two-layer GRU with 64 hidden size), GRU2-64 (two-layer GRU with 64 hidden size + GAP), TCN4: wop (four-layer TCN with 64 hidden size), TCN4 (four-layer TCN with 64 hidden size + GAP), and TCN-GRU: wop (TCN-GRU removed GAP). These comparisons demonstrate that the integration of TCN, GRU, and GAP achieves superior performance.
The structures of all networks are detailed in Table 2, where c denotes the number of channels, k the kernel size, N the prediction categories (11 for RadioML2016.10a and 10 for RadioML2016.10b), and h the hidden size. Conv refers to convolutional layers, FC to fully connected layers, Concat to concatenation, and MaxPool to max pooling.
The comparison and analysis of different network models encompass three main dimensions. They are:
  • Performance metrics: Each model’s performance is compared using the metrics outlined in Section 2.3, aiming to reflect the classification capabilities of each model.
  • Overall accuracy (OA) in different SNRs: We assess and compare the OA of all models across a spectrum of SNRs, measuring the model robustness in various noise powers.
  • Confusion matrices: We use confusion matrices to evaluate the model’s ability to recognize different modulated signals.
To mitigate the impact of high-power noise, we specifically compare each model’s OA within an SNR range of −8 dB to 18 dB. This targeted comparison helps us understand performance under common noise conditions.

3.2.1. Comparison Between Model Performance Metrics

We use seven metrics to compare model performance: the number of parameters, validation cross-entropy loss, Accuracy, F1-Score, Recall, Precision, and prediction time on the test set. Prediction time is the average time to predict 22k samples (RadioML2016.10a) and 120k samples (RadioML2016.10b) taken by the model over five trials. The quantitative comparison results on both datasets are shown in Table 3 and Table 4.
In comparison experiments, we observed that CNN-type models (CNN2 and ResNet) have difficulty completing the AMC task, showing relatively low classification accuracy on both datasets. While fast in prediction, their high parameter count limits storage and loading speed. RNN-type models (GRU2 and LSTM2) achieved high prediction speeds. These models underperformed with limited samples but improved with larger datasets. Hybrid models like CGDNet, CLDNN, MCLDNN, and the proposed TCN-GRU exhibited distinct advantages. CGDNet excelled in speed. CLDNN, despite having numerous parameters, slightly outperformed CNN-type and RNN-type models on both datasets. MCLDNN displayed high accuracy on both datasets, outperforming other networks. TCN-GRU surpasses all others in multiple evaluation metrics, with a 37.6% reduced parameters compared to MCLDNN, making it less prone to overfitting in small-sample scenarios. Besides, TCN-GRU demonstrates a more significant accuracy improvement on RadioML2016.10a, which also highlights its strong generalization capability across small sample sizes. The ablation experiment results show that the TCN-GRU outperforms all control models on both datasets, underscoring the effectiveness of our proposed architecture and the synergistic advantages of combining TCN and GRU components.
TCN-GRU outperforms the current SOTA method in storage and loading efficiency. It is noticable that TCN-GRU’s computing speed is slightly lower, positioning it as a more storage-efficient option. This reflects a trade-off between prediction speed and accuracy, with our hybrid model prioritizing accuracy at a slight cost to inference speed. Despite this, the model’s reduced parameters lower overfitting risks and enhance generalization, making it highly suitable for practical applications where accuracy is paramount.

3.2.2. Comparison of OA with Different SNR Levels

RadioML2016.10a and RadioML2016.10b include modulated signals with SNR levels from −20 dB to 18 dB. We categorized them into three intervals for analysis: extremely low SNR (−20 dB to −10 dB), low SNR (−8 dB to 0 dB), and high SNR (2 dB to 18 dB). In the extremely low SNR interval, noise is 10 to 100 times stronger than the signal, making feature extraction hard. The low SNR range, where noise power is about 1 to 6.3 times the signal power, also presents challenges in feature extraction. The high SNR interval, where signal power exceeds noise, supports higher classification accuracy.
Figure 5 provides a detailed OA overview of TCN-GRU other models across various SNR levels. The figure includes separate subgraphs to highlight the comparison between TCN-GRU and the high-performing MCLDNN model in the low SNR interval and the accuracy lines of several high accuracy networks in the high SNR interval. Specifically, for RadioML2016.10a in Figure 5a, our model demonstrates an OA improvement over MCLDNN of approximately 0.244%, 0.753%, and 0.192% in the extremely low, low, and high SNR intervals, respectively. Notably, accuracy is increased at the −4 dB and 14 dB SNR levels, with about 2.194% and 1.764%. For RadioML2016.10b in Figure 5b, the OA improvements of TCN-GRU are about 1.276%, 1.939%, and 0.052% across the three SNR intervals, with gains of approximately 4.367% and 2.861% at -6 dB and −4 dB, respectively, compared to MCLDNN. These results indicate that the TCN-GRU model’s robustness and effectiveness in challenging SNR scenarios outperform current models in AMC.
Figure 6 presents a line chart of the OA comparison between TCN-GRU and control group models at each SNR level from the ablation experiment. TCN-GRU consistently achieves the best recognition performance in both datasets, indicating that the combination of TCN, GRU, and GAP components is synergistic and effective.

3.2.3. Comparison of the Recognition Accuracy of Different Modulated Signals

We use confusion matrices to illustrate the accuracy of models for different signal modulations. Figure 7 shows the results for all models on RadioML2016.10a, while Figure 8 presents the results for RadioML2016.10b. These matrices visually represent each model’s strengths and weaknesses in accurately identifying and classifying modulated signals across categories.
For RadioML2016.10a and RadioML2016.10b, three signal pairs are commonly confused: QPSK/8PSK, QAM16/QAM64, and AM-DSB/WBFM. QPSK and 8PSK, both phase modulation techniques, share sine and cosine oscillation characteristics, differing mainly in the number of phases. BPSK, with simpler phase changes, is less frequently confused with these modulations. QAM16 and QAM64, as quadrature amplitude modulations, use different amplitude and phase combinations to represent symbols, making them prone to misidentification. AM-DSB and WBFM are analog modulations. They share similarities in amplitude, leading to potential confusion.
TCN-GRU improves discrimination between QAM16 and QAM64 on both datasets and ranks high in distinguishing QPSK and 8PSK. Regarding AM-DSB and WBFM, the TCN-GRU model shows better discrimination ability on RadioML2016.10b, though it performs slightly less effectively on RadioML2016.10a.
Uniform distribution confusions in the confusion matrices, such as AM-SSB for RadioML2016.10a and 8PSK and QPSK for RadioML2016.10b, were observed across all models. This occurs when the signal is submerged in noise at extremely low SNR, preventing effective recognition. Models tend to select the modulation mode most similar to noise as the most likely outcome. Using RadioML2016.10a, AM-SSB resembles white noise across various SNRs, leading to the misidentification signals in extremely low SNR as AM-SSB, thus creating a uniform confusion pattern in the AM-SSB column. Consequently, AM-SSB achieves high recognition accuracy at all SNR levels. Similarly, using RadioML2016.10b, where AM-SSB is absent, signals in extremely low SNR are often misclassified as QPSK or 8PSK.
Due to high-power noise interference, all models show the aforementioned errors in the extremely low SNR interval, which can obscure their true capabilities. To address this, we excluded extremely low SNR signals from the test set, focusing on recognition performance from −8 dB to 18 dB in terms of accuracy. Table 5 and Table 6 show the recognition accuracies on the RadioML2016.10a and RadioML2016.10b datasets with extremely low SNR signals removed, providing a clearer assessment of the performance of each model over the range of SNRs in which most communication devices can work properly.
On RadioML2016.10a, our model significantly improves the discrimination ability of QPSK/8PSK and QAM16/QAM64 compared with other models. It also achieves a high OA for AM-DSB/WBFM, with an average accuracy of 0.6675. This performance is surpassed only by ResNet (0.6995). For other modulations, like PAM4, CPFSK, BPSK, GFSK, and AM-SSB, differences in classification accuracy across models are relatively small. On RadioML2016.10b, our model is also improved in classifying QAM16/QAM64 over other models. The OA for QPSK/8PSK and AM-DSB/WBFM are 0.8395 and 0.6725 higher, respectively, than MCLDNN’s accuracies of 0.8160 and 0.6645. Model performance is similarly competitive for PAM4, CPFSK, BPSK, and GFSK. These results highlight the TCN-GRU model’s enhanced capability to accurately identify specific modulated signals, particularly where other models may struggle.

4. Conclusions

In this study, we introduced a novel hybrid model TCN-GRU that combines three modules: TCN, GRU, and GAP. TCN-GRU has the characteristics of multiscale feature extraction, hybrid feature fusion, and lightweight model construction. It has the ability to improve the drawbacks of existing hybrid models in low recognition accuracy of complex modulated signals and excessive parameters. Experimental results on the RadioML2016.10a and RadioML2016.10b datasets highlight the model’s superior performance, achieving an average recognition accuracy of 0.6156 and 0.6466, respectively. In particular, TCN-GRU demonstrated a notable improvement in recognition accuracy for challenging modulations. Using RadioML2016.10a, TCN-GRU enhanced the differentiation of QAM16/QAM64 by approximately 10.5% over MCLDNN, while achieving similar performance in distinguishing QPSK/8PSK. Using RadioML2016.10b, TCN-GRU improved differentiation accuracy by approximately 6.7% for QPSK/8PSK and 1.6% for AMDSB/WBFM, while maintaining comparable performance on QAM16/QAM64. This means that multiscale feature extraction can improve the network’s signal recognition ability, and will make our model valuable in some specific communication scenarios.
Achieving high-precision modulation classification under low SNR conditions remains a significant and challenging task in this field. Even the existing SOTA method (MCLDNN) achieves an overall accuracy of only 0.4163, 0.5403, and 0.7048 at −8 dB, −6 dB, and −4 dB, respectively, using RadioML2016.10b. Our model improves accuracy by 1.10%, 4.37%, and 2.86%, respectively, compared to MCLDNN, reaching 0.4274, 0.5840, and 0.7334. However, we recognize that there are still needs for further enhancement. Additionally, TCN-GRU is a lightweight network requiring fewer parameters than other hybrid models, which reduces computational demands and lowers the risk of overfitting, thereby enhancing its generalization capability and adaptability for real-world deployment.
In future work, we will focus on the current limitations of TCN-GRU, using methods such as model optimization, integration of advanced denoising techniques, and transfer learning to enhance our model. These enhancements are designed to improve inference speed and classification accuracy at low SNR, enabling the model to learn a wider variety of signals with fewer samples and achieve higher efficiency to better adapt to new wireless standards.

Author Contributions

Conceptualization, Z.W. and P.T.; methodology, Z.W. and W.Z.; software, Z.W.; validation, W.Z. and Z.Z. (Zhitao Zhao); data curation, Z.W. and Z.Z. (Zhitao Zhao); writing—original draft, Z.W.; writing—review and editing, P.T. and Z.Z. (Zheng Zhang); visualization, Z.W.; supervision, P.T. and Z.Z. (Zheng Zhang). All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Youth Innovation Promotion Association, Chinese Academy of Sciences (No. 2022127) and the Science and Disruptive Technology Program, Aerospace Information Research Institute, Chinese Academy of Sciences (No. E3Z211010F).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data are contained within the article.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Zhu, Z.; Nandi, A.K. Automatic Modulation Classification: Principles, Algorithms and Applications; John Wiley & Sons: Hoboken, NJ, USA, 2015. [Google Scholar]
  2. Zhang, Z.; Luo, H.; Wang, C.; Gan, C.; Xiang, Y. Automatic modulation classification using CNN-LSTM based dual-stream structure. IEEE Trans. Veh. Technol. 2020, 69, 13521–13531. [Google Scholar] [CrossRef]
  3. Peng, S.; Sun, S.; Yao, Y.D. A survey of modulation classification using deep learning: Signal representation and data preprocessing. IEEE Trans. Neural Netw. Learn. Syst. 2021, 33, 7020–7038. [Google Scholar] [CrossRef] [PubMed]
  4. Moser, E.; Moran, M.K.; Hillen, E.; Li, D.; Wu, Z. Automatic modulation classification via instantaneous features. In Proceedings of the 2015 National Aerospace and Electronics Conference (NAECON), Dayton, OH, USA, 15–19 June 2015; pp. 218–223. [Google Scholar]
  5. Simic, M.; Stanković, M.; Orlic, V.D. Automatic Modulation Classification of Real Signals in AWGN Channel Based on Sixth-Order Cumulants. Radioengineering 2021, 30, 204. [Google Scholar] [CrossRef]
  6. Wang, S.; Yue, J.; Lu, J.; Zheng, N.; Zhang, L.; Han, Y. Application of high-order cumulant based on the transient sequence in modulation recognition. In Proceedings of the 2023 5th International Conference on Frontiers Technology of Information and Computer (ICFTIC), Qingdao, China, 17–19 November 2023; pp. 268–273. [Google Scholar]
  7. Hazza, A.; Shoaib, M.; Alshebeili, S.A.; Fahad, A. An overview of feature-based methods for digital modulation classification. In Proceedings of the 2013 1st International Conference on Communications, Signal Processing, and Their Applications (ICCSPA), Sharjah, United Arab Emirates, 12–14 February 2013; pp. 1–6. [Google Scholar]
  8. Chen, W.; Jiang, Y.; Zhang, L.; Zhang, Y. A new modulation recognition method based on wavelet transform and high-order cumulants. J. Phys. Conf. Ser. 2021, 1738, 012025. [Google Scholar] [CrossRef]
  9. Mobasseri, B.G. Digital modulation classification using constellation shape. Signal Process. 2000, 80, 251–277. [Google Scholar] [CrossRef]
  10. Tekbıyık, K.; Akbunar, Ö.; Ekti, A.R.; Görçin, A.; Kurt, G.K. Multi–Dimensional Wireless Signal Identification Based on Support Vector Machines. IEEE Access 2019, 7, 138890–138903. [Google Scholar] [CrossRef]
  11. Sun, X.; Su, S.; Zuo, Z.; Guo, X.; Tan, X. Modulation Classification Using Compressed Sensing and Decision Tree–Support Vector Machine in Cognitive Radio System. Sensors 2020, 20, 1438. [Google Scholar] [CrossRef] [PubMed]
  12. Mao, Y.; Wang, Y.; Huang, W.; Qin, H.; Huang, D.; Guo, Y. Hidden-Markov-model-based calibration-attack recognition for continuous-variable quantum key distribution. Phys. Rev. A 2020, 101, 062320. [Google Scholar] [CrossRef]
  13. Ansari, S.; Alnajjar, K.A.; Abdallah, S.; Saad, M. Automatic Digital Modulation Recognition Based on Machine Learning Algorithms. In Proceedings of the 2020 International Conference on Communications, Computing, Cybersecurity, and Informatics (CCCI), Sharjah, United Arab Emirates, 3–5 November 2020; pp. 1–6. [Google Scholar]
  14. Wang, Y.; Wang, J.; Zhang, W.; Yang, J.; Gui, G. Deep learning-based cooperative automatic modulation classification method for MIMO systems. IEEE Trans. Veh. Technol. 2020, 69, 4575–4579. [Google Scholar] [CrossRef]
  15. Zhang, F.; Luo, C.; Xu, J.; Luo, Y.; Zheng, F.C. Deep learning based automatic modulation recognition: Models, datasets, and challenges. Digit. Signal Process. 2022, 129, 103650. [Google Scholar] [CrossRef]
  16. Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; Rabinovich, A. Going deeper with convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 1–9. [Google Scholar]
  17. Graves, A.; Mohamed, A.r.; Hinton, G. Speech recognition with deep recurrent neural networks. In Proceedings of the 2013 IEEE International Conference on Acoustics, Speech and Signal Processing, Vancouver, BC, USA, 26–31 May 2013; pp. 6645–6649. [Google Scholar]
  18. O’Shea, T.J.; Corgan, J.; Clancy, T.C. Convolutional radio modulation recognition networks. In Proceedings of the Engineering Applications of Neural Networks: 17th International Conference, EANN 2016, Aberdeen, UK, 2–5 September 2016; Proceedings 17. Springer: Cham, Switzerland, 2016; pp. 213–226. [Google Scholar]
  19. Liu, X.; Yang, D.; El Gamal, A. Deep neural network architectures for modulation classification. In Proceedings of the 2017 51st Asilomar Conference on Signals, Systems, and Computers, Grove, CA, USA, 29 October–1 November 2017; pp. 915–919. [Google Scholar]
  20. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  21. Hochreiter, S.; Schmidhuber, J. Long short-term memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef] [PubMed]
  22. Rajendran, S.; Meert, W.; Giustiniano, D.; Lenders, V.; Pollin, S. Deep learning models for wireless signal classification with distributed low-cost spectrum sensors. IEEE Trans. Cogn. Commun. Netw. 2018, 4, 433–445. [Google Scholar] [CrossRef]
  23. Xu, J.; Luo, C.; Parr, G.; Luo, Y. A spatiotemporal multi-channel learning framework for automatic modulation recognition. IEEE Wirel. Commun. Lett. 2020, 9, 1629–1632. [Google Scholar] [CrossRef]
  24. Cho, K.; Van Merriënboer, B.; Gulcehre, C.; Bahdanau, D.; Bougares, F.; Schwenk, H.; Bengio, Y. Learning phrase representations using RNN encoder-decoder for statistical machine translation. arXiv 2014, arXiv:1406.1078. [Google Scholar]
  25. Hong, D.; Zhang, Z.; Xu, X. Automatic modulation classification using recurrent neural networks. In Proceedings of the 2017 3rd IEEE International Conference on Computer and Communications (ICCC), Chengdu, China, 13–16 December 2017; pp. 695–700. [Google Scholar]
  26. West, N.E.; O’shea, T. Deep architectures for modulation recognition. In Proceedings of the 2017 IEEE International Symposium on Dynamic Spectrum Access Networks (DySPAN), Baltimore, MD, USA, 6–9 March 2017; pp. 1–6. [Google Scholar]
  27. Huang, G.; Liu, Z.; Van Der Maaten, L.; Weinberger, K.Q. Densely connected convolutional networks. In Proceedings of the Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 4700–4708. [Google Scholar]
  28. Wang, D.; Lin, M.; Zhang, X.; Huang, Y.; Zhu, Y. Automatic Modulation Classification Based on CNN-Transformer Graph Neural Network. Sensors 2023, 23, 7281. [Google Scholar] [CrossRef] [PubMed]
  29. O’shea, T.J.; West, N. Radio machine learning dataset generation with gnu radio. In Proceedings of the GNU Radio Conference, Boulder, CO, USA, 12–16 September 2016; Volume 1. [Google Scholar]
  30. Bai, S.; Kolter, J.Z.; Koltun, V. An empirical evaluation of generic convolutional and recurrent networks for sequence modeling. arXiv 2018, arXiv:1803.01271. [Google Scholar]
  31. Kalchbrenner, N.; Espeholt, L.; Simonyan, K.; Oord, A.v.d.; Graves, A.; Kavukcuoglu, K. Neural machine translation in linear time. arXiv 2016, arXiv:1610.10099. [Google Scholar]
  32. Yu, F.; Koltun, V. Multi-scale context aggregation by dilated convolutions. arXiv 2015, arXiv:1511.07122. [Google Scholar]
  33. Long, J.; Shelhamer, E.; Darrell, T. Fully convolutional networks for semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 3431–3440. [Google Scholar]
  34. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
  35. Njoku, J.N.; Morocho-Cayamcela, M.E.; Lim, W. CGDNet: Efficient hybrid deep learning model for robust automatic modulation recognition. IEEE Netw. Lett. 2021, 3, 47–51. [Google Scholar] [CrossRef]
Figure 1. TCNRB is the foundational module of TCN. It can accommodate arbitrary lengths of inputs while ensuring equal length for both input and output. TCNRB is also characterized by its flexible receptive field size [30].
Figure 1. TCNRB is the foundational module of TCN. It can accommodate arbitrary lengths of inputs while ensuring equal length for both input and output. TCNRB is also characterized by its flexible receptive field size [30].
Sensors 24 07908 g001
Figure 2. The structure of the GRU cell, which includes only two gates, taking into account both long-term and short-term memory.
Figure 2. The structure of the GRU cell, which includes only two gates, taking into account both long-term and short-term memory.
Sensors 24 07908 g002
Figure 3. The whole process of AMC in this study. The TCN network, comprising a four-layer TCNRB, a two-layer GRU, and a GAP, is engineered to discern the multiscale and temporal features of AP complex signals for effective AMC.
Figure 3. The whole process of AMC in this study. The TCN network, comprising a four-layer TCNRB, a two-layer GRU, and a GAP, is engineered to discern the multiscale and temporal features of AP complex signals for effective AMC.
Sensors 24 07908 g003
Figure 4. Three comparative experiments are designed to find a set of parameters that give the model the highest recognition accuracy. The results of these three experiments are: (a) comparison between layers, (b) comparison between hidden size, and (c) comparison between kernel size.
Figure 4. Three comparative experiments are designed to find a set of parameters that give the model the highest recognition accuracy. The results of these three experiments are: (a) comparison between layers, (b) comparison between hidden size, and (c) comparison between kernel size.
Sensors 24 07908 g004
Figure 5. Overall accuracy of models at each SNR level. (a) RadioML2016.10a; (b) RadioML2016.10b.
Figure 5. Overall accuracy of models at each SNR level. (a) RadioML2016.10a; (b) RadioML2016.10b.
Sensors 24 07908 g005
Figure 6. Overall accuracy of controlled models at each SNR level. (a) RadioML2016.10a; (b) RadioML2016.10b.
Figure 6. Overall accuracy of controlled models at each SNR level. (a) RadioML2016.10a; (b) RadioML2016.10b.
Sensors 24 07908 g006
Figure 7. Confusion matrices for RadioML2016.10a. (a) CNN2; (b) ResNet; (c) LSTM2; (d) GRU2; (e) CGDNet; (f) CLDNN; (g) MCLDNN; (h) TCN-GRU. The color depth proportional to the predicted percentage of the corresponding category.
Figure 7. Confusion matrices for RadioML2016.10a. (a) CNN2; (b) ResNet; (c) LSTM2; (d) GRU2; (e) CGDNet; (f) CLDNN; (g) MCLDNN; (h) TCN-GRU. The color depth proportional to the predicted percentage of the corresponding category.
Sensors 24 07908 g007
Figure 8. Confusion matrices for RadioML2016.10b. (a) CNN2; (b) ResNet; (c) LSTM2; (d) GRU2; (e) CGDNet; (f) CLDNN; (g) MCLDNN; (h) TCN-GRU. The color depth proportional to the predicted percentage of the corresponding category.
Figure 8. Confusion matrices for RadioML2016.10b. (a) CNN2; (b) ResNet; (c) LSTM2; (d) GRU2; (e) CGDNet; (f) CLDNN; (g) MCLDNN; (h) TCN-GRU. The color depth proportional to the predicted percentage of the corresponding category.
Sensors 24 07908 g008
Table 1. Details of the RadioML2016.10a and RadioML2016.10b datasets used in the experiment. RadioML2016.10a includes 11 types of modulated signals, while RadioML2016.10b has 10 types. Both datasets cover an SNR range from −20 dB to 18 dB in 2 dB increments.
Table 1. Details of the RadioML2016.10a and RadioML2016.10b datasets used in the experiment. RadioML2016.10a includes 11 types of modulated signals, while RadioML2016.10b has 10 types. Both datasets cover an SNR range from −20 dB to 18 dB in 2 dB increments.
DatasetSampleSNRModulation Types
RadioML2016.10a220,000−20:2:18 dBBPSK, QPSK, 8PSK, QAM16, QAM64, GFSK, CPFSK, PAM4, WBFM, AM-SSB, AM-DSB
RadioML2016.10b1,200,000−20:2:18 dBBPSK, QPSK, 8PSK, QAM16, QAM64, GFSK, CPFSK, PAM4, WBFM, AM-DSB
Table 2. The specific structure of each model, including the CNN models, the RNN models, the hybrid models, and the control group models for the ablation experiment.
Table 2. The specific structure of each model, including the CNN models, the RNN models, the hybrid models, and the control group models for the ablation experiment.
ModelStructure
CNN2 [18]I/Q-Input, Conv1( c : 64 , k : 1 × 8 ), Conv2( c : 16 , k : 2 × 8 ), FC1(128), FC2(N)
ResNet [19]I/Q-Input, Conv1( c : 256 , k : 1 × 3 ), Conv2( c : 256 , k : 2 × 3 ), Concat(I/Q-Input, Conv3( c : 80 , k : 13 )), Conv4( c : 80 , k : 13 ), FC1(128), FC2(N)
LSTM2 [22]A/P-Input, LSTM( h : 128 ) × 2, FC(N)
GRU2 [25]I/Q-Input, GRU( h : 128 ) × 2, FC(N)
CGDNet [35]I/Q-Input, Conv1( c : 50 , k : 1 × 6 ), MaxPool1( k : 2 × 2 ), Conv2( c : 50 , k : 1 × 6 ), MaxPool2( k : 2 × 2 ), Conv3( c : 50 , k : 1 × 6 ), MaxPool3( k : 2 × 2 ), Concat with MaxPool1, GRU( h : 50 ), FC1(256), FC2(N)
CLDNN [19]I/Q-Input, Conv1( c : 256 , k : 1 × 3 ), Conv2( c : 256 , k : 2 × 3 ), Conv3( c : 80 , k : 1 × 3 ) × 2, LSTM( h : 50 ), FC1(128), FC2(N)
MCLDNN [23]Concat1(I-Input, Conv1( c : 50 , k : 1 × 8 ), Q-Input, Conv2( c : 50 , k : 1 × 8 )), I/Q-Input, Conv3( c : 50 , k : 1 × 8 ), Concat2(Concat1, Conv4( c : 50 , k : 2 × 8 )), Conv5( c : 100 , k : 2 × 8 ), LSTM( h : 128 ), FC1(128), FC2(128), Softmax(N)
GRU2-h64: wopA/P-Input, GRU( h : 64 ) × 2, FC(N)
GRU2-h64A/P-Input, GRU( h : 64 ) × 2, GAP, FC(N)
TCN4: wopA/P-Input, TCN ResBlock( h : 64 , k : 7 ) × 4, FC(N)
TCN4A/P-Input, TCN ResBlock( h : 64 , k : 7 ) × 4, GAP, FC(N)
TCN-GRU: wopA/P-Input, TCN ResBlock( h : 64 , k : 7 ) × 4, GRU( h : 64 ) × 2, FC(N)
TCN-GRUA/P-Input, TCN ResBlock( h : 64 , k : 7 ) × 4, GRU( h : 64 ) × 2, GAP, FC(N)
Table 3. Comparison of parameters, performance evaluation indicators, and prediction time among models for the RadioML2016.10a dataset. It should be noted that the data types of RadioML2016.10a are uniformly distributed, which makes Accuracy and Recall equal in value.
Table 3. Comparison of parameters, performance evaluation indicators, and prediction time among models for the RadioML2016.10a dataset. It should be noted that the data types of RadioML2016.10a are uniformly distributed, which makes Accuracy and Recall equal in value.
ModelParametersValid LossAccuracyF1-ScoreRecallPrecisionPrediction Time (ms)
CNN21,592,3831.14340.56980.58190.56980.673343.7356
ResNet3,098,2831.18090.56380.58610.56380.689161.7528
LSTM2201,0991.11960.58490.60330.58490.694529.2208
GRU2151,1791.12920.56920.58660.56920.680125.2324
CGDNet124,9331.12880.57890.59350.57890.705921.1059
CLDNN517,6431.10020.60560.62750.60560.720664.4990
MCLDNN406,1991.08240.61010.63440.61010.7321128.4710
GRU2-h64: wop38,7311.13260.57080.58850.57080.678718.0023
GRU2-h6438,7311.10740.59160.60770.58970.683417.4842
TCN4: wop203,5311.12850.57580.59450.57380.7031113.3784
TCN4203,6601.11740.58200.60340.58200.7018116.3363
TCN-GRU: wop253,4511.09140.61430.62780.60850.7199132.2325
TCN-GRU253,4511.07970.61560.63890.61560.7400136.1084
Table 4. Similar to Table 3, but using the RadioML2016.10b dataset.
Table 4. Similar to Table 3, but using the RadioML2016.10b dataset.
ModelParametersValid LossAccuracyF1-ScoreRecallPrecisionPrediction Time (ms)
CNN21,592,1260.97810.59440.59300.59440.6402229.6624
ResNet3,098,1540.96310.61230.62590.61230.7001325.3973
LSTM2200,9700.88180.64420.64480.64420.6900148.6904
GRU2151,0500.89650.63930.64390.63930.7096130.3700
CGDNet124,6760.93860.62170.62550.62170.6785114.5280
CLDNN517,5140.93660.62690.63060.62690.7014346.8208
MCLDNN406,0700.88640.64620.64440.64620.6761705.3022
GRU2-h64: wop38,6660.89850.64020.64030.63860.6913103.8465
GRU2-h6438,6660.89360.63980.64470.63920.7172106.9712
TCN4: wop203,5950.92390.62730.62820.62680.6801630.0027
TCN4203,4660.88420.64310.64770.64310.7066635.1041
TCN-GRU: wop253,3860.88240.64450.65310.64390.7254732.0409
TCN-GRU253,3860.87550.64660.65910.64660.7536743.2286
Table 5. Accuracy of each modulation in low SNR and high SNR interval (−8 dB to 18 dB), RadioML2016.10a.
Table 5. Accuracy of each modulation in low SNR and high SNR interval (−8 dB to 18 dB), RadioML2016.10a.
ModelQPSK8PSKAM-DSBWBFMQAM16QAM64PAM4CPFSKBPSKGFSKAM-SSB
CNN20.7850.8180.9030.3790.4560.8860.9600.8620.8720.8590.937
ResNet0.7960.8220.9420.4570.5920.7700.9710.9260.9240.9600.950
LSTM20.8030.7610.8010.5180.6920.4640.9710.8740.8500.9060.950
GRU20.7930.7890.7020.5880.6250.4970.9680.8890.8640.9230.943
CGDNet0.8010.7370.9670.2760.6390.6130.9590.8800.8710.9050.951
CLDNN0.8010.8000.9360.3710.7420.6410.9560.8820.8570.9200.940
MCLDNN0.8510.7860.8960.4360.8560.7440.9760.8830.9010.9410.936
TCN-GRU0.8590.8280.7180.6170.8770.9010.9790.8790.8840.9240.932
Table 6. Accuracy of each modulation in low SNR and high SNR interval (−8 dB to 18 dB), RadioML2016.10b.
Table 6. Accuracy of each modulation in low SNR and high SNR interval (−8 dB to 18 dB), RadioML2016.10b.
ModelQPSK8PSKAM-DSBWBFMQAM16QAM64PAM4CPFSKBPSKGFSK
CNN20.8170.8660.9720.3620.5530.7100.9690.9290.8940.947
ResNet0.7900.9210.9500.3750.7750.8840.9600.9040.8930.945
LSTM20.8390.8230.9440.4130.8640.9600.9740.9330.8820.946
GRU20.8580.7890.9580.3880.8810.9360.9730.9310.8850.943
CGDNet0.7980.8180.9530.3710.7890.7740.9710.9230.8920.927
CLDNN0.8090.8220.9490.3540.8350.8100.9570.9150.8850.937
MCLDNN0.7980.8340.9520.3770.8900.9510.9730.9100.8750.944
TCN-GRU0.8810.7980.8930.4520.8930.9620.9750.9270.8880.952
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wang, Z.; Zhang, W.; Zhao, Z.; Tang, P.; Zhang, Z. Robust Automatic Modulation Classification via a Lightweight Temporal Hybrid Neural Network. Sensors 2024, 24, 7908. https://doi.org/10.3390/s24247908

AMA Style

Wang Z, Zhang W, Zhao Z, Tang P, Zhang Z. Robust Automatic Modulation Classification via a Lightweight Temporal Hybrid Neural Network. Sensors. 2024; 24(24):7908. https://doi.org/10.3390/s24247908

Chicago/Turabian Style

Wang, Zhao, Weixiong Zhang, Zhitao Zhao, Ping Tang, and Zheng Zhang. 2024. "Robust Automatic Modulation Classification via a Lightweight Temporal Hybrid Neural Network" Sensors 24, no. 24: 7908. https://doi.org/10.3390/s24247908

APA Style

Wang, Z., Zhang, W., Zhao, Z., Tang, P., & Zhang, Z. (2024). Robust Automatic Modulation Classification via a Lightweight Temporal Hybrid Neural Network. Sensors, 24(24), 7908. https://doi.org/10.3390/s24247908

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop