Abstract
Automatic modulation classification (AMC) is a fundamental technique in wireless communication systems, as it enables the identification of modulation schemes at the receiver without prior knowledge, thereby promoting efficient spectrum utilization. Recent advancements in deep learning (DL) have significantly enhanced classification performance by enabling neural networks (NNs) to learn complex decision boundaries directly from raw signal data. However, many existing NN-based AMC methods employ deep or specialized network architectures, which, while effective, tend to involve substantial structural complexity. To address this issue, we present a simple NN architecture that utilizes features derived from Hankelized matrices to extract informative signal representations. In the proposed approach, received signals are first transformed into Hankelized matrices, from which informative features are extracted using singular value decomposition (SVD). These features are then fed into a compact, fully connected (FC) NN for modulation classification across a wide range of signal-to-noise ratio (SNR) levels. Despite its architectural simplicity, the proposed method achieves competitive performance, offering a practical and scalable solution for AMC tasks at the receiver in diverse wireless environments.
1. Introduction
Automatic modulation classification (AMC) is a critical technique in wireless communication systems, serving as a key enabler for dynamic spectrum access and spectrum monitoring [1]. By automatically identifying the modulation schemes of received signals, AMC enables adaptive communication strategies, interference mitigation, and dynamic spectrum management [2,3]. Its importance is further magnified in modern wireless networks, where diverse standards and services coexist, necessitating robust classification under complex channel conditions [1,4]. Emerging technologies such as cognitive radio and dynamic spectrum sharing increasingly rely on AMC for efficient and intelligent spectrum utilization. In particular, adaptive modulation strategies require the receiver to identify modulation formats without prior transmitter coordination, highlighting the need for blind and reliable classification methods capable of operating under uncertainty [5,6]. Such requirements demand classification systems that not only generalize well across modulation types and channel conditions but also operate with low computational cost, especially for real-time or resource-constrained scenarios. As communication environments grow more dynamic and heterogeneous, the development of computationally efficient and reliable AMC algorithms remains a pressing research focus.
Conventional AMC methods can generally be categorized into likelihood-based and feature-based approaches [3,7]. The likelihood-based methods utilize probabilistic models and hypothesis testing to classify modulation schemes accurately under precise channel conditions but often incur prohibitively high computational complexity due to extensive parameter estimation processes, limiting practical deployment [8,9,10]. Conversely, feature-based methods involve classical machine learning (ML) algorithms such as support vector machines (SVMs), decision trees, and the k-nearest neighbors (k-NN) algorithm, which classify handcrafted signal features [3,11]. Although feature-based methods are generally simpler and computationally efficient, they often suffer from limitations related to the discriminative power of manually designed features and restricted learning capabilities inherent in conventional classifiers [12,13]. Moreover, such handcrafted approaches may not generalize well across different channel environments or unforeseen modulation formats.
Recently, neural networks (NNs) have gained considerable attention in AMC research due to their ability to automatically learn hierarchical features directly from raw input signals. By leveraging architectures such as convolutional neural networks (CNNs) [4,14,15], recurrent neural networks (RNNs) [16], and Transformers [17,18], deep learning (DL)-based approaches have demonstrated strong performance across a wide range of channel conditions and signal-to-noise ratio (SNR) levels [4,19]. These models offer greater flexibility and accuracy compared to conventional methods by capturing complex signal patterns without the need for manual feature engineering. In particular, their ability to learn from diverse signal variations makes them suitable for handling the increasing diversity of modulation schemes and channel impairments in wireless systems. Despite these advantages, DL-based AMC models typically require large amounts of data and computational resources and often utilize high-dimensional input representations that limit their practical applicability in real-time or resource-constrained environments. Moreover, many DL frameworks do not explicitly exploit the structured properties of communication signals, potentially overlooking meaningful low-dimensional features.
To overcome these limitations, we propose a novel AMC framework that transforms the received signal into a Hankelized matrix and extracts its singular values to construct a compact and information-rich feature representation. The singular values, obtained through singular value decomposition (SVD), capture key structural properties of the signal and act as a highly discriminative, low-dimensional descriptor. These singular value vectors are then used as direct inputs to a compact NN, trained to classify modulation schemes under wireless channel conditions. A distinctive advantage of the proposed method lies in its ability to effectively leverage the informative nature of singular values, which encapsulate modulation-relevant features while significantly reducing input dimensionality. This compactness allows the use of a simple NN architecture without sacrificing classification performance. In contrast to existing methods that depend on deep or complex models to learn from raw or high-dimensional data, our approach achieves competitive accuracy using a minimal NN structure thanks to the expressiveness of the singular value features. While the proposed NN is not an extremely shallow network, it maintains a relatively compact structure compared to many state-of-the-art AMC architectures, i.e., models based on deep convolutional layers, recurrent units, or Transformer blocks.
The remainder of this paper is organized as follows: Section 2 describes the proposed method, including the system model, problem formulation, and NN design via Hankelization-based preprocessing. Section 3 reports the simulation results, including the experimental setup, comparative analysis of the latent representations produced by different methods, and modulation recognition performance. Finally, Section 4 concludes the paper with a summary of key findings and discussions.
2. Proposed Framework
2.1. System Model and Problem Formulation
Consider the received signal for a single-input single-output (SISO) system, represented as
where L denotes the number of multipaths, is the transmitted modulation symbol sequence, is the frequency offset, represents the initial phase error, is the sampling rate, is the channel impulse response (CIR), is the timing offset, and is additive white Gaussian noise (AWGN) [20].
The objective is to reliably classify the modulation types of received signals embedded in noisy environments. To address this, we first define a Hankelization operator , which transforms a vector into a Hankelized matrix as follows:
where and denote the row and column dimensions of , respectively, with indicating the floor operation. We adopt a Hankelized matrix structure that closely approximates a square matrix.
In our proposed framework, we consider two choices for the input vector to the Hankelization operator ℏ. First, we apply ℏ directly to the received time-domain signal to obtain . Second, we apply the fast Fourier transform (FFT) operator to , resulting in , and construct . This frequency-domain representation is expected to highlight modulation-dependent spectral characteristics, thereby enhancing the separability of different modulation types, especially under noisy conditions. For notational convenience, we denote the resulting Hankelized matrix as , where represents the time-domain t and frequency-domain f variants, respectively. We then apply SVD to as
where and are the matrices consisting of left and right singular vectors of , respectively. Also, is a diagonal matrix whose diagonal elements are the descending ordered singular values, i.e., , .
We assume that the extracted singular values can effectively form a discriminative latent space suitable for the modulation classification task. To empirically support this assumption, we generate noise-free samples of various modulation types in the MATLAB R2024b environment and visualize their feature distributions using t-distributed stochastic neighbor embedding (t-SNE), a nonlinear dimensionality reduction technique that transforms high-dimensional data into a low-dimensional representation [21].
As shown in Figure 1a, the absolute values of the signal show relatively dispersed and overlapping patterns across different modulation types. In contrast, the singular values derived from the Hankelized matrix in Figure 1b,c exhibit more compact distributions with improved separation between modulation categories, particularly in Figure 1c. These results suggest that the singular values better capture the intrinsic structure of the signal representations.
Figure 1.
Data visualization based on t-SNE analysis. (a) Absolute values of signal. (b) Singular values of (with original signal). (c) Singular values of (with FFT-processed signal).
To provide a quantitative counterpart to the qualitative t-SNE visualization results shown in Figure 1, we compute the silhouette score [22], which evaluates how well each sample is clustered with respect to its assigned label, i.e., modulation type.
where denotes the average distance from sample i to all other samples in the same class and denotes the minimum average distance from i to all samples in any other class. The silhouette score , and the overall score is computed as the average over all samples. Values close to 1 indicate well-separated and compact clusters, while values near or below 0 suggest overlapping or poorly separated classes. As shown in Figure 2, we evaluate the silhouette score using three types of features. In Figure 2a, we vary the number of retained singular values while keeping the Hankelized matrix structure to be nearly square, following the definition in (2). On the other hand, in Figure 2b, we fix the number of singular values to use all available values and vary the row size when forming the Hankelized matrix. In both cases, singular values of (with the FFT-processed signal) consistently yield higher silhouette scores than the other approaches. Notably, as the structure of Hankelized matrix becomes more square-like, the FFT-based features show improved class separability. Therefore, in this work, we follow the Hankelization strategy that attempts to construct a nearly square matrix.
Figure 2.
Silhouette score according to (a) the number of singular values; (b) row size for the Hankelized matrix.
Based on this observation, we design a NN architecture that takes the singular values of the Hankelized matrix as input and performs modulation classification. The overall architecture of the proposed model is illustrated in Figure 3.
Figure 3.
Overview of the architecture underlying the proposed framework.
2.2. NN Design via Hankelization-Based Preprocessing
Now, we propose a harmonized framework with Hankelization and NN utilization. The relevant terms are defined as follows:
- : The loss function.
- : The collection of all trainable parameters, including weight matrices and bias vectors.
- : The parameter gradient computed with respect to .
- C: The number of modulation classes.
- M: The number of training datasets.
- N: The number of test datasets.
Let denote the ground truth modulation class index for the m-th training sample. The corresponding one-hot encoding vector is defined as:
For notational convenience, we define the vector of singular values of the Hankelized matrix as follows:
We denote the fully connected (FC) NN model by , which is defined using the above terms. Then, the optimized NN model, denoted by , is defined as follows:
where is the loss function, and we use the cross-entropy in this work.
To optimize NN model, we adopt the adaptive moment estimation (Adam) optimizer as the update rule for the model; thus, the parameter at iteration t is computed as follows:
where refers to the set of indices associated with the input/output samples in the current minibatch. The learning rate is denoted by , while and represent the first and second moment estimates, respectively. And and are the exponential moving average of momentum and the exponential moving average of root mean square propagation (RMSProp), respectively. Now, in accordance with (7), (8), and (9), we can obtain the optimized . Then, the result vector from can be obtained as follows:
Finally, the predicted modulation class index is determined as the index corresponding to the maximum value of the NN output vector:
Algorithm 1 shows the procedure of the proposed method. The performance of the optimized NN model is evaluated through numerical simulations in the following section.
| Algorithm 1 The Procedure of the Proposed Method |
| [Training phase] |
| 1: Collect training dataset . |
| 2: Select domain for all samples: (time-domain) or (frequency-domain via FFT). |
| 3: for to M do |
| 4: Build Hankelized matrix: |
| 5: Compute SVD based on (3): . |
| 6: Extract the vector consisting of singular values based on (6) . |
| 7: end for |
| 8: Optimize NN based on (7), (8), and (9). |
| [Test phase] |
| 9: Collect test dataset . |
| 10: Employ the same domain as chosen during training phase. |
| 11: for to N do |
| 12: Build Hankelized matrix as in Step 4. |
| 13: Make the input vector referring to Step 5 and 6. |
| 14: Predict modulation class index based on (10) and (11):, . |
| 15: end for |
3. Simulation Results
3.1. Simulation Configurations
The experiments in this study are conducted using the widely used RadioML2016.10a dataset introduced in [23], which is commonly adopted for evaluating AMC performance. The dataset configuration, including channel model parameters and signal generation settings, is summarized in Table 1. Additionally, Table 2 presents the default experimental setup, including the NN architecture, hyperparameters, and computational environment.
Table 1.
Settings of the RadioML2016.10a dataset [20,23].
Table 2.
Default configuration in the experiments.
The performance metric employed in this paper is the detection rate, defined as the ratio of correctly classified samples to the total number of test samples N. The detection rate is computed as
where denotes the cardinality of the set. This metric reflects the classification accuracy across all test samples.
The effectiveness of the proposed method was assessed through comparison with four baseline methods:
- Deep learning (CNN with real and imaginary values of the signal: The NN-based method that utilizes a CNN architecture, in which the real and imaginary parts of the input signal are treated as two separate input channels. This structure follows the design presented in [24].
- SCNN2: The NN-based method that transforms raw complex signals into spectrogram images via discrete short-time Fourier transform (STFT), applies Gaussian filtering for noise reduction, and performs classification using a dedicated CNN architecture optimized for time-frequency representations, as described in [14].
- CLDNN: The NN-based method that combines convolutional layers for local feature extraction, long short-term memory (LSTM) layers for modeling temporal dependencies, and FC layers for classification. This hybrid architecture leverages both spatial and temporal information embedded in the received signal, following the design principles in [16].
- LSTMDAE: The NN-based method that employs a denoising autoencoder (DAE) based on LSTM networks, which learns robust latent representations of noisy signals through temporal masking and reconstruction. The decoder output is jointly optimized with a modulation classification objective, thereby improving classification performance under noisy conditions, as introduced in [25].
Among the 11 modulation types in the RadioML2016.10a dataset, we select BPSK, QPSK, 16QAM, and 64QAM for evaluation. These schemes are chosen to cover both phase modulation and amplitude–phase combined modulation, allowing us to assess classification performance using a simple NN and minimal preprocessing.
3.2. Empirical Latent Space Comparison Across Models
To investigate the feature representation capability of each model, we empirically visualize the latent vectors obtained from the final hidden layer prior to classification. Let denote the set of considered models, including the proposed methods and three baseline schemes. For each model , the optimized NN model is denoted by , which can be decomposed as
where ∘ is the composition operator of the function, denotes the feature mapping function up to the penultimate layer, and is the final classification layer of model g.
Given a model-specific input , the corresponding latent feature vector is obtained as
To facilitate visualization, the output dimension of was fixed to 3 for all in this analysis. Note, however, that the original models used for classification in the next subsection employ higher-dimensional latent spaces.
Figure 4 shows the resulting 3D scatters of latent vectors , categorized by modulation classes, i.e., BPSK, QPSK, 16QAM, and 64QAM. Each subfigure corresponds to a different model . As seen in Figure 4a,b, the proposed methods result in more structured and discriminative embeddings for modulation types, whereas the baseline models exhibit more entangled and less separable latent representations.
Figure 4.
Data visualization based on latent space (SNR = 10 dB). (a) Proposed method (with original signal). (b) Proposed method (with FFT-processed signal). (c) Deep learning (CNN with real and imaginary values of signal). (d) SCNN2. (e) CLDNN. (f) LSTMDAE.
3.3. Performance Evaluation for Modulation Recognition
In Figure 5, confusion matrices are shown for six methods, each evaluated at 10 dB SNR conditions across four modulation schemes. In each matrix, rows indicate the ground-truth modulation classes, whereas columns correspond to the predicted classes. For instance, as seen in Figure 5a a value of in the third row and third column indicates that of 16QAM samples were correctly classified as 16QAM, whereas were incorrectly classified as 16QAM, and so forth. As shown in Figure 5b, the proposed method (with the FFT-processed signal) achieves the highest classification accuracy across almost all modulation types compared to other baseline methods. Among the baseline methods, SCNN2 in Figure 5d shows competitive performance, likely due to its spectrogram-based preprocessing with Gaussian filtering, which effectively suppresses noise. While the proposed method with original signals (Figure 5a) does not employ explicit denoising, it still achieves high accuracy for BPSK, 16QAM, and 64QAM, while exhibiting limited discriminability between QPSK and BPSK.
Figure 5.
Confusion matrices (SNR = 10 dB). (a) Proposed method (with original signal). (b) Proposed method (with FFT-processed signal). (c) Deep learning (CNN with real and imaginary values of signal). (d) SCNN2. (e) CLDNN. (f) LSTMDAE.
Figure 6 shows the detection rate according to the number of training datasets. For instance, a total of 2800 training samples corresponds to 700 samples per modulation class when four modulation types are considered.
Figure 6.
Detection rate according to the number of training datasets. (a) SNR = 4 dB. (b) SNR = 10 dB. (c) SNR = 16 dB.
Across all SNR levels and training sizes, the proposed method (with FFT-processed signal), shown in orange, consistently achieves the highest detection rate. As shown in Figure 6a, even when trained with only 560 samples in total—corresponding to 140 samples per class under the SNR condition of 4 dB—the proposed method (with FFT-processed signal) maintains a detection rate approaching , demonstrating strong robustness under limited data availability.
Table 3 presents the average detection rate and corresponding standard deviation across 10 independent trials, each conducted with different random seeds, for various SNR levels ranging from to 18 dB. The boldfaced values indicate the SNR points where the proposed methods achieve the highest mean detection rate among all evaluated methods. From 0 dB and above, the proposed method (with the FFT-processed signal) demonstrates superiority, yielding the highest detection rates across all conditions in this range. These results collectively validate that the proposed framework, particularly when combined with FFT preprocessing, achieves superior classification performance compared to baseline methods, demonstrating robustness and stability across various SNR conditions.
Table 3.
Mean of detection rate ± standard deviation according to SNR.
Figure 7 shows the test error curves across training epochs. The proposed method (with FFT-processed signal), shown in orange, consistently exhibits the fastest convergence and lowest test error across all SNR levels. Notably, under the SNR condition of 4 dB, as shown in Figure 7a, the proposed method (with FFT-processed signal) achieves a near-minimal test error within 30 epochs. The results highlight that the FFT-based method achieves faster convergence and learns robust, discriminative features effectively across varying conditions.
Figure 7.
Test error according to epochs. (a) SNR = 4 dB. (b) SNR = 10 dB. (c) SNR = 16 dB.
Figure 8 shows the detection performance of the proposed method under various SNR levels, with respect to the number of retained singular values. The Hankelized matrix is constructed to be approximately square, while the number of singular values is varied to assess its influence on classification accuracy. Across all SNR conditions, the detection rate improves as more singular values are retained, highlighting the benefit of preserving sufficient spectral information. Furthermore, the FFT-processed variant consistently outperforms the original signal variant, particularly when a sufficient number of singular values is used.
Figure 8.
Detection rate according to the number of singular values. (a) SNR = 4 dB. (b) SNR = 10 dB. (c) SNR = 16 dB.
Figure 9 presents the detection performance of the proposed method under varying row sizes of the Hankelized matrix across different SNR conditions. The number of singular values retained is fixed to the row dimension, while the row size is varied to assess the effect of the matrix shape. The results indicate that a well-chosen row size, even without strictly enforcing a near-square matrix structure, can still lead to a high detection rate. Notably, the FFT-processed variant achieves relatively higher accuracy in most SNR settings when the row size is sufficiently large.
Figure 9.
Detection rate according to row size for the Hankelized matrix. (a) SNR = 4 dB. (b) SNR = 10 dB. (c) SNR = 16 dB.
These observations suggest that high detection performance can be attained by appropriately selecting the number of singular values and the row size of the Hankelized matrix.
Table 4 presents the computational complexity of each model in terms of the number of floating-point operations (FLOPs) and network parameters required for a single inference. Recalling the width of each hidden layer, denoted by , and the depth of the NN d, the FLOP complexity of an FC model is given by . Although the proposed methods introduce an additional computational cost of due to singular value extraction via Hankelization, this step is performed as part of offline preprocessing. As shown in Table 4, both variants of the proposed method—one using the original signal and the other using the FFT-processed signal—require only FLOPs and parameters, demonstrating significantly lower complexity compared to other baseline methods. In particular, CNN-based architectures such as SCNN2 and CLDNN require up to and FLOPs, respectively, with network sizes ranging from to parameters. The baseline methods follow the same NN configuration described in Section 3.1, ensuring a fair comparison in terms of network depth and width. The findings highlight that, although a mathematically complex preprocessing step is employed, the proposed method maintains high computational efficiency.
Table 4.
Comparison of computational complexity in terms of FLOPs and the number of parameters.
4. Conclusions
In this paper, we proposed a NN-based framework for AMC, incorporatingHankelization preprocessing. The proposed approach transforms the received signals into Hankelized matrices, from which singular values are extracted to form low-dimensional yet informative feature vectors. Two variants of the method were investigated—one based on time-domain signals and the other incorporating frequency-domain features via FFT. These features are then fed into an NN composed of a simple FC structure to perform classification across various SNR environments. The key strength of the proposed framework lies in its ability to effectively extract modulation-relevant features without relying on deep or complex NN architectures. Unlike conventional methods that depend on either handcrafted features or high-capacity models, our approach employs a concise design with minimal layers and moderate width, demonstrating that high performance can still be achieved when combined with appropriate signal representations.
Experimental results confirm that the FFT-enhanced variant consistently achieves the highest detection accuracy across a wide range of SNR levels. Notably, it performs reliably even under low-SNR conditions and with limited training data, maintaining stable detection rates. In addition, the model converges rapidly within a small number of training epochs. Latent space analysis further supports the ability of the model to form well-separated representations across modulation types. In summary, the proposed Hankelization-based AMC framework provides a practical and effective solution by integrating signal structure-aware preprocessing with a simple NN architecture. Future work may extend this framework to broader modulation families or more complex channel environments.
Author Contributions
Conceptualization, W.-H.L., J.-H.L. and O.-S.S.; software, J.-H.K.; writing—original draft preparation, J.-H.K.; writing—review and editing, W.-H.L., J.-H.L. and O.-S.S.; visualization, J.-H.K.; funding acquisition, J.-H.L. All authors have read and agreed to the published version of the manuscript.
Funding
This work was supported in part by the Dongguk University Research Fund of 2024 (S-2024-G0001-00025) and in part by the National Research Foundation of Korea (NRF) grant funded by the Korea government (MSIT) (No. RS-2023-00239349).
Institutional Review Board Statement
Not applicable.
Informed Consent Statement
Not applicable.
Data Availability Statement
The data presented in this study are available on request from the corresponding author. The data are not publicly available due to privacy.
Conflicts of Interest
The authors declare no conflicts of interest.
Abbreviations
The following abbreviations are used in this manuscript:
| AMC | Automatic modulation classification |
| ML | Machine learning |
| SVM | Support vector machine |
| k-NN | k-nearest neighbors |
| NN | Neural network |
| CNN | Convolutional neural network |
| RNN | Recurrent neural network |
| DL | Deep learning |
| SNR | Signal-to-noise ratio |
| SVD | Singular value decomposition |
| SISO | Single-input single-output |
| CIR | Channel impulse response |
| AWGN | Additive white Gaussian noise |
| FFT | Fast Fourier transform |
| t-SNE | t-distributed stochastic neighbor embedding |
| FC | Fully connected |
| Adam | Adaptive moment estimation |
| RMSProp | Root mean square propagation |
| MLP | Multilayer perceptron |
| STFT | Short-time Fourier transform |
| LSTM | Long short-term memory |
| DAE | Denoising autoencoder |
| FLOP | Floating-point operation |
References
- Gui, G.; Liu, M.; Tang, F.; Kato, N.; Adachi, F. 6G: Opening new horizons for integration of comfort, security, and intelligence. IEEE Wirel. Commun. 2020, 27, 126–132. [Google Scholar] [CrossRef]
- Dobre, O.A.; Abdi, A.; Bar-Ness, Y.; Su, W. Survey of automatic modulation classification techniques: Classical approaches and new trends. IET Commun. 2007, 1, 137–156. [Google Scholar] [CrossRef]
- Huynh-The, T.; Pham, Q.V.; Nguyen, T.V.; Nguyen, T.T.; Ruby, R.; Zeng, M.; Kim, D.S. Automatic modulation classification: A deep architecture survey. IEEE Access 2021, 9, 142950–142971. [Google Scholar] [CrossRef]
- O’Shea, T.J.; Roy, T.; Clancy, T.C. Over-the-air deep learning based radio signal classification. IEEE J. Sel. Top. Signal Process. 2018, 12, 168–179. [Google Scholar] [CrossRef]
- Häring, L.; Chen, Y.; Czylwik, A. Efficient modulation classification for adaptive wireless OFDM systems in TDD mode. In Proceedings of the 2010 IEEE Wireless Communications and Networking Conference, Sydney, Australia, 18–21 April 2010; pp. 1–6. [Google Scholar]
- Harper, C.A.; Thornton, M.A.; Larson, E.C. Automatic modulation classification with deep neural networks. Electronics 2023, 12, 3962. [Google Scholar] [CrossRef]
- Fu, X.; Gui, G.; Wang, Y.; Gacanin, H.; Adachi, F. Automatic modulation classification based on decentralized learning and ensemble learning. IEEE Trans. Veh. Technol. 2022, 71, 7942–7946. [Google Scholar] [CrossRef]
- Su, W.; Xu, J.L.; Zhou, M. Real-time modulation classification based on maximum likelihood. IEEE Commun. Lett. 2008, 12, 801–803. [Google Scholar] [CrossRef]
- Xu, J.L.; Su, W.; Zhou, M. Likelihood-ratio approaches to automatic modulation classification. IEEE Trans. Syst. Man Cybern. Part C (Appl. Rev.) 2010, 41, 455–469. [Google Scholar] [CrossRef]
- Ozdemir, O.; Li, R.; Varshney, P.K. Hybrid maximum likelihood modulation classification using multiple radios. IEEE Commun. Lett. 2013, 17, 1889–1892. [Google Scholar] [CrossRef]
- Li, J.; Meng, Q.; Zhang, G.; Sun, Y.; Qiu, L.; Ma, W. Automatic modulation classification using support vector machines and error correcting output codes. In Proceedings of the 2017 IEEE 2nd Information Technology, Networking, Electronic and Automation Control Conference (ITNEC), Chengdu, China, 15–17 December 2017; pp. 60–63. [Google Scholar]
- Orlic, V.D.; Dukic, M.L. Automatic modulation classification algorithm using higher-order cumulants under real-world channel conditions. IEEE Commun. Lett. 2009, 13, 917–919. [Google Scholar] [CrossRef]
- Su, W. Feature space analysis of modulation classification using very high-order statistics. IEEE Commun. Lett. 2013, 17, 1688–1691. [Google Scholar] [CrossRef]
- Zeng, Y.; Zhang, M.; Han, F.; Gong, Y.; Zhang, J. Spectrum analysis and convolutional neural network for automatic modulation recognition. IEEE Wirel. Commun. Lett. 2019, 8, 929–932. [Google Scholar] [CrossRef]
- Huynh-The, T.; Hua, C.H.; Pham, Q.V.; Kim, D.S. MCNet: An efficient CNN architecture for robust automatic modulation classification. IEEE Commun. Lett. 2020, 24, 811–815. [Google Scholar] [CrossRef]
- West, N.E.; O’shea, T. Deep architectures for modulation recognition. In Proceedings of the 2017 IEEE International Symposium on Dynamic Spectrum Access networks (DySPAN), Baltimore, MD, USA, 6–9 March 2017; pp. 1–6. [Google Scholar]
- Hamidi-Rad, S.; Jain, S. Mcformer: A transformer based deep neural network for automatic modulation classification. In Proceedings of the 2021 IEEE Global Communications Conference (GLOBECOM), Madrid, Spain, 7–11 December 2021; pp. 1–6. [Google Scholar]
- Chen, Y.; Dong, B.; Liu, C.; Xiong, W.; Li, S. Abandon locality: Frame-wise embedding aided transformer for automatic modulation recognition. IEEE Commun. Lett. 2022, 27, 327–331. [Google Scholar] [CrossRef]
- Mao, Q.; Hu, F.; Hao, Q. Deep learning for intelligent wireless networks: A comprehensive survey. IEEE Commun. Surv. Tutor. 2018, 20, 2595–2621. [Google Scholar] [CrossRef]
- Sathyanarayanan, V.; Gerstoft, P.; El Gamal, A. RML22: Realistic dataset generation for wireless modulation classification. IEEE Trans. Wirel. Commun. 2023, 22, 7663–7675. [Google Scholar] [CrossRef]
- Van der Maaten, L.; Hinton, G. Visualizing data using t-SNE. J. Mach. Learn. Res. 2008, 9, 2579–2605. [Google Scholar]
- Rousseeuw, P.J. Silhouettes: A graphical aid to the interpretation and validation of cluster analysis. J. Comput. Appl. Math. 1987, 20, 53–65. [Google Scholar] [CrossRef]
- O’shea, T.J.; West, N. Radio machine learning dataset generation with GNU radio. In Proceedings of the GNU Radio Conference, Charlotte, NC, USA, 20–24 September 2016; Volume 1. [Google Scholar]
- Tekbıyık, K.; Ekti, A.R.; Görçin, A.; Kurt, G.K.; Keçeci, C. Robust and fast automatic modulation classification with CNN under multipath fading channels. In Proceedings of the 2020 IEEE 91st Vehicular Technology Conference (VTC2020-Spring), Antwerp, Belgium, 25–28 May 2020; pp. 1–6. [Google Scholar]
- Ke, Z.; Vikalo, H. Real-time radio technology and modulation classification via an LSTM auto-encoder. IEEE Trans. Wirel. Commun. 2021, 21, 370–382. [Google Scholar] [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).