Next Article in Journal
Non-Vertical Well Trajectory Design Based on Multi-Objective Optimization
Previous Article in Journal
Simultaneous Submicron Temperature Mapping of Substrate and Channel in P-GaN/AlGaN/GaN HEMTs Using Raman Thermometry
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Novel Neural Network Framework for Automatic Modulation Classification via Hankelization-Based Signal Transformation

by
Jung-Hwan Kim
1,
Jong-Ho Lee
2,
Oh-Soon Shin
2 and
Woong-Hee Lee
1,*
1
Division of Electronics and Electrical Engineering, Dongguk University-Seoul, Seoul 04620, Republic of Korea
2
School of Electronic Engineering, Soongsil University, Seoul 06978, Republic of Korea
*
Author to whom correspondence should be addressed.
Appl. Sci. 2025, 15(14), 7861; https://doi.org/10.3390/app15147861
Submission received: 15 June 2025 / Revised: 6 July 2025 / Accepted: 11 July 2025 / Published: 14 July 2025
(This article belongs to the Section Electrical, Electronics and Communications Engineering)

Abstract

Automatic modulation classification (AMC) is a fundamental technique in wireless communication systems, as it enables the identification of modulation schemes at the receiver without prior knowledge, thereby promoting efficient spectrum utilization. Recent advancements in deep learning (DL) have significantly enhanced classification performance by enabling neural networks (NNs) to learn complex decision boundaries directly from raw signal data. However, many existing NN-based AMC methods employ deep or specialized network architectures, which, while effective, tend to involve substantial structural complexity. To address this issue, we present a simple NN architecture that utilizes features derived from Hankelized matrices to extract informative signal representations. In the proposed approach, received signals are first transformed into Hankelized matrices, from which informative features are extracted using singular value decomposition (SVD). These features are then fed into a compact, fully connected (FC) NN for modulation classification across a wide range of signal-to-noise ratio (SNR) levels. Despite its architectural simplicity, the proposed method achieves competitive performance, offering a practical and scalable solution for AMC tasks at the receiver in diverse wireless environments.

1. Introduction

Automatic modulation classification (AMC) is a critical technique in wireless communication systems, serving as a key enabler for dynamic spectrum access and spectrum monitoring [1]. By automatically identifying the modulation schemes of received signals, AMC enables adaptive communication strategies, interference mitigation, and dynamic spectrum management [2,3]. Its importance is further magnified in modern wireless networks, where diverse standards and services coexist, necessitating robust classification under complex channel conditions [1,4]. Emerging technologies such as cognitive radio and dynamic spectrum sharing increasingly rely on AMC for efficient and intelligent spectrum utilization. In particular, adaptive modulation strategies require the receiver to identify modulation formats without prior transmitter coordination, highlighting the need for blind and reliable classification methods capable of operating under uncertainty [5,6]. Such requirements demand classification systems that not only generalize well across modulation types and channel conditions but also operate with low computational cost, especially for real-time or resource-constrained scenarios. As communication environments grow more dynamic and heterogeneous, the development of computationally efficient and reliable AMC algorithms remains a pressing research focus.
Conventional AMC methods can generally be categorized into likelihood-based and feature-based approaches [3,7]. The likelihood-based methods utilize probabilistic models and hypothesis testing to classify modulation schemes accurately under precise channel conditions but often incur prohibitively high computational complexity due to extensive parameter estimation processes, limiting practical deployment [8,9,10]. Conversely, feature-based methods involve classical machine learning (ML) algorithms such as support vector machines (SVMs), decision trees, and the k-nearest neighbors (k-NN) algorithm, which classify handcrafted signal features [3,11]. Although feature-based methods are generally simpler and computationally efficient, they often suffer from limitations related to the discriminative power of manually designed features and restricted learning capabilities inherent in conventional classifiers [12,13]. Moreover, such handcrafted approaches may not generalize well across different channel environments or unforeseen modulation formats.
Recently, neural networks (NNs) have gained considerable attention in AMC research due to their ability to automatically learn hierarchical features directly from raw input signals. By leveraging architectures such as convolutional neural networks (CNNs) [4,14,15], recurrent neural networks (RNNs) [16], and Transformers [17,18], deep learning (DL)-based approaches have demonstrated strong performance across a wide range of channel conditions and signal-to-noise ratio (SNR) levels [4,19]. These models offer greater flexibility and accuracy compared to conventional methods by capturing complex signal patterns without the need for manual feature engineering. In particular, their ability to learn from diverse signal variations makes them suitable for handling the increasing diversity of modulation schemes and channel impairments in wireless systems. Despite these advantages, DL-based AMC models typically require large amounts of data and computational resources and often utilize high-dimensional input representations that limit their practical applicability in real-time or resource-constrained environments. Moreover, many DL frameworks do not explicitly exploit the structured properties of communication signals, potentially overlooking meaningful low-dimensional features.
To overcome these limitations, we propose a novel AMC framework that transforms the received signal into a Hankelized matrix and extracts its singular values to construct a compact and information-rich feature representation. The singular values, obtained through singular value decomposition (SVD), capture key structural properties of the signal and act as a highly discriminative, low-dimensional descriptor. These singular value vectors are then used as direct inputs to a compact NN, trained to classify modulation schemes under wireless channel conditions. A distinctive advantage of the proposed method lies in its ability to effectively leverage the informative nature of singular values, which encapsulate modulation-relevant features while significantly reducing input dimensionality. This compactness allows the use of a simple NN architecture without sacrificing classification performance. In contrast to existing methods that depend on deep or complex models to learn from raw or high-dimensional data, our approach achieves competitive accuracy using a minimal NN structure thanks to the expressiveness of the singular value features. While the proposed NN is not an extremely shallow network, it maintains a relatively compact structure compared to many state-of-the-art AMC architectures, i.e., models based on deep convolutional layers, recurrent units, or Transformer blocks.
The remainder of this paper is organized as follows: Section 2 describes the proposed method, including the system model, problem formulation, and NN design via Hankelization-based preprocessing. Section 3 reports the simulation results, including the experimental setup, comparative analysis of the latent representations produced by different methods, and modulation recognition performance. Finally, Section 4 concludes the paper with a summary of key findings and discussions.

2. Proposed Framework

2.1. System Model and Problem Formulation

Consider the received signal r C P for a single-input single-output (SISO) system, represented as
r [ n ] = e j ( 2 π f e r r n T s + θ e r r ) l = 0 L 1 h [ l ] s [ n l ζ err ] + w [ n ] , n { 0 , 1 , , P 1 } ,
where L denotes the number of multipaths, s is the transmitted modulation symbol sequence, f e r r is the frequency offset, θ e r r represents the initial phase error, T s is the sampling rate, h is the channel impulse response (CIR), ζ e r r is the timing offset, and w is additive white Gaussian noise (AWGN) [20].
The objective is to reliably classify the modulation types of received signals embedded in noisy environments. To address this, we first define a Hankelization operator : C P × 1 C P T × P T , which transforms a vector x into a Hankelized matrix X as follows:
X = ( x ) s . t . X [ i , j ] = x [ i + j 1 ] , i , j { 1 , 2 , } ,
where P T = ( P + 1 / 2 ) and P T = P P T + 1 denote the row and column dimensions of X , respectively, with · indicating the floor operation. We adopt a Hankelized matrix structure that closely approximates a square matrix.
In our proposed framework, we consider two choices for the input vector x to the Hankelization operator . First, we apply directly to the received time-domain signal r to obtain R t = ( r ) . Second, we apply the fast Fourier transform (FFT) operator F to r , resulting in r ˜ = F ( r ) , and construct R f = ( r ˜ ) . This frequency-domain representation is expected to highlight modulation-dependent spectral characteristics, thereby enhancing the separability of different modulation types, especially under noisy conditions. For notational convenience, we denote the resulting Hankelized matrix as R d , where d { t , f } represents the time-domain t and frequency-domain f variants, respectively. We then apply SVD to R d as
R d = U d Σ d V d H ,
where U d C P T × P T and V d C P T × P T are the matrices consisting of P T left and P T right singular vectors of R d , respectively. Also, Σ d R P T × P T is a diagonal matrix whose diagonal elements are the descending ordered singular values, i.e., σ ( i ) d , i { 1 , , P T } .
We assume that the extracted singular values σ ( i ) d can effectively form a discriminative latent space suitable for the modulation classification task. To empirically support this assumption, we generate noise-free samples of various modulation types in the MATLAB R2024b environment and visualize their feature distributions using t-distributed stochastic neighbor embedding (t-SNE), a nonlinear dimensionality reduction technique that transforms high-dimensional data into a low-dimensional representation [21].
As shown in Figure 1a, the absolute values of the signal show relatively dispersed and overlapping patterns across different modulation types. In contrast, the singular values derived from the Hankelized matrix in Figure 1b,c exhibit more compact distributions with improved separation between modulation categories, particularly in Figure 1c. These results suggest that the singular values better capture the intrinsic structure of the signal representations.
To provide a quantitative counterpart to the qualitative t-SNE visualization results shown in Figure 1, we compute the silhouette score [22], which evaluates how well each sample is clustered with respect to its assigned label, i.e., modulation type.
s ( i ) = b ( i ) a ( i ) max { a ( i ) , b ( i ) } ,
where a ( i ) denotes the average distance from sample i to all other samples in the same class and b ( i ) denotes the minimum average distance from i to all samples in any other class. The silhouette score s ( i ) [ 1 , 1 ] , and the overall score is computed as the average over all samples. Values close to 1 indicate well-separated and compact clusters, while values near or below 0 suggest overlapping or poorly separated classes. As shown in Figure 2, we evaluate the silhouette score using three types of features. In Figure 2a, we vary the number of retained singular values while keeping the Hankelized matrix structure to be nearly square, following the definition in (2). On the other hand, in Figure 2b, we fix the number of singular values to use all available values and vary the row size when forming the Hankelized matrix. In both cases, singular values of R f (with the FFT-processed signal) consistently yield higher silhouette scores than the other approaches. Notably, as the structure of Hankelized matrix becomes more square-like, the FFT-based features show improved class separability. Therefore, in this work, we follow the Hankelization strategy that attempts to construct a nearly square matrix.
Based on this observation, we design a NN architecture that takes the singular values of the Hankelized matrix as input and performs modulation classification. The overall architecture of the proposed model is illustrated in Figure 3.

2.2. NN Design via Hankelization-Based Preprocessing

Now, we propose a harmonized framework with Hankelization and NN utilization. The relevant terms are defined as follows:
  • L : The loss function.
  • Θ : The collection of all trainable parameters, including weight matrices and bias vectors.
  • Θ : The parameter gradient computed with respect to Θ .
  • C: The number of modulation classes.
  • M: The number of training datasets.
  • N: The number of test datasets.
Let c ( m ) { 1 , 2 , , C } denote the ground truth modulation class index for the m-th training sample. The corresponding one-hot encoding vector e ( m ) { 0 , 1 } C is defined as:
e ( m ) [ c ] = 1 if c = c ( m ) , 0 otherwise .
For notational convenience, we define the vector of singular values σ d ¯ ( m ) R + P T of the Hankelized matrix as follows:
σ d ¯ ( m ) = diag ( Σ d ) .
We denote the fully connected (FC) NN model by N p , which is defined using the above terms. Then, the optimized NN model, denoted by N p , is defined as follows:
N p = arg min N p m = 1 M L ( e ( m ) , N p ( σ d ¯ ( m ) ) ) ,
where L is the loss function, and we use the cross-entropy in this work.
To optimize NN model, we adopt the adaptive moment estimation (Adam) optimizer as the update rule for the model; thus, the parameter Θ t at iteration t is computed as follows:
Θ t + 1 Θ t η · m t + 1 v t + 1 + ϵ ,
m t + 1 β 1 m t + ( 1 β 1 ) m B t Θ L ( e ( m ) , N p ( σ d ¯ ( m ) ; Θ t ) ) , v t + 1 β 2 v t + ( 1 β 2 ) m B t Θ L ( z ( m ) , N p ( σ d ¯ ( m ) ; Θ t ) ) 2 ,
where B t refers to the set of indices associated with the input/output samples in the current minibatch. The learning rate is denoted by η , while m and v represent the first and second moment estimates, respectively. And β 1 and β 2 are the exponential moving average of momentum and the exponential moving average of root mean square propagation (RMSProp), respectively. Now, in accordance with (7), (8), and (9), we can obtain the optimized N p . Then, the result vector e ^ from σ d ¯ can be obtained as follows:
e ^ = N p ( σ d ¯ ) .
Finally, the predicted modulation class index c ^ is determined as the index corresponding to the maximum value of the NN output vector:
c ^ = arg max c e ^ [ c ] .
Algorithm 1 shows the procedure of the proposed method. The performance of the optimized NN model N p is evaluated through numerical simulations in the following section.
Algorithm 1 The Procedure of the Proposed Method
  [Training phase]
    1:     Collect training dataset { ( r ( m ) , e ( m ) ) } m = 1 M .
    2:     Select domain d { t , f } for all samples:
                 d = t (time-domain) or d = f (frequency-domain via FFT).
    3:     for m 1  to M do
    4:        Build Hankelized matrix: R d ( m ) = ( r ( m ) ) if d = t , ( F ( r ( m ) ) ) if d = f .
    5:        Compute SVD based on (3): R d ( m ) = U d Σ d V d .
    6:        Extract the vector consisting of singular values based on (6)
            σ d ¯ ( m ) = diag ( Σ d ) .
    7:     end for
    8:     Optimize NN N p based on (7), (8), and (9).
  [Test phase]
    9:     Collect test dataset { r ( n ) } n = 1 N .
   10:     Employ the same domain d { t , f } as chosen during training phase.
   11:     for  n 1  to N do
   12:        Build Hankelized matrix R d ( n ) as in Step 4.
   13:        Make the input vector σ d ¯ ( n ) referring to Step 5 and 6.
   14:        Predict modulation class index based on (10) and (11): e ^ ( n ) = N p ( σ d ¯ ( n ) ) ,       
                     c ^ ( n ) = arg max c e ^ ( n ) [ c ] .
   15:     end for

3. Simulation Results

3.1. Simulation Configurations

The experiments in this study are conducted using the widely used RadioML2016.10a dataset introduced in [23], which is commonly adopted for evaluating AMC performance. The dataset configuration, including channel model parameters and signal generation settings, is summarized in Table 1. Additionally, Table 2 presents the default experimental setup, including the NN architecture, hyperparameters, and computational environment.
The performance metric employed in this paper is the detection rate, defined as the ratio of correctly classified samples to the total number of test samples N. The detection rate is computed as
detection rate = card { n c ^ ( n ) = c ( n ) } N , n { 1 , , N } ,
where card { · } denotes the cardinality of the set. This metric reflects the classification accuracy across all test samples.
The effectiveness of the proposed method was assessed through comparison with four baseline methods:
  • Deep learning (CNN with real and imaginary values of the signal: The NN-based method that utilizes a CNN architecture, in which the real and imaginary parts of the input signal are treated as two separate input channels. This structure follows the design presented in [24].
  • SCNN2: The NN-based method that transforms raw complex signals into spectrogram images via discrete short-time Fourier transform (STFT), applies Gaussian filtering for noise reduction, and performs classification using a dedicated CNN architecture optimized for time-frequency representations, as described in [14].
  • CLDNN: The NN-based method that combines convolutional layers for local feature extraction, long short-term memory (LSTM) layers for modeling temporal dependencies, and FC layers for classification. This hybrid architecture leverages both spatial and temporal information embedded in the received signal, following the design principles in [16].
  • LSTMDAE: The NN-based method that employs a denoising autoencoder (DAE) based on LSTM networks, which learns robust latent representations of noisy signals through temporal masking and reconstruction. The decoder output is jointly optimized with a modulation classification objective, thereby improving classification performance under noisy conditions, as introduced in [25].
Among the 11 modulation types in the RadioML2016.10a dataset, we select BPSK, QPSK, 16QAM, and 64QAM for evaluation. These schemes are chosen to cover both phase modulation and amplitude–phase combined modulation, allowing us to assess classification performance using a simple NN and minimal preprocessing.

3.2. Empirical Latent Space Comparison Across Models

To investigate the feature representation capability of each model, we empirically visualize the latent vectors z R 3 obtained from the final hidden layer prior to classification. Let G denote the set of considered models, including the proposed methods and three baseline schemes. For each model g G , the optimized NN model is denoted by N g , which can be decomposed as
N g = C g f latent ( g )
where ∘ is the composition operator of the function, f latent ( g ) ( · ) denotes the feature mapping function up to the penultimate layer, and C g ( · ) is the final classification layer of model g.
Given a model-specific input y , the corresponding latent feature vector is obtained as
z = f latent ( g ) ( y ) .
To facilitate visualization, the output dimension of f latent ( g ) was fixed to 3 for all g G in this analysis. Note, however, that the original models used for classification in the next subsection employ higher-dimensional latent spaces.
Figure 4 shows the resulting 3D scatters of latent vectors z R 3 , categorized by modulation classes, i.e., BPSK, QPSK, 16QAM, and 64QAM. Each subfigure corresponds to a different model g G . As seen in Figure 4a,b, the proposed methods result in more structured and discriminative embeddings for modulation types, whereas the baseline models exhibit more entangled and less separable latent representations.

3.3. Performance Evaluation for Modulation Recognition

In Figure 5, confusion matrices are shown for six methods, each evaluated at 10 dB SNR conditions across four modulation schemes. In each matrix, rows indicate the ground-truth modulation classes, whereas columns correspond to the predicted classes. For instance, as seen in Figure 5a a value of 94.0 % in the third row and third column indicates that 94.0 % of 16QAM samples were correctly classified as 16QAM, whereas 6.0 % were incorrectly classified as 16QAM, and so forth. As shown in Figure 5b, the proposed method (with the FFT-processed signal) achieves the highest classification accuracy across almost all modulation types compared to other baseline methods. Among the baseline methods, SCNN2 in Figure 5d shows competitive performance, likely due to its spectrogram-based preprocessing with Gaussian filtering, which effectively suppresses noise. While the proposed method with original signals (Figure 5a) does not employ explicit denoising, it still achieves high accuracy for BPSK, 16QAM, and 64QAM, while exhibiting limited discriminability between QPSK and BPSK.
Figure 6 shows the detection rate according to the number of training datasets. For instance, a total of 2800 training samples corresponds to 700 samples per modulation class when four modulation types are considered.
Across all SNR levels and training sizes, the proposed method (with FFT-processed signal), shown in orange, consistently achieves the highest detection rate. As shown in Figure 6a, even when trained with only 560 samples in total—corresponding to 140 samples per class under the SNR condition of 4 dB—the proposed method (with FFT-processed signal) maintains a detection rate approaching 0.9 , demonstrating strong robustness under limited data availability.
Table 3 presents the average detection rate and corresponding standard deviation across 10 independent trials, each conducted with different random seeds, for various SNR levels ranging from 20 to 18 dB. The boldfaced values indicate the SNR points where the proposed methods achieve the highest mean detection rate among all evaluated methods. From 0 dB and above, the proposed method (with the FFT-processed signal) demonstrates superiority, yielding the highest detection rates across all conditions in this range. These results collectively validate that the proposed framework, particularly when combined with FFT preprocessing, achieves superior classification performance compared to baseline methods, demonstrating robustness and stability across various SNR conditions.
Figure 7 shows the test error curves across training epochs. The proposed method (with FFT-processed signal), shown in orange, consistently exhibits the fastest convergence and lowest test error across all SNR levels. Notably, under the SNR condition of 4 dB, as shown in Figure 7a, the proposed method (with FFT-processed signal) achieves a near-minimal test error within 30 epochs. The results highlight that the FFT-based method achieves faster convergence and learns robust, discriminative features effectively across varying conditions.
Figure 8 shows the detection performance of the proposed method under various SNR levels, with respect to the number of retained singular values. The Hankelized matrix is constructed to be approximately square, while the number of singular values is varied to assess its influence on classification accuracy. Across all SNR conditions, the detection rate improves as more singular values are retained, highlighting the benefit of preserving sufficient spectral information. Furthermore, the FFT-processed variant consistently outperforms the original signal variant, particularly when a sufficient number of singular values is used.
Figure 9 presents the detection performance of the proposed method under varying row sizes of the Hankelized matrix across different SNR conditions. The number of singular values retained is fixed to the row dimension, while the row size is varied to assess the effect of the matrix shape. The results indicate that a well-chosen row size, even without strictly enforcing a near-square matrix structure, can still lead to a high detection rate. Notably, the FFT-processed variant achieves relatively higher accuracy in most SNR settings when the row size is sufficiently large.
These observations suggest that high detection performance can be attained by appropriately selecting the number of singular values and the row size of the Hankelized matrix.
Table 4 presents the computational complexity of each model in terms of the number of floating-point operations (FLOPs) and network parameters required for a single inference. Recalling the width of each hidden layer, denoted by P W , and the depth of the NN d, the FLOP complexity of an FC model is given by O ( d P T P W ) . Although the proposed methods introduce an additional computational cost of O ( P T 2 P T ) due to singular value extraction via Hankelization, this step is performed as part of offline preprocessing. As shown in Table 4, both variants of the proposed method—one using the original signal and the other using the FFT-processed signal—require only 5.72 × 10 3 FLOPs and 2.86 × 10 3 parameters, demonstrating significantly lower complexity compared to other baseline methods. In particular, CNN-based architectures such as SCNN2 and CLDNN require up to 1.42 × 10 8 and 7.89 × 10 6 FLOPs, respectively, with network sizes ranging from 1.73 × 10 5 to 3.50 × 10 4 parameters. The baseline methods follow the same NN configuration described in Section 3.1, ensuring a fair comparison in terms of network depth and width. The findings highlight that, although a mathematically complex preprocessing step is employed, the proposed method maintains high computational efficiency.

4. Conclusions

In this paper, we proposed a NN-based framework for AMC, incorporatingHankelization preprocessing. The proposed approach transforms the received signals into Hankelized matrices, from which singular values are extracted to form low-dimensional yet informative feature vectors. Two variants of the method were investigated—one based on time-domain signals and the other incorporating frequency-domain features via FFT. These features are then fed into an NN composed of a simple FC structure to perform classification across various SNR environments. The key strength of the proposed framework lies in its ability to effectively extract modulation-relevant features without relying on deep or complex NN architectures. Unlike conventional methods that depend on either handcrafted features or high-capacity models, our approach employs a concise design with minimal layers and moderate width, demonstrating that high performance can still be achieved when combined with appropriate signal representations.
Experimental results confirm that the FFT-enhanced variant consistently achieves the highest detection accuracy across a wide range of SNR levels. Notably, it performs reliably even under low-SNR conditions and with limited training data, maintaining stable detection rates. In addition, the model converges rapidly within a small number of training epochs. Latent space analysis further supports the ability of the model to form well-separated representations across modulation types. In summary, the proposed Hankelization-based AMC framework provides a practical and effective solution by integrating signal structure-aware preprocessing with a simple NN architecture. Future work may extend this framework to broader modulation families or more complex channel environments.

Author Contributions

Conceptualization, W.-H.L., J.-H.L. and O.-S.S.; software, J.-H.K.; writing—original draft preparation, J.-H.K.; writing—review and editing, W.-H.L., J.-H.L. and O.-S.S.; visualization, J.-H.K.; funding acquisition, J.-H.L. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the Dongguk University Research Fund of 2024 (S-2024-G0001-00025) and in part by the National Research Foundation of Korea (NRF) grant funded by the Korea government (MSIT) (No. RS-2023-00239349).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are available on request from the corresponding author. The data are not publicly available due to privacy.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
AMCAutomatic modulation classification
MLMachine learning
SVMSupport vector machine
k-NNk-nearest neighbors
NNNeural network
CNNConvolutional neural network
RNNRecurrent neural network
DLDeep learning
SNRSignal-to-noise ratio
SVDSingular value decomposition
SISOSingle-input single-output
CIRChannel impulse response
AWGNAdditive white Gaussian noise
FFTFast Fourier transform
t-SNEt-distributed stochastic neighbor embedding
FCFully connected
AdamAdaptive moment estimation
RMSPropRoot mean square propagation
MLPMultilayer perceptron
STFTShort-time Fourier transform
LSTMLong short-term memory
DAEDenoising autoencoder
FLOPFloating-point operation

References

  1. Gui, G.; Liu, M.; Tang, F.; Kato, N.; Adachi, F. 6G: Opening new horizons for integration of comfort, security, and intelligence. IEEE Wirel. Commun. 2020, 27, 126–132. [Google Scholar] [CrossRef]
  2. Dobre, O.A.; Abdi, A.; Bar-Ness, Y.; Su, W. Survey of automatic modulation classification techniques: Classical approaches and new trends. IET Commun. 2007, 1, 137–156. [Google Scholar] [CrossRef]
  3. Huynh-The, T.; Pham, Q.V.; Nguyen, T.V.; Nguyen, T.T.; Ruby, R.; Zeng, M.; Kim, D.S. Automatic modulation classification: A deep architecture survey. IEEE Access 2021, 9, 142950–142971. [Google Scholar] [CrossRef]
  4. O’Shea, T.J.; Roy, T.; Clancy, T.C. Over-the-air deep learning based radio signal classification. IEEE J. Sel. Top. Signal Process. 2018, 12, 168–179. [Google Scholar] [CrossRef]
  5. Häring, L.; Chen, Y.; Czylwik, A. Efficient modulation classification for adaptive wireless OFDM systems in TDD mode. In Proceedings of the 2010 IEEE Wireless Communications and Networking Conference, Sydney, Australia, 18–21 April 2010; pp. 1–6. [Google Scholar]
  6. Harper, C.A.; Thornton, M.A.; Larson, E.C. Automatic modulation classification with deep neural networks. Electronics 2023, 12, 3962. [Google Scholar] [CrossRef]
  7. Fu, X.; Gui, G.; Wang, Y.; Gacanin, H.; Adachi, F. Automatic modulation classification based on decentralized learning and ensemble learning. IEEE Trans. Veh. Technol. 2022, 71, 7942–7946. [Google Scholar] [CrossRef]
  8. Su, W.; Xu, J.L.; Zhou, M. Real-time modulation classification based on maximum likelihood. IEEE Commun. Lett. 2008, 12, 801–803. [Google Scholar] [CrossRef]
  9. Xu, J.L.; Su, W.; Zhou, M. Likelihood-ratio approaches to automatic modulation classification. IEEE Trans. Syst. Man Cybern. Part C (Appl. Rev.) 2010, 41, 455–469. [Google Scholar] [CrossRef]
  10. Ozdemir, O.; Li, R.; Varshney, P.K. Hybrid maximum likelihood modulation classification using multiple radios. IEEE Commun. Lett. 2013, 17, 1889–1892. [Google Scholar] [CrossRef]
  11. Li, J.; Meng, Q.; Zhang, G.; Sun, Y.; Qiu, L.; Ma, W. Automatic modulation classification using support vector machines and error correcting output codes. In Proceedings of the 2017 IEEE 2nd Information Technology, Networking, Electronic and Automation Control Conference (ITNEC), Chengdu, China, 15–17 December 2017; pp. 60–63. [Google Scholar]
  12. Orlic, V.D.; Dukic, M.L. Automatic modulation classification algorithm using higher-order cumulants under real-world channel conditions. IEEE Commun. Lett. 2009, 13, 917–919. [Google Scholar] [CrossRef]
  13. Su, W. Feature space analysis of modulation classification using very high-order statistics. IEEE Commun. Lett. 2013, 17, 1688–1691. [Google Scholar] [CrossRef]
  14. Zeng, Y.; Zhang, M.; Han, F.; Gong, Y.; Zhang, J. Spectrum analysis and convolutional neural network for automatic modulation recognition. IEEE Wirel. Commun. Lett. 2019, 8, 929–932. [Google Scholar] [CrossRef]
  15. Huynh-The, T.; Hua, C.H.; Pham, Q.V.; Kim, D.S. MCNet: An efficient CNN architecture for robust automatic modulation classification. IEEE Commun. Lett. 2020, 24, 811–815. [Google Scholar] [CrossRef]
  16. West, N.E.; O’shea, T. Deep architectures for modulation recognition. In Proceedings of the 2017 IEEE International Symposium on Dynamic Spectrum Access networks (DySPAN), Baltimore, MD, USA, 6–9 March 2017; pp. 1–6. [Google Scholar]
  17. Hamidi-Rad, S.; Jain, S. Mcformer: A transformer based deep neural network for automatic modulation classification. In Proceedings of the 2021 IEEE Global Communications Conference (GLOBECOM), Madrid, Spain, 7–11 December 2021; pp. 1–6. [Google Scholar]
  18. Chen, Y.; Dong, B.; Liu, C.; Xiong, W.; Li, S. Abandon locality: Frame-wise embedding aided transformer for automatic modulation recognition. IEEE Commun. Lett. 2022, 27, 327–331. [Google Scholar] [CrossRef]
  19. Mao, Q.; Hu, F.; Hao, Q. Deep learning for intelligent wireless networks: A comprehensive survey. IEEE Commun. Surv. Tutor. 2018, 20, 2595–2621. [Google Scholar] [CrossRef]
  20. Sathyanarayanan, V.; Gerstoft, P.; El Gamal, A. RML22: Realistic dataset generation for wireless modulation classification. IEEE Trans. Wirel. Commun. 2023, 22, 7663–7675. [Google Scholar] [CrossRef]
  21. Van der Maaten, L.; Hinton, G. Visualizing data using t-SNE. J. Mach. Learn. Res. 2008, 9, 2579–2605. [Google Scholar]
  22. Rousseeuw, P.J. Silhouettes: A graphical aid to the interpretation and validation of cluster analysis. J. Comput. Appl. Math. 1987, 20, 53–65. [Google Scholar] [CrossRef]
  23. O’shea, T.J.; West, N. Radio machine learning dataset generation with GNU radio. In Proceedings of the GNU Radio Conference, Charlotte, NC, USA, 20–24 September 2016; Volume 1. [Google Scholar]
  24. Tekbıyık, K.; Ekti, A.R.; Görçin, A.; Kurt, G.K.; Keçeci, C. Robust and fast automatic modulation classification with CNN under multipath fading channels. In Proceedings of the 2020 IEEE 91st Vehicular Technology Conference (VTC2020-Spring), Antwerp, Belgium, 25–28 May 2020; pp. 1–6. [Google Scholar]
  25. Ke, Z.; Vikalo, H. Real-time radio technology and modulation classification via an LSTM auto-encoder. IEEE Trans. Wirel. Commun. 2021, 21, 370–382. [Google Scholar] [CrossRef]
Figure 1. Data visualization based on t-SNE analysis. (a) Absolute values of signal. (b) Singular values of R t (with original signal). (c) Singular values of R f (with FFT-processed signal).
Figure 1. Data visualization based on t-SNE analysis. (a) Absolute values of signal. (b) Singular values of R t (with original signal). (c) Singular values of R f (with FFT-processed signal).
Applsci 15 07861 g001
Figure 2. Silhouette score according to (a) the number of singular values; (b) row size for the Hankelized matrix.
Figure 2. Silhouette score according to (a) the number of singular values; (b) row size for the Hankelized matrix.
Applsci 15 07861 g002
Figure 3. Overview of the architecture underlying the proposed framework.
Figure 3. Overview of the architecture underlying the proposed framework.
Applsci 15 07861 g003
Figure 4. Data visualization based on latent space (SNR = 10 dB). (a) Proposed method (with original signal). (b) Proposed method (with FFT-processed signal). (c) Deep learning (CNN with real and imaginary values of signal). (d) SCNN2. (e) CLDNN. (f) LSTMDAE.
Figure 4. Data visualization based on latent space (SNR = 10 dB). (a) Proposed method (with original signal). (b) Proposed method (with FFT-processed signal). (c) Deep learning (CNN with real and imaginary values of signal). (d) SCNN2. (e) CLDNN. (f) LSTMDAE.
Applsci 15 07861 g004
Figure 5. Confusion matrices (SNR = 10 dB). (a) Proposed method (with original signal). (b) Proposed method (with FFT-processed signal). (c) Deep learning (CNN with real and imaginary values of signal). (d) SCNN2. (e) CLDNN. (f) LSTMDAE.
Figure 5. Confusion matrices (SNR = 10 dB). (a) Proposed method (with original signal). (b) Proposed method (with FFT-processed signal). (c) Deep learning (CNN with real and imaginary values of signal). (d) SCNN2. (e) CLDNN. (f) LSTMDAE.
Applsci 15 07861 g005
Figure 6. Detection rate according to the number of training datasets. (a) SNR = 4 dB. (b) SNR = 10 dB. (c) SNR = 16 dB.
Figure 6. Detection rate according to the number of training datasets. (a) SNR = 4 dB. (b) SNR = 10 dB. (c) SNR = 16 dB.
Applsci 15 07861 g006
Figure 7. Test error according to epochs. (a) SNR = 4 dB. (b) SNR = 10 dB. (c) SNR = 16 dB.
Figure 7. Test error according to epochs. (a) SNR = 4 dB. (b) SNR = 10 dB. (c) SNR = 16 dB.
Applsci 15 07861 g007
Figure 8. Detection rate according to the number of singular values. (a) SNR = 4 dB. (b) SNR = 10 dB. (c) SNR = 16 dB.
Figure 8. Detection rate according to the number of singular values. (a) SNR = 4 dB. (b) SNR = 10 dB. (c) SNR = 16 dB.
Applsci 15 07861 g008
Figure 9. Detection rate according to row size for the Hankelized matrix. (a) SNR = 4 dB. (b) SNR = 10 dB. (c) SNR = 16 dB.
Figure 9. Detection rate according to row size for the Hankelized matrix. (a) SNR = 4 dB. (b) SNR = 10 dB. (c) SNR = 16 dB.
Applsci 15 07861 g009
Table 1. Settings of the RadioML2016.10a dataset [20,23].
Table 1. Settings of the RadioML2016.10a dataset [20,23].
ParameterValue
Channel modelRician fading
K-factor4 dB
Sampling rate T s 200 kHz
Number of multipaths L3
Doppler frequencymax = 1 Hz
Initial phase error θ e r r U(0, 2 π )
Frequency offset f e r r Standard deviation per sample = 0.01 , Maximum deviation: 500 Hz
Timing offset ζ e r r Standard deviation per sample = 0.01 , Maximum deviation: 500 Hz
Number of samples per modulation per SNR1000
Length of samples128
SNR 20 to 20 dB in steps of 2 dB
Table 2. Default configuration in the experiments.
Table 2. Default configuration in the experiments.
ParameterValue
Data split (per modulation class)70% training/10% validation/20% testing
NN depth d3
NN width P W 24
Epochs120
Mini-batch size64
NN connection typeFC
Learning rate 10 2
Activation functionHyperbolic tangent function
Loss functionCross-entropy
OptimizerAdam
β 1 , β 2 0.9 , 0.99
ϵ 10 8
Experimental setupPython 3.10.16, PyTorch 2.5.1, CUDA 12.1, cuDNN 9.1.0, NVDIA RTX A4000 GPU, Intel Core i9-11900K CPU
Table 3. Mean of detection rate ± standard deviation according to SNR.
Table 3. Mean of detection rate ± standard deviation according to SNR.
SNR [dB]Proposed Method (with Original Signal)Proposed Method (with FFT-Processed Signal)Deep Learning (CNN with Real and Imaginary Values of Signal)SCNN2CLDNNLSTMDAE
−20 0.2500 ± 0.0000 0.2500 ± 0.0000 0.2460 ± 0.0102 0.2500 ± 0.0196 0.2545 ± 0.01130.2490 ± 0.0123
−18 0.2500 ± 0.0000 0.2500 ± 0.0000 0.2428 ± 0.0149 0.2636 ± 0.0167 0.2529 ± 0.00990.2454 ± 0.0066
−16 0.3046 ± 0.0164 0.2504 ± 0.0011 0.2826 ± 0.0199 0.3299 ± 0.0198 0.3394 ± 0.01380.2476 ± 0.0054
−14 0.3881 ± 0.0146 0.3433 ± 0.0934 0.4354 ± 0.0181 0.4505 ± 0.0152 0.4627 ± 0.02820.2984 ± 0.0692
−12 0.5205 ± 0.0217 0.5630 ± 0.0279 0.5496 ± 0.0215 0.5693 ± 0.0103 0.5888 ± 0.01760.6123 ± 0.0145
−10 0.6560 ± 0.0097 0.6671 ± 0.0164 0.6389 ± 0.0112 0.6491 ± 0.0175 0.6834 ± 0.01760.6912 ± 0.0131
−8 0.7206 ± 0.0056 0.7036 ± 0.0211 0.6894 ± 0.0128 0.6915 ± 0.0208 0.7140 ± 0.00770.7251 ± 0.0078
−6 0 . 7394 ± 0.0083 0 . 7334 ± 0.0060 0.7184 ± 0.0138 0.7129 ± 0.0160 0.7292 ± 0.01460.7329 ± 0.0148
−4 0.7395 ± 0.0052 0.7380 ± 0.0038 0.7205 ± 0.0142 0.7336 ± 0.0092 0.7274 ± 0.00820.7409 ± 0.0138
−2 0.7346 ± 0.0040 0.7285 ± 0.0121 0.7125 ± 0.0264 0.7674 ± 0.0321 0.7264 ± 0.00980.7428 ± 0.0609
0 0.7406 ± 0.0137 0 . 8194 ± 0.0445 0.6911 ± 0.0217 0.8005 ± 0.0418 0.7297 ± 0.06550.7316 ± 0.0119
2 0.8102 ± 0.0317 0 . 9084 ± 0.0171 0.6986 ± 0.0277 0.8603 ± 0.0564 0.8229 ± 0.08430.8408 ± 0.0097
4 0.8085 ± 0.0379 0 . 9439 ± 0.0147 0.7368 ± 0.0180 0.9133 ± 0.0500 0.8396 ± 0.06140.8456 ± 0.0196
6 0.8569 ± 0.0346 0 . 9614 ± 0.0046 0.7458 ± 0.0261 0.9151 ± 0.0493 0.8306 ± 0.05600.8508 ± 0.0120
8 0.8641 ± 0.0217 0 . 9574 ± 0.0132 0.6789 ± 0.0256 0.8835 ± 0.0558 0.8316 ± 0.06070.8545 ± 0.0112
10 0.8870 ± 0.0395 0 . 9699 ± 0.0073 0.7286 ± 0.0240 0.9261 ± 0.0118 0.8482 ± 0.01250.8448 ± 0.0182
12 0.8899 ± 0.0165 0 . 9484 ± 0.0561 0.6784 ± 0.0234 0.9089 ± 0.0558 0.8341 ± 0.01590.8512 ± 0.0281
14 0.8698 ± 0.0214 0 . 9647 ± 0.0106 0.6959 ± 0.0157 0.8958 ± 0.0434 0.8336 ± 0.05350.8619 ± 0.0221
16 0.8656 ± 0.0282 0 . 9669 ± 0.0064 0.6790 ± 0.0252 0.9014 ± 0.0544 0.8326 ± 0.06760.8945 ± 0.0196
18 0.8473 ± 0.0536 0 . 9623 ± 0.0093 0.6876 ± 0.0226 0.9111 ± 0.0329 0.8403 ± 0.06870.8898 ± 0.0159
Table 4. Comparison of computational complexity in terms of FLOPs and the number of parameters.
Table 4. Comparison of computational complexity in terms of FLOPs and the number of parameters.
MethodFLOPs# Params
Proposed method (with original signal) 5.72 × 10 3 2.86 × 10 3
Proposed method (with FFT-processed signal) 5.72 × 10 3 2.86 × 10 3
Deep learning (CNN with real and imaginary values of signal) 2.04 × 10 6 1.58 × 10 5
SCNN2 1.42 × 10 8 1.73 × 10 5
CLDNN 7.89 × 10 6 3.50 × 10 4
LSTMDAE 3.53 × 10 6 1.48 × 10 4
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Kim, J.-H.; Lee, J.-H.; Shin, O.-S.; Lee, W.-H. A Novel Neural Network Framework for Automatic Modulation Classification via Hankelization-Based Signal Transformation. Appl. Sci. 2025, 15, 7861. https://doi.org/10.3390/app15147861

AMA Style

Kim J-H, Lee J-H, Shin O-S, Lee W-H. A Novel Neural Network Framework for Automatic Modulation Classification via Hankelization-Based Signal Transformation. Applied Sciences. 2025; 15(14):7861. https://doi.org/10.3390/app15147861

Chicago/Turabian Style

Kim, Jung-Hwan, Jong-Ho Lee, Oh-Soon Shin, and Woong-Hee Lee. 2025. "A Novel Neural Network Framework for Automatic Modulation Classification via Hankelization-Based Signal Transformation" Applied Sciences 15, no. 14: 7861. https://doi.org/10.3390/app15147861

APA Style

Kim, J.-H., Lee, J.-H., Shin, O.-S., & Lee, W.-H. (2025). A Novel Neural Network Framework for Automatic Modulation Classification via Hankelization-Based Signal Transformation. Applied Sciences, 15(14), 7861. https://doi.org/10.3390/app15147861

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop