Next Article in Journal
Effects of Sampling Frequency on Human Activity Recognition with Machine Learning Aiming at Clinical Applications
Previous Article in Journal
Quantitative Evaluation of Mechanical Properties of Hydrogen Transmission Pipelines Based on Weak Magnetic Detection
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Adversarial Sample Generation Method Based on Frequency Domain Transformation and Channel Awareness

Institute of Cyberspace Security, College of Information Engineering, Zhejiang University of Technology, Hangzhou 310023, China
*
Author to whom correspondence should be addressed.
Sensors 2025, 25(12), 3779; https://doi.org/10.3390/s25123779
Submission received: 29 April 2025 / Revised: 10 June 2025 / Accepted: 14 June 2025 / Published: 17 June 2025
(This article belongs to the Section Communications)

Abstract

:
In OFDM wireless communication systems, low-resolution channel characteristics and noise interference pose significant challenges to accurate channel estimation. To solve these problems, we propose a super-resolution denoising residual network (SDRNet), which combines the advantages of the super-resolution convolutional neural network (SRCNN) and the denoising convolutional neural network (DnCNN) to construct a pilot-based OFDM signal model, train SDRNet using OFDM pilot data containing Gaussian noise, and optimize its feature enhancement ability in frequency-selective fading channels. To further explore the role of channel estimation in communication security, we propose a frequency-domain adversarial attack method based on SDRNet output. This method first converts the time-domain signal to the frequency domain by using the Fourier transform and then applies Gaussian noise and selective masking. By integrating the channel gradient information, the adversarial perturbation we generated significantly improves the attack success rate compared with the non-channel awareness method. The experimental results show that SDRNet is superior to traditional algorithms (such as the least square method, minimum mean square error estimation, etc.) in both mean square error and bit error rate. Furthermore, the adversarial samples optimized through channel awareness frequency-domain masking exhibit stronger attack performance, confirming that accurate channel estimation can not only enhance communication reliability but also provide key guidance for adversarial perturbation. The experimental results show that under the same noise conditions, the MSE of SDRNet is significantly lower than that of LS and MMSE. The bit error rate is lower than 0.01 when the signal-to-noise ratio is 10 dB, which is significantly better than the traditional algorithm. The attack success rate of the proposed adversarial attack method reached 79.9%, which was 16.3% higher than that of the non-channel aware method, verifying the key role of accurate channel estimation in enhancing the effectiveness of the attack.

1. Introduction

In the field of communication, channel estimation is a key link to ensure the reliable transmission of signals. However, traditional channel estimation methods have certain limitations when facing complex environments or scenarios with channel sparsity. Although the application of deep learning has brought about some progress, there are still flaws in the network structure and data form. Exploring more accurate and effective channel estimation methods has become an important research direction at present. In today’s technological field, deep learning has been widely applied, yet adversarial attacks have become a major obstacle to its development. The core of adversarial attacks lies in inputting deceptive data into machine learning models, resulting in misclassification of the models. Deep learning adversarial attacks originated in the field of images [1] and then gradually expanded to multiple other fields. During the implementation of adversarial attacks, attackers will perturb the original samples based on their in-depth understanding of the model, such as neural network architecture, learning parameters, loss functions, etc., thereby maximizing the loss function of the classifier and ultimately leading to classification errors. With the advancement of research, adversarial attack algorithms continue to evolve. On the one hand, attackers begin to focus on frequency-domain information because different frequency components have a non-negligible impact on the judgment of the model. By using techniques such as frequency-domain adversarial generative networks [2], spectral attacks [3], and frequency-domain adversarial transport [4], the model can be ingeniously deceived at the frequency-domain level, thereby improving the success rate of attacks. Frequency-domain analysis also helps to deeply understand the model’s response to different frequency components, providing strong guidance for the generation of adversarial samples. This is of great significance for improving the defense mechanism and enhancing the robustness of the model.
In the generation of adversarial samples based on frequency-domain transformation considering channel information, if channel estimation is lacking, the generated adversarial samples are difficult to adapt to the characteristics of the actual communication channel. Due to the failure to consider factors such as multipath propagation, fading, and frequency offset in the channel, in the wireless communication scenario, the adversarial samples may undergo severe distortion after being transmitted through the channel, which greatly reduces the success rate of the attack and makes it impossible to effectively evaluate the security of the model in the real channel environment. Channel estimation can accurately describe the channel characteristics, providing a key basis for the generation of adversarial samples in frequency-domain transformation. This enables the generated samples to optimize the perturbation addition strategy based on the channel conditions, ensuring that they remain deceptive after being transmitted through the channel. Therefore, it is extremely urgent to explore new channel estimation methods, improve the estimation accuracy, and verify their importance in the generation of adversarial samples. This paper proposes an innovative channel estimation method for these problems and verifies it with the aid of adversarial sample generation based on frequency-domain transformation. The specific contributions are as follows:
  • A channel estimation method based on deep learning is proposed. This method fully exploits the potential of deep learning in data processing and feature extraction and is committed to achieving high-precision estimation of sparse channels in OFDM systems.
  • Based on the above channel estimation methods, a channel-aware adversarial sample generation method based on frequency-domain transformation is further proposed to verify the importance of SDRNet channel estimation and significantly improve the effect of adversarial sample generation.
  • The experimental results show that SDRNet significantly outperforms other traditional algorithms in terms of accuracy and robustness in the channel estimation task. Meanwhile, the proposed adversarial attack method also demonstrates a higher attack success rate.

2. Background Introduction

Channel estimation is essential for ensuring communication quality, with traditional OFDM-based methods exhibiting distinct advantages and limitations. The least squares (LSs) method [5,6], widely used for its computational simplicity, suffers from severe performance degradation in low-SNR scenarios due to its disregard for noise. Discrete Fourier transform (DFT)-based techniques [7] effectively suppress noise through frequency-domain filtering while maintaining low complexity; however, their performance rapidly deteriorates when frequency offsets exceed 5% of the subcarrier spacing. The linear minimum mean square error (LMMSE) method [8,9], though capable of achieving near-optimal estimation under moderate-to-high SNRs, depends on prior knowledge of channel covariance and exhibits cubic computational complexity, limiting its scalability in large-scale MIMO systems.
Recent developments in deep learning have introduced data-driven alternatives. Deep neural networks (DNNs) [10,11,12] demonstrate strong performance across varying pilot lengths through end-to-end learning, while convolutional neural networks (CNNs) [13] utilize pilot position information to enhance estimation accuracy. Advanced frameworks such as CsiNet and CsiNet-LSTM [14] improve the robustness of CSI feedback, and tensor-train DNNs (TT-DNNs) [15] reduce parameter dimensionality for high-dimensional CSI, albeit with slower convergence. Hybrid approaches that integrate traditional techniques with DNNs [13,16] seek to balance complexity and adaptability, for instance, by leveraging spectral-time averaging to track channel variations.
For sparse channel estimation, traditional algorithms such as orthogonal matching pursuit (OMP) [17] remain prevalent, though their effectiveness heavily relies on accurate sparsity level estimation. Enhanced OMP variants [14,15,18,19] exploit angular sparsity or introduce adaptive mechanisms to improve robustness, often at the cost of increased complexity. In contrast, deep learning-based methods, including CNN-based MIMO-OFDM estimators [20] and DNN-enhanced OTFS systems [21,22] have shown superior performance over OMP, particularly in delay-Doppler domains.
Channel estimation guarantees communication quality and is related to the accuracy and stability of signal transmission. Modulation classification, as the basis of communication systems, is a key prerequisite for subsequent signal processing and information interpretation. The two are closely related and jointly promote the development of communication technology. Under this broad background, the security and stability of communication systems have become key research directions, and adversarial attacks and channel estimation are precisely the core issues among them. In the field of adversarial attacks, many research achievements have been remarkable. In the early days, after Szegedy et al. [23] discovered that deep neural networks were vulnerable to slight adversarial interference, Wu et al. [24,25] proposed the adversarial transformation-enhanced transfer Attack (ATTA), which constructs an adversarial transformation network through adversarial learning to generate adversarial noise and thereby resist the distortion problem caused by the network.
The high-frequency component semantic similarity attack proposed by Luo et al. [26] focuses on the high-frequency noise of the image. Chen et al.’s [27] adversarial attack method based on adversarial generative networks (GANs), adGAN, reduces the performance of intelligent systems by leveraging adversarial generative networks. Xu et al. [28] studied the perturbation pattern analysis of radio frequency signals. Wang et al. [29] utilized Hamiltonian Monte Carlo to generate a series of adversarial samples.
With the in-depth research on the security of communication systems, the correlation between modulation classification and adversarial attacks has gradually emerged. In the context of modulation classification, Zhang et al. [30] evaluated the performance and adversarial sensitivity of Transformer-based neural networks. Manoj et al. [31] introduced multiple training methods to construct robust DNN models and evaluate them. Kotak and Elovicii [32] applied attack assessment to evaluate the vulnerability of the Internet of Things device identification system and discovered new attack methods.
As an important application scenario of communication systems, wireless communication has concrete manifestations and further development in which the above research results are presented. In wireless communication, Sadeghi et al. [33] proposed an adversarial attack method for automatic modulation identification. Lin et al. [34] explored its threats and impacts on automatic modulation recognition. Sandler et al. [35] verified the effectiveness of the attack in the form of external interference. Cohen et al. [36] improved robustness by increasing the noisy training data. Kim et al. [37] studied the channel influence and proposed a channel-aware attack method. Meanwhile, frequency-domain attacks emerged. Guo et al. [38] utilized the low-frequency component attack algorithm, and Sharma et al. [39] discovered different responses of the defense model to high- and low-frequency disturbances. Duan et al. [40] proposed the adversarial attack on DNNs by dropping information (AdvDrop) attack on neural networks.
Despite these advancements, significant challenges remain. Traditional methods struggle in complex and dynamic environments, sparse estimators falter when ideal sparsity assumptions are violated, and deep learning-based approaches still require improved generalization and robustness. To address these issues, we propose a novel channel estimation framework that enhances both accuracy and robustness. Additionally, we introduce a frequency-domain adversarial sample generation method that leverages channel state information to assess the critical role of accurate estimation in communication security. This work bridges the gap between performance and robustness, contributing to the development of reliable and secure wireless communication systems.

3. Related Work

3.1. Mathematical Model of OFDM Channel Estimation System

The channel estimation system model consists of two parts: the transmitting end and the receiving end. At the transmitting end, the input bit stream is modulated and mapped onto the OFDM subcarriers, and pilot symbols are inserted at specific subcarrier positions for subsequent channel estimation. Then, the frequency-domain signal is converted into a time-domain signal through the inverse fast Fourier transform (IFFT), and a cyclic prefix is inserted. CP is used to resist Inter-symbol interference (ISI), and finally the signal is transmitted through the additive white Gaussian noise (AWGN) channel. At the receiving end, the signal successively removes the cyclic prefix, undergoes fast Fourier transform (FFT) to convert back to the frequency domain, extracts the pilot symbol, and uses the pilot signal for channel estimation, thereby obtaining the channel state information (CSI). Subsequently, the channel influence is eliminated through frequency-domain equalization, the amplitude and phase of the signal are restored, and then the original bit stream is restored through subcarrier demapping and demodulation. The specific process is shown in Figure 1, which shows the training data sources of SDRNet (such as OFDM pilot signals) and the processing flow.
According to the above system model block diagram, we can further establish the above signal processing flow as a mathematical model as follows. Suppose the total number of subcarriers in an OFDM system is N, and a subframe contains I OFDM symbols in total. For the ith OFDM symbol, suppose the nth subcarrier symbol on the ith symbol is s i ( n ) , and there is s i = [ s i ( 0 ) , , s i ( n ) , , s i ( N 1 ) ] T . After OFDM modulation of the frequency-domain symbol through discrete Fourier transform (IFFT):
S i = F H s i
where S i C N × 1 represents the transmitted time-domain sequence and F H is the Fourier transform matrix, specifically expressed as:
[ F ] n , k = 1 N exp ( j 2 π N k n )
The OFDM transmission model can be further constructed as follows:
y i = H i s i + z i
where the frequency-domain symbol vector received on the ith OFDM symbol block is y i = [ y i ( 0 ) , , y i ( n ) , , y i ( N 1 ) ] T , z i is the additive white Gaussian noise of the channel, the covariance matrix is Q z = σ 2 I N , σ 2 is the noise variance, I N represents the N × N dimensional identity matrix, and H i C N × N represents the frequency-domain response matrix of the channel on:
H i = F G i F H
where G i C N × N represents the impulse response matrix of the ith symbolic time channel, expressed as:
G i = h i ( 0 , 0 ) 0 h i ( 0 , L 1 ) h i ( 0 , 1 ) h i ( 1 , 1 ) h i ( 1 , 0 ) 0 h i ( 1 , 2 ) 0 0 h i ( N 1 , L 1 ) h i ( N 1 , 0 )
where h i ( k , l ) represents the kth sampling point of the channel impulse response at the lth tap at the ith symbol time. Equation (3) can be considered as the mathematical model of baseband signal transmission in the OFDM system, and H i is the parameter to be estimated in the channel estimation link.
For the channel estimation method of OFDM systems, the first step is to estimate the channel response at the pilot symbol position, and then further obtain the channel response on the complete resource block through time-domain interpolation and frequency-domain interpolation. In this paper, it is assumed that the channel impulse response is invariant within an OFDM symbol time, that is, if h i ( k , l ) in G i satisfies that the sampling points are equal at different times, then it can be ensured that:
h i ( l ) = h i ( 0 , l ) = = h i ( N 1 , l ) = 1 N k = 0 N 1 h i ( k , l )
where h i ( l ) represents the channel response on the lth tap within the time of the Ith OFDM symbol. Based on this assumption, the channel impulse response matrix G i becomes:
G i = h i ( 0 ) 0 h i ( L 1 ) h i ( 1 ) h i ( 1 ) h i ( 0 ) 0 h i ( 2 ) 0 0 h i ( L 1 ) h i ( 0 )
where G i is obtained by continuously rotating the channel impulse response h i ( l ) on different taps. According to the simulation of the properties of the channel frequency-domain response matrix in the previous section and the properties of the discrete Fourier transform, it can be seen that the channel frequency-domain response matrix will become a diagonal matrix as follows:
H i = H i ( 0 ) 0 0 0 H i ( 0 ) 0 0 0 0 0 H i ( N 1 )
where H i represents the channel response of the channel in the ith OFDM symbol time. Further, we simplify Equation (3) based on this assumption, with:
y i = X i h i + z i
X i = d i a g ( s i ) = s i ( 0 ) 0 0 0 s i ( 0 ) 0 0 0 0 0 s i ( N 1 )
where h i = v e c ( H i ) = [ H i ( 0 ) , H i ( 0 ) , , H i ( N 1 ) ] T is the channel impulse response vector; as shown in the formula, X i represents the channel frequency-domain transmission symbol matrix; function v e c ( · ) represents the operation of extracting the diagonal elements of a matrix and forming a column vector; function d i a g ( · ) represents the operation of constructing a diagonal matrix with vector E as the main diagonal element.

3.2. Adversarial Disturbance Generation Method

The purpose of adversarial attacks is to add adversarial perturbations to the original benign samples to maximize the loss function value of the classifier and interfere with the classifier’s incorrect judgment. For the input sample x and the true label y, let the classification model f be able to correctly identify sample x, then f ( x ) = y . The goal of the adversarial attack is to generate a subtle adversarial perturbation δ to form an adversarial sample x = x + δ , which can enable the classification model to make a wrong judgment, namely f ( x ) y . Attackers usually formulate various strategies for generating perturbations based on the prior knowledge of deep learning models, such as neural network structures, model weights, and model loss functions. For example, FGSM generates adversarial perturbations based on the gradient of the loss function. The adversarial samples are as follows:
x = x + α · s i g n ( x J ( x , y ) )
where α is the perturbation step size, the adversarial sample x x ε , x + ε , J represents the loss function, and y represents an arbitrary label outside the true sample.
During the attack process, adversarial perturbation ε , also known as perturbation budget, represents the maximum extent of changes that the attacker is allowed to make to benign samples, and δ represents adversarial perturbation. The perturbation budget keeps the perturbations generated by attackers within a displayable and controllable range. Under normal circumstances, after the adversarial sample x is generated, it should exist within the L p norm sphere of the original sample x. Commonly used p values include 0, 1, 2, and . In this paper, the focus will be placed on the L 2 norm, that is, the power limit of the signal. To generate more effective adversarial samples, attackers usually operate with more complex algorithms. For instance, the PGD algorithm gradually adds perturbations to the original data until it successfully misleads the model. The adversarial sample optimization process can be expressed as:
x t + 1 = c l i p x , ε { x t + α · s i g n ( x t J ( x t , y ) ) }
where c l i p represents the clipping function and J ( · ) represents the cross-entropy loss.

4. Channel Estimation Method Based on Deep Learning

In this section, firstly, the signal generation, processing and the traditional OMP algorithm are described, and then the network super-resolution denoising network (SRCNN DNCNN network, SDNet) after model fusion is introduced in detail, and the optimized super-resolution denoising residual network (SDRNet) on this basis is introduced. Finally, the generation of the channel-aware adversarial sample algorithm based on frequency-domain transformation was elaborated in detail, verifying the importance and accuracy of SDRNet channel estimation. As shown in Figure 2, it represents the training process of SDRNet, and Figure 3 represents the generation process of adversarial samples.

4.1. Residual Channel Estimation Framework for Super-Resolution Denoising Based on Deep Learning

In the SDR channel estimation framework, data preparation should be carried out first and the collected data should be preprocessed. Then, the OMP algorithm is used to estimate the signal, followed by model fusion to obtain the score. Finally, the signal is estimated by integrating the above results to accurately obtain the channel information and achieve efficient and accurate channel estimation.

4.1.1. Date Preparation

In the channel estimation task, in order to simulate the diversified channel environment in a real communication scenario and improve the generalization of the model, we generated signals with different signal-to-noise ratios (SNRs). First, an original signal is generated at the transmitter, which can be represented as a complex sequence:
x ( t ) = x r e a l ( t ) + j x i m a g ( t )
where x r e a l ( t ) and x i m a g ( t ) represent the real and imaginary components of the signal. Subsequently, we transmit this signal through the channel and receive a signal with added noise at the receiving end. To simulate various SNRs, Gaussian white noise of varying intensity is added to the received signal for each ratio. Specifically, for each SNR, we add noise and generate the corresponding received signal:
y i ( t ) = x ( t ) + z i ( t )
where z i represents Gaussian white noise. y i represents the signal received by the receiver.
Next, the generated signals are longitudinally concatenated according to different signal-to-noise ratios to construct a comprehensive training dataset containing multiple channel environments. Suppose N signal segments with different signal-to-noise ratios are generated. The operation of vertical splicing can be expressed as:
Y = y 1 ( t ) y 2 ( t ) y N ( t )
In this way, a matrix containing signal segments with different signal-to-noise ratios was generated, which can simulate the diverse channel conditions in the actual communication environment more comprehensively. In channel estimation, the orthogonal matching pursuit (OMP) algorithm is widely used as an effective sparse signal recovery technique. Therefore, the OMP estimator will be used next to estimate the generated and processed signals. It is applicable to the estimation problem of sparse signals, with low computational complexity and good noise robustness. This algorithm adopts the idea of the greedy algorithm. Essentially, it selects the atoms with the greatest correlation to the signal from a group of atoms called the atomic set. Next, it uses the known atoms to construct estimation coefficients, gradually reduces the residuals, and obtains a sparse signal representation. This algorithm is easy to understand and can accurately restore high-dimensional sparse information.
First, the residual r is initialized, and the atom with the highest correlation to the residual is selected from the set of atoms:
λ k = arg max i = 1 , , N r k , A ( i ) , k = 0 , 1 , , K
where λ k is the kth column index determined by the maximum absolute value of the correlation; r 0 = b represents the residual of the first iteration; A ( i ) is the ith term of A, representing the estimated value of the channel matrix.
Next, the column index with the highest correlation is added to the index set Λ k , and the data corresponding to these indicators in the observation matrix is updated to the reconstructed set of atoms A Λ k :
Λ k = Λ k 1 λ k
A Λ k = A Λ k 1 A ( Λ k )
Using the LS method, we obtain an approximate solution:
x k = arg min x A Λ k x b 2 = A Λ k 1 b
Calculate the new approximation of the data and the new residual:
r k = b b k
b k = A x k
This algorithm selects the vector that best matches the current residual through repeated iterations, and performs orthokerization processing on the selected vectors at each step to gradually construct an approximate solution. When the residuals are small enough or reach the preset sparsity, the iteration terminates. The final solution is a linear combination of all the previously selected vectors. This process employs a greedy strategy to ensure that each step of the choice can minimize the residuals to the greatest extent. It is worth noting that the residuals are always orthogonal to the column space of the selected vectors. When the residuals are not zero and the matrix has a full column rank, the solution of the least square method is unique, thereby ensuring the stability of the algorithm.

4.1.2. Model Integration

In the task of channel estimation, the time-frequency domain characteristics of the channel are complex and changeable, and are vulnerable to noise interference, resulting in limited estimation accuracy. In the field of image processing, super-resolution technology can restore high-resolution details from low-resolution data, and denoising technology can effectively remove noise from images. Since the signal data structure in channel estimation is similar to the image data in features such as multi-channel and two-dimensional distribution, drawing on the SRCNN and DNCNN networks in image processing, the two are integrated and applied to channel estimation. By taking advantage of their abilities to extract details and reduce noise, the accuracy and reliability of channel estimation are improved.
The purpose of SRCNN is to improve the spatial resolution of images. The core idea of SRCNN is to learn the mapping from low-resolution images to high-resolution images through convolutional neural networks. SRCNN consists of three convolutional layers:
Y s r c n n = f 3 ( w 3 × f 2 ( w 2 × f 1 ( w 1 × X ) ) )
where X is the input signal, Y s r c n n is the output signal, w i represents the convolutional kernel of layer i, f i represents the activation function.
SRCNN achieves super-resolution through three layers of convolution. Feature extraction layer f 1 : Map the low-resolution input to the high-dimensional feature space to capture the macroscopic characteristics of the channel response. Nonlinear mapping layer f 2 : Enhance the feature expression ability and restore the high-frequency details lost due to the multipath effect. Reconstruction layer f 3 : Synthesize high-resolution output, whose output dimension matches the input dimension of DNCNN.
The purpose of DNCNN is to remove noise from images, and its core idea is to reduce noise through residual learning. DNCNN consists of multiple convolutional layers and batch normalization layers:
Y d n c n n = X F ( X ; Θ )
where X is the input signal, Y d n c n n is the output signal, F ( X ; Θ ) is the convolutional network part of DNCNN, which is capable of extracting the mapping relationship of noise.
In order to integrate these two networks and apply them to channel estimation, this paper designs a model (SDNet) that combines the advantages of both. First, the real and imaginary parts of the complex signal are regarded as two channels of the image, respectively, to construct the input data. Next, the input data is sent to SRCNN to improve the resolution of the signal and obtain the preliminary estimated signal. Finally, the estimated signal output by SRCNN is sent to DNCNN to further remove the noise in the signal and obtain the final accurate channel estimation result. Channel estimation requires restoring the frequency domain details (super-resolution) first and then suppressing the noise (denoising), which is consistent with the image processing flow. The real and imaginary parts of the complex signal are used as dual-channel inputs, retaining the joint phase-amplitude information and enabling the image processing method to be transferred to channel estimation. The key to model fusion lies in the reasonable connection of the two networks, so that the output of the former network can be seamlessly used as the input of the latter network.

4.2. The Super-Resolution Denoising Network Model Based on OMP Algorithm

In the field of signal processing, SRCNN and DNCNN are integrated into the super-resolution denoising network (SDNet) model, aiming to give full play to the super-resolution advantage of SRCNN and the efficient denoising ability of DNCNN, thereby simultaneously improving the accuracy and robustness of channel estimation. Based on this, the SDNet channel estimation framework based on the OMP algorithm was further developed. This framework consists of two stages: discrete training and online estimation. The entire network structure proposed is shown in Figure 4. In the offline training stage, after obtaining the channel estimation result of the OMP algorithm, it is input as the training set into the neural network, and the standard deep neural network training process is used to obtain a well-trained model. Subsequently, in the online estimation stage, the test data is input into the well-trained SDNet to obtain the estimated channel coefficients. Although Figure 5 shows a general process, the proposed algorithm embeds a unique design in each step. In the training network of offline training, we used the specially designed OMP algorithm combined with the deep learning channel estimation algorithm to complete the channel estimation task. During the online testing stage, well-trained networks apply SRCNN and DNCNN to improve the accuracy of channel estimation in our specific scenarios. The overall process is shown in Figure 5 as follows:

4.2.1. Offline Training

The architecture of the SDNet framework consists of the following parts.
Training data generation: prepare a large amount of labeled training data represented as:
( x , y ) = ( x ( 1 ) , y ( 1 ) ) , , ( x ( N ) , y ( N ) )
where N represents the number of training samples, each ( x ( i ) , y ( i ) ) represents the ith ( i 1 , 2 , , N ) training samples in the dataset, where x ( i ) represents the input data (features) and y ( i ) represents the output data used to train the neural network.
According to Equation (3), for the ith user, the input data should be the sent signal S i and set the corresponding output as the channel impulse response H i o m p . Since the inputs and outputs of the neural network must be real numbers, we need to convert the complex-valued matrix into a real-valued matrix.
For the input S i C m × 1 , we reshape it into a 2D matrix S i C m 1 × m 2 , where m = m 1 × m 2 , and both m 1 and m 2 are integers. Then, by superimposing the real part and the imaginary part into two channels, the input of OMP-SDNet can be represented as:
x ( i ) = F { [ Re ( S i ) , Im ( S i ) ] } R m 1 × m 2 × 2
where F ( · , · ) represents conversion to a real function. For the label X r e a l i , the real part and the imaginary part are superimposed in turn and reconstructed into a vector with a real value of 2 m , which can be expressed as:
y ( i ) = Re H i o m p , Im H i o m p R 2 m × 1
Network architecture SRCNN→DnCNN→FC: the proposed network is shown in Figure 5. Compared with the classical convolutional layer, the network has better performance in extracting channel features and de-noising. The network architecture adopts a three-level series structure of SRCNN→DnCNN→FC. First, through the three-layer convolution of SRCNN (Conv1: 3 × 3 × 64, ReLU; Conv2: 3 × 3 × 128, ReLU; Conv3: 3 × 3 × 256, ReLU) to achieve super-resolution reconstruction of channel features and extract high-frequency details; subsequently, the five-layer residual module of DnCNN (each module contains 3 × 3 convolution, batch normalization and ReLU activation) is connected to suppress the noise layer by layer and retain the effective signal. Finally, the features are mapped to the channel estimates through the fully connected layer (FC), and the dual-channel results of the real part and the imaginary part are output.
First comes the input layer x ( i ) , which is responsible for receiving and carrying the input data. The second, third, and fourth layers immediately following are convolutional layers, each equipped with a different number of filters with varying sizes of convolution kernels. After each convolutional layer, a rectification linear unit activation function is connected to perform initial feature extraction and transformation of the data. At this point, the obtained intermediate output is saved. Subsequently, the data flows through multiple convolutional layers of size (3 × 3) with the same fill settings. Each Conv layer here is successively connected to a batch normalization layer and a rectification linear unit activation function. Through this series of processing, the noisy image is obtained. Ultimately, subtract the resulting noisy image from the previously saved output, and the difference obtained is the final output result of the entire network. It is worth emphasizing that each connection in the above architecture corresponds to a specific weight. As shown in Figure 6, the output of the entire network can be expressed by the following formula:
y S D N e t ( i ) = f ( θ ; x ( i ) )
where θ represents the weight that plays a key role in determining the performance and performance of the model, and f ( · ) represents the forward propagation function of the whole neural network.
Training network: we will describe in detail how to use the training set to train the proposed OMP-SDNet framework.
First, the weight θ of each layer needs to be randomly initialized. The purpose of this step is to ensure that the network has a different initial state at the beginning of training so that it can learn different feature representations.
Then it enters the forward propagation stage. During this process, each layer of the neural network processes the input data x ( i ) in sequence. Firstly, after three convolution and linear activation operations, the purpose is to accurately extract image blocks from the input low-resolution image and map them to the high-resolution space. Then, these high-resolution image blocks are aggregated and integrated to generate the final high-resolution image. The data then passes successively through the module composed of multiple layers of convolution, batch normalization layers, and rectification linear unit activation functions. This module can effectively extract the features contained in the image, and simultaneously complete the tasks of noise removal and image restoration. The output generated by each layer in the above process will be used as the input for the next layer, thus enabling the information to be smoothly transmitted layer by layer until it reaches the output layer. Ultimately, the output result of the network can be represented by f ( θ ; x ( i ) ) .
The loss calculation: after each step of forward propagation, we calculate the loss between the predicted output f ( θ ; x ( i ) ) and the actual output y ( i ) . The goal of training is to minimize losses by adjusting network parameters. In this paper, we use MSE to define the loss function of the network:
M S E = 1 N i = 1 N ( y e s t i y r e a l ) 2
where y e s t i represents the predicted value, y r e a l represents the actual value, and N represents the sample size. Our training goal is to find a well-trained set to θ ^ to minimize the loss function, so the optimization problem in this paper can be expressed as:
min θ L S D N e t ( θ ) = 1 N i = 1 N f ( θ ; x ( i ) ) y ( i ) 2
In neural network training, we use backpropagation to calculate the loss gradients and apply the Adam optimization algorithm to update the network parameters according to these gradients. When training, the data are divided into small batches and processed in multiple cycles. Each cycle, the model predicts by forward propagation, calculates MSE losses, backpropagation acquires gradients, and updates parameters to minimize losses. After the training is completed, a model with optimal parameters can be obtained, which can accurately predict new data. The final trained model can be expressed as:
y S D N e t = f ( θ ^ ; x )
where x represents the input data and y S D N e t represents the output data, which needs to be converted into complex value channel information H S D N e t . We can express the estimated channel matrix H S D N e t as:
H S D N e t = y S D N e t X e s t i

4.2.2. Online Estimation

In the online estimation stage of channel estimation, the optimal parameter model obtained after offline training of the neural network will be deployed to the wireless communication system. In the process of online estimation, the system will first capture the real-time received signal and convert it into a real-valued matrix suitable for neural network processing according to the formula. These real-valued matrices are then fed into a trained neural network, SDNet, and the predicted channel information is obtained through forward propagation.
However, with the increase in network depth in the OMP-SDNet, the problem of over-fitting or gradient disappearance may be encountered, resulting in a poor estimation effect. Therefore, in order to further improve the accuracy of channel estimation, we will propose a neural network architecture based on ResNet in the next section.

4.3. The Super-Resolution Denoising Residual Network Model Based on OMP Algorithm

In order to further improve the accuracy of channel estimation, a channel estimation framework was developed. This framework integrates the residual network (Resnet) module based on SDNet and is denoted as SDRNet.
Here, the ResNet module is integrated into the channel estimation framework. Adding residual connections in the part after three convolutions can help the network better learn the residuals to solve the problem of vanishing gradients and thereby improve the denoising performance.
The improved channel estimation framework OMP-SDRNet also includes two stages: (1) offline training and (2) online estimation. The entire workflow is shown in Figure 6. Similar to the SDNet framework, the training data generation method of SDRNet is the same, but their network architectures are different. The ResBlock component of the super-resolution denoising residual network (SDRNet) is a key architectural innovation. This section focuses on the design, functions, and impacts of ResBlocks, which are integrated to solve the vanishing gradient problem in deep networks and enhance the channel estimation performance in OFDM systems. The ResBlock in Figure 6 consists of a skip connection and two convolutional sub-layers, aiming to facilitate deep learning by enabling the network to learn residual mappings rather than direct input–output relationships. The detailed structure is as follows. Convolutional layer: Two consecutive 3 × 3 convolutional layers, filled to maintain the dimension of the feature map. After each convolutional layer, there is batch normalization (BN) to standardize activation and accelerate training, and there is also a ReLU activation function to introduce nonlinearity. Skip the connection: Directly connect the input of the block to the output of the convolutional layer, allowing the network to learn the residual between the input and the expected output. This design ensures that the network can learn incremental improvements (residuals) instead of reconstructing the entire signal from scratch, reducing gradient vanishing. In Figure 6, after the initial feature extraction layer based on srcnn, multiple Resblocks are cascaded in the denoising module of SDRNet. The workflow is as follows. Feature extraction: The SRCNN module (the first few layers) processes the OFDM pilot signal to restore the high-resolution channel features. Residual denoising: Cascades ResBlocks receive features from SRCNN and gradually suppress noise by learning residuals (i.e., the difference between noisy features and clean channel responses). Channel reconstruction: The output of ResBlocks is fed into the final convolutional layer to generate an estimated channel matrix, leveraging denoising and enhanced features. The skip connections in ResBlocks bypass the deep convolutional layers, ensuring that the gradients propagate effectively through the network. Compared with the non-residual SDNet that lacks such connections, this allows SDRNet to train deeper architectures. Noise suppression: By focusing on residual learning, SDRNet more effectively separates noise from the effective channel features. For instance, in low signal-to-noise ratio scenarios, ResBlocks can distinguish between multipath fading (expected signal) and Gaussian noise (residual), thereby generating clearer estimates. At this time, the output result can be expressed as:
x ( j ) = F c o n v [ x ( i ) w ( i ) ]
The improved channel estimation framework OMP-SDRNet also includes two phases: (1) offline training and (2) online estimation. The whole workflow is shown in Figure 4. where F c o n v · represents the convolution operation; x ( i ) indicates the input training data; w ( i ) represents the set of parameters of the three convolutions operation processes.
Next, the results after three convolutions and linear activation are used as the input of the first Residual block. The basic structure of the residual block consists of two parts: direct mapping and residual mapping, and the output of each block can be expressed as:
y ( j ) = F x ( j ) w ( j ) + x ( j )
where j represents the number of residual blocks, x ( j ) and y ( j ) are the input and output of the residual blocks, and w ( j ) is the set of parameters for these operations.
Residual mapping F x ( i ) w ( i ) usually consists of two or three convolution layers, each of which is followed by batch normalization (BN) and ReLU activation functions.
The final output result can be expressed as:
y S D R N e t = y ( j )
where j represents the value of the last residual block, which needs to be converted into complex-valued channel information H S D R N e t . We can express the estimated channel matrix H S D R N e t as:
H S D R N e t = y S D R N e t X e s t i
Through adequate training, SDRNet successfully introduced the concept of residual learning into the task of channel estimation. The structure of the residual block allows the network to learn the residual representation between the input and output, thereby simplifying the learning difficulty and avoiding the problem of vanishing gradients. The integration method adopted in this paper can significantly improve the denoising performance of neural networks. The loss calculation and backpropagation sections are similar to the SDNet framework proposed in the previous section and will not be elaborated on in detail here.

5. Channel Aware Adversarial Sample Generation Method Based on Frequency Domain Transformation

Channel estimation guarantees communication quality and is related to the accuracy and stability of signal transmission. Modulation classification, as the basis of communication systems, is a key prerequisite for subsequent signal processing and information interpretation. The two are closely related and jointly promote the development of communication technology. Under this broad background, the security and stability of communication systems have become key research directions, and adversarial attacks and channel estimation are precisely the core issues among them.
In recent years, the research on adversarial attacks based on frequency-domain information has revealed the frequency-domain sensitivity characteristics of deep learning from two dimensions: attack methods and model mechanisms. At the level of attack methods, researchers have found that low-frequency components and high-frequency components have the same influence on model decision-making. Although the existing defense mechanisms can effectively suppress high-frequency disturbances, there are still significant loopholes in the defense against low-frequency disturbances.
At the model mechanism level, research shows that deep neural networks have significant frequency-domain perception preferences. On the one hand, the model can capture high-frequency features that are difficult for humans to detect, but is extremely sensitive to high-frequency noise. On the other hand, network decision-making overly relies on spectral amplitude information while ignoring the robustness characterization of phase characteristics. Further research has found that the key discrimination frequency bands of samples of different categories are specific. This category difference in frequency sensitivity provides an opportunity for targeted frequency-domain attacks. The current defense methods enhance robustness through strategies such as high-frequency noise suppression and phase protection. However, breakthroughs are still needed in cross-band attack defense and dynamic spectrum registration, which points out the direction for subsequent research.
From the above content, it can be known that different frequency components in the existing research contribute differently to the model decision-making. Then, different frequency components also have guiding significance for the generation of adversarial samples. Based on this, this paper proposes an adversarial sample generation method based on channel awareness for frequency-domain transformation. At the same time, it can verify the importance of accurate channel estimation in adversarial attacks on communication systems. Provide new solutions for the security and reliability of communication systems.
In this section, the Fourier transform is adopted to transform the signal from a continuous time series signal in the time-domain to the frequency domain for analysis. The specific mathematical expression is as follows:
F ( w ) = f ( t ) × e j w t d t
e j w t = cos ( w t ) j × sin ( w t )
where e j w t represents an odd function. When both the function and its Fourier transform undergo discretization processing, the discrete Fourier transform (DFT) can be obtained. For radio signals that are usually modulated by I/Q modulators, the I/Q two-channel signals can be expressed as:
x ( t ) = I ( t ) + j Q ( t ) = e j φ ( t )
For a signal with a sampling length of L, it can be expressed as:
x = I Q , I = i 1 , i 2 , , i L Q = q 1 , q 2 , , q L
Therefore, the discrete Fourier transform of the modulated signal of length L in the I/Q two channels can be expressed as:
X ( m ) = n = 0 L 1 x ( n ) · e j 2 π L m n
Correspondingly, the frequency-domain signal can be converted into a time-domain signal through the inverse discrete Fourier transform, and its expression is:
x ( n ) = 1 N m = 0 L 1 X ( m ) · e j 2 π L m n
Based on the existing research and analysis, if the sensitivity of different frequency components of the sample to model recognition can be explored and malicious disturbances can be generated guided by this, the generated adversarial samples will be more targeted and threatening. For this purpose, this section proposes a frequency-domain signal processing method based on random masks. By inputting samples under different frequency distributions and observing the feedback of the model, the specific operation process T ( x ) can be expressed as:
T ( x ) = I D F T ( ( D F T ( x ) + D F T ( ξ ) ) M ) = I D F T ( D F T ( x + ξ ) M )
where D F T ( · ) and I D F T ( · ) represent the discrete Fourier transform and inverse discrete Fourier transform, ξ N ( 0 , δ 2 ) represents random noise obeying a Gaussian distribution. M U ( 1 ρ , 1 + ρ ) represents the mask matrix, whose elements are random samples in a uniform distribution, and ⊙ represents the Hadamar product, which is the product of each element in matrix operations. The frequency domain conversion process F can be seen as shown on Figure 2.
Furthermore, the signal data processed by Equation (42) are input into the target classification model and the model gradient information g is obtained. In order to obtain more reliable frequency-domain sensitivity information, in this section, the process T is selected for N times, that is, the noise ξ i and mask M n generated in each round, and a set of gradient information M n is obtained, where n = 1 , 2 , N . After completion, continue to sum the N gradients and calculate the average value. The total gradient information highlights the sensitive regions of the model for robust rows and key features, thereby guiding adversarial samples to be generated in a more threatening direction. Each generated mask element follows a uniform distribution. In this section, the weights of the time-domain and frequency-domain components of each gradient are set to 1.
Finally, the obtained gradient information is combined with the attack algorithm in Equation (12) to generate adversarial samples. To sum up, the complete adversarial attack algorithm based on frequency-domain transformation can be summarized as the following formula:
x i + 1 = c l i p x , ε x i + α · s i g n 1 N n = 1 N x i J T x i , y , H
where ε represents the limit of the disturbance L , α = ε / I , I represents the number of iterations s i g n · represents the sign function, α represents the iteration step size, J · represents the target model loss function, and H represents the channel matrix information.
Next, consider a wireless communication system composed of one transmitter, m receivers, and one adversary. All nodes are equipped with an antenna and operate on the same channel. Each receiver uses a neural network to classify the signals it receives into the modulation type used by the transmitter. Meanwhile, the opponent transmits disturbance signals through the air, deceiving the classifier on the receiver into making mistakes in modulation classification, thereby enabling the attack to succeed.
The deep neural network classifier at the ith receiving end is denoted as f i · ; θ i : χ R C , where a represents the parameters of the neural network at the ith receiving end and C is the number of modulation types. Here, χ C p , and p is the dimension of the complex input (in-phase/orthogonal component), which can also be expressed as the concatenation of two real input numbers. The classifier f i assigns the modulation type l ^ i x , θ i = argmax k f k i x , θ i to each input x χ , where f k i x , θ i is the output of the ith classifier for the kth modulation type.
The channel from the transmitting end to the ith receiving end is denoted as h t r i , and the channel from the opponent to the receiving end is denoted as h a r i . The vector forms are represented by h t r i = h t r i , 1 , h t r i , 2 , , h t r i , p T C p × 1 and h a r i = h a r i , 1 , h a r i , 2 , , h a r i , p T C p × 1 . When there is no adversarial attack, the transmitting end sends signal x, and the signal received by the ith receiving end is r t r i = H t r i x + n i . When there is an adversarial attack, if the attacker transmits a perturbation signal δ , the signal at the ith receiving end is r a r i δ = H t r i x + H a r i δ + n i , where H t r i and H a r i are diagonal matrices as mentioned earlier, and n i is Gaussian noise.
Suppose the adversarial disturbance δ and the transmitting signal x are synchronously superimposed at the receiving end to ensure the effectiveness of the attack. To achieve the concealment and energy efficiency of the attack, the adversarial perturbation δ needs to satisfy the power constraint δ 2 2 P max , where P max is the preset maximum perturbation power budget. The attacker needs to design a universal perturbation δ for the input signal x and all receiver classifiers f i by solving the following optimization problems (44):
arg min δ   δ 2 subject to   l ^ i r t r i , θ i l ^ i r a r i δ , θ i i = 1 , 2 , , m δ 2 2 P max
In optimization problems (44), the objective is to minimize the perturbation power (minimize the L 2 norm) to ensure that the perturbation power does not exceed the budget P max while satisfying all receiver classification errors. It should be noted that due to the complexity of the decision boundary of deep neural networks, the optimal solution may not be obtained at point δ 2 2 = P max .
The analysis will be carried out from the single-receiver scenario (m = 1), and the receiver index i will be omitted to simplify the expression. For targeted attacks, the attacker designs the perturbation δ by minimizing the loss function J r a r , y , ϕ . Based on the fast gradient method (FGM), the loss function can be linearly approximated as J r a r , y , ϕ J r t r , y , ϕ + H a r δ T x J r t r , y , ϕ . Minimization is achieved by setting H a r δ = β J r t r , y , ϕ , and β is the scaling factor used to constrain the adversarial disturbance power to P max .
In the MRPP attack [37], the attacker maximizes the perturbation power at the receiving end by selecting perturbations and analyzes the impact of this power on the classifier decision-making process. To achieve this goal, attackers need to make full use of the channel characteristics between the attacker and the receiving end. Specifically, if the target attack disturbance δ t a r g e t is multiplied by the conjugate h a r * of channel h a r , the received power can be maximized along the channel direction. After being transmitted through the channel, the disturbance power at the receiving end becomes h a r 2 2 δ t a r g e t . Through this operation, the adversarial attack not only maintains the consistency of the perturbation direction with the channel but also maximizes the transmission efficiency of the perturbation energy through the channel gain. Ultimately, the attacker needs to generate targeted perturbations for all possible modulation types and calculate the scaling factor to meet the power constraints of the opponent. The calculation of the scaling factor β has been obtained from reference [33] and will not be elaborated here.
Based on the above derivation and combined with the methods mentioned in the previous section, the adversarial sensing adversarial perturbation generation algorithm based on frequency-domain transformation can be obtained. The specific details are given in Algorithm 1.
Algorithm 1 Adversarial example generation algorithm channel sensing and frequency-domain transformation (FTHA).
Input :
Classification model f with parameters ϕ , clean sample x, true label y, L norm bound ϵ , iteration count I, frequency transformation count N, coordination factor ρ , noise ξ with standard deviation σ , mask M, H a r representing channel matrix to receiver, H a r * representing conjugate of channel matrix
Output :
Adversarial sample x
 1:
α = ϵ / I , x i = x , ξ follows Gaussian distribution ( 0 , δ 2 ) , M follows uniform distribution
 
    ( 1 ρ , 1 + ρ )
 2:
for  i = 1 to I do
 3:
    for  n = 1 to N do
 4:
         Random frequency transformation:
 
             T ( x ) = I D F T ( D F T ( x + ξ ) M )
 5:
         Gradient calculation: g n = x i J T x i , y ; ϕ
 6:
    end for
 7:
    Compute average gradient: g n = 1 N n = 1 N g n
 8:
    Channel information perturbation:
 
        δ t arg e t = H a r * x i J T x i , y ; ϕ H a r * x i J T x i , y ; ϕ 2
 9:
     x i + 1 = c l i p c l i p x , ε x i + α · H a r · δ t arg e t , 1 , 1
10:
end for
11:
x = x i
12:
return  x i

6. Performance Indicators

6.1. Channel Estimation Performance Indicators

In order to evaluate the performance of the proposed OMP-SDRNet frameworks in OFDM systems, we choose the mean square error and the bit error rate to measure. Firstly, the MSE performance of channel estimation is analyzed as follows:
M S E = 1 N i = 1 N ( H e s t i H r e a l ) 2
where H e s t i represents the estimated channel information, i.e., H S D N e t and H S D R N e t , and H r e a l represents the actual channel information. The BER performance of channel estimation is analyzed as follows:
B E R = B i t e r r B i t s u m
where B i t e r r represents the number of incorrect bits and B i t s u m represents the total number of bits.

6.2. Adversarial Attack Performance Indicators

To compare the adversarial attack methods in this section, the adversarial performance is measured by the misclassification rate, the average confidence of the wrong class prediction, the average confidence of the correct class prediction, the perturbation-to-signal ratio and the L 2 norm.
The misclassification ratio (MR) is the most important attribute in adversarial attacks. In untargeted attacks, MR is defined as the proportion of all samples that are successfully misclassified into any class, and therefore can also be called the attack success rate. Specifically as follows:
M R = 1 N i = 1 N c o u n t ( f ( x i a ) y i )
where f · represents the classification model, there are f ( x ) = y , x represents the normal sample, and x a represents the adversarial sample. In the experiment, all the adversarial samples were from the samples that were correctly classified by the target model.
The average confidence of adversarial class (ACAC) represents the model’s confidence in predicting the class of the sample after perturbation as follows:
A C A C = 1 n i = 1 n Q ( x i a ) f ( x i a )
where Q is the softmax layer output of the classifier f, there exists f ( x ) = arg max j Q ( x ) j , and Q ( x ) j represents the probability of the jth class.
The average confidence of true class (ACTC) represents the model’s confidence in the correct class of the sample after the attack. ACAC is used to evaluate the degree to which the adversarial sample deviates from the true class as follows:
A C T C = 1 n i = 1 n Q ( x i a ) y i
The L 2 norm is the most frequently used norm for calculating Euclidean distances and is also often employed as the regularization term for optimizing the objective function.
The perturbation-to-noise ratio is the ratio of perturbation power to noise power, which can be calculated through perturbation-to-signal ratio (PSR) and signal-to-noise ratio (SNR), as follows:
P N R = p p e r p n = P S R × S N R
where p p e r represents the disturbance power and p n represents the noise power.

7. Simulation Result

7.1. Channel Estimation

In this section, the experiments for verifying the effectiveness of the proposed method are described in detail, and the experimental results are presented. This experiment is built based on the OFDM communication system. The application scenario is set as transmission by a single antenna and reception by a single user. The channel model selected is the Rayleigh channel. The Pytorch toolkit is selected as the deep learning development tool, and all deep learning models are trained on NVIDIA Tesla V100-PCIE. We make full use of its powerful functions of deep learning model construction and training to provide strong support for the implementation of the SDNet and SDRNet frameworks. The sample size is set at 2000, and the signal data are generated by using the QPSK modulation method. To obtain the test set, the output of the OMP algorithm is used as the input data of the neural network, thereby providing basic data support for the subsequent network performance evaluation.
In the network compilation stage, the Adam optimizer is selected for training because it performs well in parameter optimization and can effectively adjust network parameters to improve model performance. The learning rate is set to 0.0005 to maintain a stable update during the training process. The mean square error is adopted as the loss function throughout the training process, and the loss of the validation set is taken as the key evaluation index to measure the training effect and generalization ability of the model. Furthermore, in all experiments, other parameters were kept uniform, and the number of loop iterations was fixed at 100 times to ensure the scientificity and comparability of the experimental results, facilitating the accurate evaluation of the performance of different methods. Secondly, the default system parameters used in the channel simulation are summarized as shown in Table 1.
Figure 7 and Figure 8 present a detailed analysis of the mean square error (MSE) and bit error rate (BER) of different channel estimation algorithms under a signal-to-noise ratio of 0–20 dB. The results show that the MSE and BER of all methods decrease as the signal-to-noise ratio increases. This is because a higher transmission power can effectively resist noise interference.
Based on the LS algorithm in low SNR and poor performance, a high signal-to-noise ratio is better. This is because the LS algorithm regards the channel as a definite but unknown constant and uses linear estimation, which is extremely sensitive to noise. In comparison, the performance of the MMSE algorithm is slightly better because it takes into account the prior statistical characteristics of the channel and uses weighted coefficients for linear estimation. The OMP algorithm estimates by gradually selecting the most relevant sparse channel components, which can effectively utilize the channel sparsity and have a higher estimation accuracy than the previous two.
Different from traditional algorithms, the framework proposed in this paper adopts a nonlinear and multi-layer neural network structure, which can estimate more complex channels. Overall, SDNet and SDRNet perform exceptionally well across the full signal-to-noise ratio range, with their MSE significantly lower than that of OMP, LS, and MMSE. Among them, SDRNet performed the most prominently. At 20 dB, the MSE was as low as 0.00004, showing significant advantages. This is attributed to SDRNet integrating multiple residual blocks on the basis of SDNet, retaining the output results of the previous module, and making full use of the features. Therefore, regardless of the environment of low signal-to-noise ratio (0–4 dB), medium signal-to-noise ratio (6–12 dB), or high signal-to-noise ratio (14–20 dB), the MSE and BER of SDRNet remain at a relatively low level, highlighting its superiority and robustness in channel estimation. These data strongly prove the effectiveness of SDNet and SDRNet in channel estimation, and further verify the performance advantages of the proposed method.
Figure 9 shows the MSE comparison of the proposed method under the same number of pilots, the same channel sparsity, and different pilot interval schemes. For SDNet, as the pilot interval I increases, the MSE initially shows an upward trend (from I = 4 to I = 8), and then significantly increases at I = 12. This indicates that under a smaller pilot interval, SDNet can better utilize the pilot signal for channel estimation. However, when the pilot interval is too large, the performance will decline significantly.
For SDRNet, MSE decreases slowly with the increase in the pilot interval I, but the overall change is not significant. This indicates that SDRNet is relatively insensitive to the change in pilot intervals and can maintain relatively stable performance under different pilot intervals. Under most pilot intervals, the MSE of SDRNet is generally lower than that of SDNet, especially at larger pilot intervals (such as I = 12), the performance advantage of SDRNet is more obvious. This indicates that SDRNet may have adopted more effective algorithms or structures in channel estimation and can provide better performance under different pilot intervals. Moreover, the MSE curve of SDRNet is smoother, showing better stability and robustness.
Figure 10 shows the comparison of MSE performance of different pilot quantity schemes under the same pilot interval and channel sparsity. For SDNet, as the number of pilots increases (Nc = 8 to 32), the MSE decreases significantly, mainly due to the cyclic prefix (CP) improving signal continuity and reducing time-domain aliasing. However, when the pilot is too much (Nc = 64), the MSE slightly rebounds, which might be due to an increase in pilot overhead or interference. The MSE of SDRNet continuously and smoothly decreases with the increase in pilot frequency, indicating that it can utilize pilot resources more efficiently. Overall, the variation range of MSE in the two methods is limited, indicating that they are not sensitive to the number of pilots, which is conducive to reducing the system overhead.
Figure 11 compares the MSE performance changes in SDNet and SDRNet under different channel sparsity (K = 3, 6, 9). When the sparsity of SDNet increases from K = 3 to K = 6, the MSE slightly decreases, indicating that a moderate increase in sparsity is conducive to improving the accuracy of channel estimation. However, when the sparsity further increases to K = 9 and K = 12, the MSE rises significantly, indicating that an excessively high sparsity will make the channel overly complex and reduce the estimation performance. In contrast, the MSE of SDRNet slightly decreased from K = 3 to K = 6, and then remained stable, demonstrating strong robustness to changes in sparsity. This indicates that SDRNet can better adapt to channel environments with different sparsity, while the performance of SDNet depends more on the reasonable selection of sparsity.

7.2. Adversarial Sample Generation Based on Frequency Conversion and Channel Awareness

To verify the effectiveness of the FTHA method, this study conducted a phased and progressive experiment for verification. Firstly, in the benchmark environment of the ideal channel, by comparing with traditional attack methods, the advantages of FTA in terms of perturbation efficiency and concealment are verified. Further, we introduce multi-dimensional channel conditions to construct an adversarial attack and defense test platform in real communication scenarios, and compare and analyze the performance differences between FTHA and the classic channel-aware attack strategies, as well as traditional attack methods considering channel conditions in key indicators such as attack success rates.
This section mainly introduces the specific implementation process and testing of adversarial samples based on frequency-domain transformation with channel awareness. As a result, the task experiments on signal modulation recognition in this chapter were conducted in the RadiOML 2016.10A public dataset Proceed. RadioML2016 is an open-source benchmark dataset in the field of wireless communication, mainly used for modulation identification tasks. This dataset was released by Tim O’Shea et al. in 2016, aiming to provide a standardized test platform for the performance evaluation of deep learning models in complex wireless signal processing tasks. Its core objective is to promote the application of machine learning in fields such as wireless communication security and spectrum sensing by simulating the signal characteristics in real communication environments. It includes 11 modulation methods, covering common digital modulations (such as BPSK, QPSK, 16QAM, 64QAM) and analog modulations (such as AM and FM). Each sample is IQ data In complex form (in-phase and quadrature components), with a sampling length of 128 time points, which can completely capture the time-domain characteristics of the signal. By adding Gaussian white noise (AWGN) with different signal-to-noise ratios, the noise interference in actual communication is simulated.
Furthermore, some data introduce channel distortions such as multipath fading and frequency shift. The dataset contains approximately 2.2 million samples in total, which are evenly distributed hierarchically by modulation type and SNR to ensure the comprehensiveness of training and evaluation. It is usually divided into the training set (80%), the validation set (10%), and the test set (10%), supporting the model development of supervised learning tasks.
For the target task, namely the model of automatic modulation recognition, the commonly used models of automatic modulation recognition selected in the experiments of this chapter are as follows:
(1) CNN1D: Modulation recognition model based on one-dimensional convolutional residual network;
(2) CNN2D: Modulation recognition model based on two-dimensional convolutional neural networks.
Firstly, without considering the channel information, in order to verify the effectiveness of the proposed adversarial attack algorithm based on frequency-domain transformation, this section selects multiple adversarial attack algorithms as baselines and compares them with the proposed method:
(1) FGSM: It generates adversarial samples by using the gradient information of the input data, and implements attacks by adding or subtracting the sign of the gradient on each element of the input data and multiplying it by a tiny perturbation.
(2) PGD: During the iterative process, adversarial samples are constructed by maximizing the loss function within the perturbation range to ensure that the generated samples have stronger interference.
(3) BIM: By applying minor disturbances to the original input and conducting multiple iterations to generate adversarial samples, the loss function is maximized within the disturbance range in each iteration.
(4) Autoattack (AA): Combining multiple attack methods to generate more deceptive adversarial samples in an automated manner;
(5) MIFGSM: On the basis of FGSM, a momentum term is introduced to generate adversarial samples through accelerated gradient descent.
Secondly, in order to verify the effectiveness of the proposed frequency-domain transformation adversarial attack algorithm based on channel awareness, the following attack algorithms based on channel information are selected as comparative experiments in this paper:
(1) Channel inversion attack: The attacker alters the characteristics of the wireless communication channel (such as channel gain, phase, etc.), causing abnormal changes in the channel state;
(2) Maximum disturbance power attack (MRPP): The attacker exploits the phase of the target communication channel obtained.
Adjust the power of the interfering signal in a targeted manner based on relevant information (such as channel status information, noise characteristics, etc.) and the size of the disturbance limited by power is changed to complete the attack.
As shown Table 2, under the CNN1D model and the RadioML2016.10a dataset, the frequency-domain adversarial attack method (FTA) is significantly superior to other attack algorithms in terms of attack efficiency and perturbation concealment. Specifically, the misclassification rate of FTA is 2.3% higher than that of the suboptimal method AA. Meanwhile, its perturbation energy is the lowest among the comparison methods, decreasing by 3.9% and 10.7% respectively compared with PGD and BIM. This result indicates that FTA, through the frequency-domain sparse perturbation injection strategy, can effectively cross the classifier decision boundary under extremely small perturbations.
In terms of confidence offset, the adversarial samples generated by FTA show a high degree of directional misleading. The mean confidence of the error class is close to the theoretical upper limit 1, while the confidence of the correct class is compressed to a value close to zero, which is 85.3% lower than that of FGSM. Although the ACAC and ACTC indicators of FTA are not significantly different from some mature methods (such as MIFGSM), its unique frequency-domain masking mechanism achieves the optimal balance between attack effectiveness and concealment by suppressing the disturbance of redundant frequency bands.
As shown in Table 3, under the experimental framework of the CNN2D model and the RadioML2016.10a dataset, the frequency-domain adversarial attack method (FTA) demonstrates balanced performance advantages. Although its attack success rate is slightly lower than that of the current optimal automatic attack AA, its disturbance concealment is significantly better than AA, reducing by 11.8% and 19.8%, respectively, compared with MIFGSM and FGSM. This result indicates that FTA, through the frequency-domain sparsification perturbation generation strategy, effectively inhibits the diffusion of redundant energy while ensuring the attack effectiveness, achieving an efficient balance between attack intensity and concealment.
Further analysis reveals that the confidence misleading ability of FTA differs from that of AA by only 2.3%, while its cross-model transferability is superior to that of AA, indicating that its perturbation mode has greater generalization potential. It is worth noting that all adversarial samples in the experimental design are derived from clean samples that can be correctly classified by the target model, resulting in a non-uniform distribution of the disturbance-to-signal ratio (PNR). This will cause local inconsistencies in the relationship between the L 2 norm and PNR, and it is necessary to optimize the data balance through dynamic signal-to-noise ratio sampling in subsequent studies.
Figure 12 shows the relationship between the classifier accuracy and PNR under the proposed target white-box adversarial attack with precise channel information and compares it with the channel inversion attack and the maximum perturbation power attack considering the information. It can be observed that the FTA algorithm without considering the channel effect has very poor performance, close to the no-attack situation in the low PNR region. This is because the wireless channel changes the phase and amplitude of the disturbance perceived by the receiver. Furthermore, compared with the MRPP attack, the target channel inversion attack performs poorly, which indicates the importance of the received perturbation power to the performance of the classifier at the receiver. The classification accuracy of the FTHA proposed in this paper is higher than that of MRPP and the channel inversion attack, indicating that it has better performance within a certain range of perturbation power.
It can be seen from Table 4 and Table 5 that under Gaussian channel conditions, the performance of the FTHA method on the CNN1D and CNN2D models is significantly better than that of other attack methods, especially in terms of the error classification rate and the average confidence of error class prediction. FTHA, as an improved method of adding channel information on the basis of FTA, can better adapt to the Gaussian channel environment and make full use of the channel characteristics to generate more threatening adversarial samples. On the CNN1D model, the misclassification rate of FTHA increased from 68.7% of FTA to 79.90%, while maintaining a relatively high average confidence level of error class prediction (0.935), indicating that it can mislead the model classification more effectively under channel conditions.
On the CNN2D model, FTHA further increased the misclassification rate to 80.50%, while ACAC reached 0.912, significantly higher than other methods, demonstrating its strong attack capability under complex models. In contrast, the performance of other methods such as FGSM and PGD in the Gaussian channel environment is relatively weak, especially in the CNN2D model. The misclassification rate of FGSM is only 28.00%, significantly lower than that of FTHA, which further highlights the advantages of FTHA under channel conditions. Overall, FTHA has demonstrated stronger attack capabilities and robustness in the Gaussian channel environment by combining channel information. Whether on the CNN1D or CNN2D models, it provides a better solution for adversarial sample generation.
It can be seen from the results of Table 6 and Table 7 that the accuracy of channel estimation has a decisive influence on the performance of the FTHA method, which directly reflects the key role of precise channel information in adversarial attacks. In CNN1D and CNN2D models, the performance of FTHA (SDRNet) has always been superior to other methods, especially in terms of the misclassification rate and the average confidence of misclass prediction. This is mainly because SDRNet can estimate the channel state more accurately, thereby generating more targeted adversarial samples. In contrast, FTHA (MMSE) has a significantly weaker attack effect than other methods due to its lower channel estimation accuracy. Especially in the CNN2D model, its MR Is only 77.50%, which is significantly lower than 80.50% of FTHA (SDRNet). This gap indicates that the error of channel estimation will directly affect the generation quality of adversarial samples, resulting in a decline in the attack effect.

8. Conclusions

To sum up, this study successfully proposed a new channel estimation method based on deep learning, giving full play to the advantages of deep learning in data processing and feature extraction, and achieving high-precision estimation of sparse channels in OFDM systems. On this basis, further innovation was carried out and a channel-aware adversarial sample generation method based on frequency-domain transformation was proposed. It effectively integrates channel estimation and frequency-domain transformation techniques, significantly improving the generation effect of adversarial samples.
Judging from the phenomena revealed by the research results, deep learning models can capture channel features more accurately. In the complex OFDM system environment, they show stronger adaptability and accuracy compared with traditional algorithms. This theoretical achievement not only enriches the research methods in the field of channel estimation, but also provides new ideas for the optimization of communication systems. In terms of practical significance, high-precision channel estimation and more effective adversarial sample generation methods contribute to enhancing the stability and security of communication systems, providing more reliable technical support for practical communication applications.
Although this study has achieved remarkable results, there is still room for further research. In the future, the performance optimization of deep learning models under different channel conditions can be deeply explored, and studies can be conducted on how to further enhance the universality and concealment of adversarial samples. Meanwhile, studies can apply these methods to more complex communication scenarios, verify their effectiveness and feasibility, and contribute more to the development of communication technology.

Author Contributions

Conceptualization, Y.G.; methodology, Y.G.; software, Y.G.; validation, Y.G. and H.Z.; formal analysis, Y.G. and H.Z.; investigation, Y.G.; resources, Y.G.; data curation, Y.G.; writing—original draft preparation, Y.G.; writing—review and editing, Y.G.; visualization, Y.G.; supervision, D.X.; project administration, Q.X. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The first dataset was random. Relevant data files can be obtained from the corresponding author upon reasonable request.The second dataset used in this study is a publicly available dataset RadiOML 2016.10A, which can be accessed via https://www.deepsig.ai/datasets (accessed on 29 April 2025). Detailed information about the dataset is provided in the Methods section. Experimental results supporting the findings of this study are presented in the manuscript (Tables and Figures). Additional raw data or analysis results can be provided by the authors upon request.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Ren, K.; Zheng, T.; Qin, Z.; Liu, X. Adversarial attacks and defenses in deep learning. Engineering 2020, 6, 346–360. [Google Scholar] [CrossRef]
  2. Tian, Q.; Chen, Y.; Zhang, Z.; Lu, H.; Chen, L.; Xie, L.; Liu, S. TFGAN: Time and frequency domain based generative adversarial network for high-fidelity speech synthesis. arXiv 2020, arXiv:2011.12206. [Google Scholar]
  3. Lin, L.; Blaser, E.; Wang, H. Graph structural attack by perturbing spectral distance. In Proceedings of the 28th ACM SIGKDD Conference on Knowledge Discovery and Data Mining, Washington, DC, USA, 14–18 August 2022; pp. 989–998. [Google Scholar]
  4. Lu, Y.; Ma, T.; Pang, Z.; Chai, X.; Chen, Z.; Tang, Z. Frequency domain-based reversible adversarial attacks for privacy protection in Internet of Things. J. Electron. Imaging 2024, 33, 043049. [Google Scholar] [CrossRef]
  5. Zarrinkoub, H. Understanding LTE with MATLAB: From Mathematical Modeling to Simulation and Prototyping; John Wiley & Sons: Hoboken, NJ, USA, 2014. [Google Scholar]
  6. Hussein, Y.S.; Alias, M.Y.; Abdulkafi, A.A. On performance analysis of LS and MMSE for channel estimation in VLC systems. In Proceedings of the 2016 IEEE 12th International Colloquium on Signal Processing & Its Applications (CSPA), Melaka, Malaysia, 4–6 March 2016; pp. 204–209. [Google Scholar]
  7. Mingming, F.; Jin, H. An improved channel estimation algorithm based on DFT in OFDM system. In Proceedings of the 2020 International Conference on Computer Information and Big Data Applications (CIBDA), Guiyang, China, 17–19 April 2020; pp. 321–325. [Google Scholar]
  8. Zettas, S.; Lazaridis, P.I.; Zaharis, Z.D.; Kasampalis, S.; Prasad, N.; Glover, I.A.; Cosmas, J.P. Performance comparison of LS, LMMSE and adaptive averaging channel estimation (AACE) for DVB-T2. In Proceedings of the 2015 IEEE International Symposium on Broadband Multimedia Systems and Broadcasting, Ghent, Belgium, 17–19 June 2015; pp. 1–5. [Google Scholar]
  9. Wu, H. LMMSE channel estimation in OFDM systems: A vector quantization approach. IEEE Commun. Lett. 2021, 25, 1994–1998. [Google Scholar] [CrossRef]
  10. Chun, C.J.; Kang, J.M.; Kim, I.M. Deep learning-based channel estimation for massive MIMO systems. IEEE Wirel. Commun. Lett. 2019, 8, 1228–1231. [Google Scholar] [CrossRef]
  11. Wen, C.K.; Shih, W.T.; Jin, S. Deep learning for massive MIMO CSI feedback. IEEE Wirel. Commun. Lett. 2018, 7, 748–751. [Google Scholar] [CrossRef]
  12. Wang, T.; Wen, C.K.; Jin, S.; Li, G.Y. Deep learning-based CSI feedback approach for time-varying massive MIMO channels. IEEE Wirel. Commun. Lett. 2018, 8, 416–419. [Google Scholar] [CrossRef]
  13. Zhang, J.; Ma, X.; Qi, J.; Jin, S. Designing tensor-train deep neural networks for time-varying MIMO channel estimation. IEEE J. Sel. Top. Signal Process. 2021, 15, 759–773. [Google Scholar] [CrossRef]
  14. Gao, S.; Chen, R.; Feng, R. Channel Estimation Based on Improved Orthogonal Matching Pursuit Algorithm. In Proceedings of the 2023 4th International Symposium on Computer Engineering and Intelligent Communications (ISCEIC), Nanjing, China, 18–20 August 2023; pp. 472–475. [Google Scholar]
  15. Korrai, P.K.; Sen, D. Performance analysis of OFDM mmWave communications with compressive sensing based channel estimation and impulse noise suppression. In Proceedings of the 2016 IEEE International Conference on Advanced Networks and Telecommunications Systems (ANTS), Bangalore, India, 6–9 November 2016; pp. 1–6. [Google Scholar]
  16. Gizzini, A.K.; Chafii, M.; Nimr, A.; Fettweis, G. Deep learning based channel estimation schemes for IEEE 802.11 p standard. IEEE Access 2020, 8, 113751–113765. [Google Scholar] [CrossRef]
  17. Rao, K.D.; Kartheek, B. Comparative performance analysis of omp and sabmp for massive mimo ofdm channel estimation. In Proceedings of the 2018 5th IEEE Uttar Pradesh Section International Conference on Electrical, Electronics and Computer Engineering (UPCON), Gorakhpur, India, 2–4 November 2018; pp. 1–5. [Google Scholar]
  18. Wael, C.B.A.; Suyoto, S.; Armi, N.; Rohman, B.P.A.; Kurniawan, D.; Praludi, T.; Satyawan, A.S.; Subekti, A.; Adhi, P. Sparse channel estimation using orthogonal matching pursuit (OMP) for MIMO-OFDM system. In Proceedings of the 2022 International Conference on Radar, Antenna, Microwave, Electronics, and Telecommunications (ICRAMET), Bandung, Indonesia, 6–7 December 2022; pp. 258–261. [Google Scholar]
  19. Masood, M.; Al-Naffouri, T.Y. Sparse reconstruction using distribution agnostic Bayesian matching pursuit. IEEE Trans. Signal Process. 2013, 61, 5298–5309. [Google Scholar] [CrossRef]
  20. Liu, S.; Huang, X. Sparsity-aware channel estimation for mmWave massive MIMO: A deep CNN-based approach. China Commun. 2021, 18, 162–171. [Google Scholar] [CrossRef]
  21. Muzavazi, R. Channel Estimation and Data Detection Schemes for Orthogonal Time Frequency Space Massive. Master’s Thesis, University of the Witwatersrand, Johannesburg, South Africa, 2022. [Google Scholar]
  22. Deng, Q.; Ge, Y.; Ding, Z. A unifying view of OTFS and its many variants. IEEE Commun. Surv. Tutor. 2025. [Google Scholar] [CrossRef]
  23. Szegedy, C.; Zaremba, W.; Sutskever, I.; Bruna, J.; Erhan, D.; Goodfellow, I.; Fergus, R. Intriguing properties of neural networks. arXiv 2013, arXiv:1312.6199. [Google Scholar]
  24. Hasan, M.M.; Islam, R.; Mamun, Q.; Islam, M.Z.; Gao, J. Adversarial Attacks on Deep Learning-Based Network Intrusion Detection Systems: A Taxonomy and Review. 2025. Available online: https://ssrn.com/abstract=4863302 (accessed on 29 April 2025).
  25. Wu, W.; Su, Y.; Lyu, M.R.; King, I. Improving the transferability of adversarial samples with adversarial transformations. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 9024–9033. [Google Scholar]
  26. Luo, C.; Lin, Q.; Xie, W.; Wu, B.; Xie, J.; Shen, L. Frequency-driven imperceptible adversarial attack on semantic similarity. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 15315–15324. [Google Scholar]
  27. Chen, S.; Chen, Z.; Wang, D. Adaptive adversarial training for meta reinforcement learning. In Proceedings of the 2021 International Joint Conference on Neural Networks (IJCNN), Shenzhen, China, 18–22 July 2021; pp. 1–8. [Google Scholar]
  28. Xu, Y.; Xu, G.; An, Z.; Nielsen, M.H.; Shen, M. Adversarial attacks and active defense on deep learning based identification of GaN power amplifiers under physical perturbation. AEU-Int. J. Electron. Commun. 2023, 159, 154478. [Google Scholar] [CrossRef]
  29. Wang, H.; Li, G.; Liu, X.; Lin, L. A Hamiltonian Monte Carlo method for probabilistic adversarial attack and learning. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 44, 1725–1737. [Google Scholar] [CrossRef]
  30. Zhang, L.; Lambotharan, S.; Zheng, G. Adversarial learning in transformer based neural network in radio signal classification. In Proceedings of the ICASSP 2022–2022 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Singapore, 23–27 May 2022; pp. 9032–9036. [Google Scholar]
  31. Manoj, B.; Santos, P.M.; Sadeghi, M.; Larsson, E.G. Toward robust networks against adversarial attacks for radio signal modulation classification. In Proceedings of the 2022 IEEE 23rd International Workshop on Signal Processing Advances in Wireless Communication (SPAWC), Oulu, Finland, 4–6 July 2022; pp. 1–5. [Google Scholar]
  32. Kotak, J.; Elovici, Y. Adversarial attacks against IoT identification systems. IEEE Internet Things J. 2022, 10, 7868–7883. [Google Scholar] [CrossRef]
  33. Sadeghi, M.; Larsson, E.G. Adversarial attacks on deep-learning based radio signal classification. IEEE Wirel. Commun. Lett. 2018, 8, 213–216. [Google Scholar] [CrossRef]
  34. Lin, Y.; Zhao, H.; Tu, Y.; Mao, S.; Dou, Z. Threats of adversarial attacks in DNN-based modulation recognition. In Proceedings of the IEEE INFOCOM 2020-IEEE Conference on Computer Communications, Toronto, ON, Canada, 6–9 July 2020; pp. 2469–2478. [Google Scholar]
  35. Sandler, R.A.; Relich, P.K.; Cho, C.; Holloway, S. Real-time over-the-air adversarial perturbations for digital communications using deep neural networks. arXiv 2022, arXiv:2202.11197. [Google Scholar]
  36. Cohen, J.; Rosenfeld, E.; Kolter, Z. Certified adversarial robustness via randomized smoothing. In Proceedings of the International Conference on Machine Learning. PMLR, Long Beach, CA, USA, 9–15 June 2019; pp. 1310–1320. [Google Scholar]
  37. Kim, B.; Sagduyu, Y.E.; Davaslioglu, K.; Erpek, T.; Ulukus, S. Channel-aware adversarial attacks against deep learning-based wireless signal classifiers. IEEE Trans. Wirel. Commun. 2021, 21, 3868–3880. [Google Scholar] [CrossRef]
  38. Guo, C.; Frank, J.S.; Weinberger, K.Q. Low frequency adversarial perturbation. arXiv 2018, arXiv:1809.08758. [Google Scholar]
  39. Sharma, Y.; Ding, G.W.; Brubaker, M. On the effectiveness of low frequency perturbations. arXiv 2019, arXiv:1903.00073. [Google Scholar]
  40. Duan, R.; Chen, Y.; Niu, D.; Yang, Y.; Qin, A.K.; He, Y. Advdrop: Adversarial attack to dnns by dropping information. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Virtual, 11–17 October 2021; pp. 7506–7515. [Google Scholar]
Figure 1. OFDM system model.
Figure 1. OFDM system model.
Sensors 25 03779 g001
Figure 2. SDRNet channel estimation framework.
Figure 2. SDRNet channel estimation framework.
Sensors 25 03779 g002
Figure 3. Flowchart of adversarial sample production method based on frequency-domain algorithm and channel awareness.
Figure 3. Flowchart of adversarial sample production method based on frequency-domain algorithm and channel awareness.
Sensors 25 03779 g003
Figure 4. Deep neural network-based channel estimation framework.
Figure 4. Deep neural network-based channel estimation framework.
Sensors 25 03779 g004
Figure 5. Network architecture of the SDNet.
Figure 5. Network architecture of the SDNet.
Sensors 25 03779 g005
Figure 6. Network architecture of the SDRNet.
Figure 6. Network architecture of the SDRNet.
Sensors 25 03779 g006
Figure 7. Comparison of MSE results for different channel estimation methods.
Figure 7. Comparison of MSE results for different channel estimation methods.
Sensors 25 03779 g007
Figure 8. Comparison of BER results for different channel estimation methods.
Figure 8. Comparison of BER results for different channel estimation methods.
Sensors 25 03779 g008
Figure 9. Comparison of MSE results for different pilot spacing I.
Figure 9. Comparison of MSE results for different pilot spacing I.
Sensors 25 03779 g009
Figure 10. Comparison of MSE results for different cyclic prefix lengths (CP).
Figure 10. Comparison of MSE results for different cyclic prefix lengths (CP).
Sensors 25 03779 g010
Figure 11. Comparison of MSE results for different channel sparsity levels (K).
Figure 11. Comparison of MSE results for different channel sparsity levels (K).
Sensors 25 03779 g011
Figure 12. Classification accuracy under different attacks.
Figure 12. Classification accuracy under different attacks.
Sensors 25 03779 g012
Table 1. Default parameters of the channel simulation system.
Table 1. Default parameters of the channel simulation system.
Parameter NameParameter Valve
Symbol period 6.4 × 10 6
Multipath quantity3
Sampling period 1 × 10 6
Number of loop iterations100
Delay time0, 1 × 10 6 , 2 × 10 6
Root mean square delay spread 4 × 10 6
Table 2. Performance of different methods on CNN1D under ideal channel conditions.
Table 2. Performance of different methods on CNN1D under ideal channel conditions.
Attack MethodMR (%)ACACACTC L 2
PGD81.00.9900.0071.104
FGSM31.70.8350.0721.420
BIM81.30.9020.0071.190
AA91.10.0440.0101.202
MIFGSM82.90.9660.0051.325
FTA93.40.9380.0111.062
Table 3. Performance of different methods on CNN2D under ideal channel conditions.
Table 3. Performance of different methods on CNN2D under ideal channel conditions.
Attack MethodMR (%)ACACACTC L 2
PGD72.200.8490.0851.247
FGSM32.400.8820.0481.579
BIM71.000.8480.0871.233
AA81.010.8760.0691.332
MIFGSM71.900.8440.0811.437
FTA76.820.8560.0711.267
Table 4. Performance of different methods on CNN1D under gaussian channel conditions.
Table 4. Performance of different methods on CNN1D under gaussian channel conditions.
Attack MethodMR (%)ACACACTC L 2
PGD65.000.9500.0201.300
FGSM25.000.8000.1001.600
BIM65.050.8500.0201.350
AA69.600.8000.0301.400
MIFGSM67.000.9200.0151.500
FTA68.700.9000.0251.250
FTHA79.900.9350.0191.268
Table 5. Performance of different methods on CNN2D under gaussian channel conditions.
Table 5. Performance of different methods on CNN2D under gaussian channel conditions.
Attack MethodMR (%)ACACACTC L 2
PGD65.000.8200.0901.412
FGSM28.000.8500.0601.734
BIM64.000.8200.0901.425
AA75.000.8500.0701.511
MIFGSM65.500.8100.0851.669
FTA67.580.8300.0751.458
FTHA80.500.9120.0081.318
Table 6. Performance of FTHA of different channel estimation methods on CNN1D.
Table 6. Performance of FTHA of different channel estimation methods on CNN1D.
Attack MethodMR (%)ACACACTC L 2
FTA (No Channel)68.700.9000.0251.250
FTHA (MMSE)76.700.8580.0141.200
FTHA (OMP)78.800.9350.0081.250
FTHA (LS)78.500.8680.0051.300
FTHA (SDRNet)79.900.9350.0051.350
Table 7. Performance of FTHA of different channel estimation methods on CNN2D.
Table 7. Performance of FTHA of different channel estimation methods on CNN2D.
Attack MethodMR (%)ACACACTC L 2
FTA (No Channel)67.580.8300.0751.458
FTHA (MMSE)77.500.8310.0181.242
FTHA (OMP)75.300.8200.0111.132
FTHA (LS)78.500.8430.0081.305
FTHA (SDRNet)80.500.9120.0081.318
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Gao, Y.; Xu, D.; Zhu, H.; Xuan, Q. Adversarial Sample Generation Method Based on Frequency Domain Transformation and Channel Awareness. Sensors 2025, 25, 3779. https://doi.org/10.3390/s25123779

AMA Style

Gao Y, Xu D, Zhu H, Xuan Q. Adversarial Sample Generation Method Based on Frequency Domain Transformation and Channel Awareness. Sensors. 2025; 25(12):3779. https://doi.org/10.3390/s25123779

Chicago/Turabian Style

Gao, Yalin, Dongwei Xu, Huiyan Zhu, and Qi Xuan. 2025. "Adversarial Sample Generation Method Based on Frequency Domain Transformation and Channel Awareness" Sensors 25, no. 12: 3779. https://doi.org/10.3390/s25123779

APA Style

Gao, Y., Xu, D., Zhu, H., & Xuan, Q. (2025). Adversarial Sample Generation Method Based on Frequency Domain Transformation and Channel Awareness. Sensors, 25(12), 3779. https://doi.org/10.3390/s25123779

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop