Next Article in Journal
An Optimized Deep Learning Approach for Multiclass Anomaly Detection
Previous Article in Journal
UDPLDP-Tree: Range Queries Under User-Distinguished Personalized Local Differential Privacy
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Adversarial Example Generation Method Based on Wavelet Transform

1
School of Software, Shenyang University of Technology, Shenyang 110000, China
2
Bank of Xinjiang Co., Ltd., Tianshan District, Urumqi 830000, China
3
College of Electrical Engineering, Shenyang University of Technology, Shenyang 110000, China
*
Author to whom correspondence should be addressed.
Information 2026, 17(2), 182; https://doi.org/10.3390/info17020182
Submission received: 8 January 2026 / Revised: 31 January 2026 / Accepted: 5 February 2026 / Published: 10 February 2026

Abstract

Adversarial examples are crucial tools for assessing the robustness of deep neural networks (DNNs) and revealing potential security vulnerabilities. Adversarial example generation methods based on Generative Adversarial Networks (GANs) have made significant progress in generating image adversarial examples, but still suffer from insufficient sparsity and transferability. To address these issues, this study proposes a novel semi-white-box untargeted adversarial example generation method named Wavelet-AdvGAN, with an explicit threat model defined as follows. Specifically, the attack is strictly untargeted without predefined target categories, aiming solely to mislead DNNs into classifying adversarial examples into any category other than the original label. It adopts a semi-white-box setting where attackers are denied access to the target model’s private information. Regarding the generator’s information dependence, the training phase only utilizes public resources (i.e., the target model’s public architecture and CIFAR-10 public training data), while the test phase generates adversarial examples through one-step feedforward of clean images without interacting with the target model. The method incorporates a Frequency Sub-band Difference (FSD) module and a Wavelet Transform Local Feature (WTLF) extraction module, evaluating the differences between original and adversarial examples from the frequency domain perspective. This approach constrains the magnitude of perturbations, reinforces feature regions, and further enhances the attack effectiveness, thereby improving the sparsity and transferability of adversarial examples. Experimental results demonstrate that the Wavelet-AdvGAN method achieves an average increase of 1.26% in attack success rates under two defense strategies—data augmentation and adversarial training. Additionally, the adversarial transferability improves by an average of 2.7%. Moreover, the proposed method exhibits a lower l 0 norm, indicating better perturbation sparsity. Consequently, it effectively evaluates the robustness of deep neural networks.

1. Introduction

In recent years, with the rapid advancement of machine learning technologies, deep neural networks (DNNs) have been widely applied in fields such as image recognition [1] and analog compute in memory [2]. They have achieved remarkable progress, particularly in safety-critical domains like autonomous driving [3] and facial recognition [4]. However, despite their excellent performance across various tasks, DNNs exhibit significant vulnerability to adversarial robustness. Szegedy et al. [5] first pointed out that by introducing subtle, imperceptible perturbations into image data, classification models could be misled into making incorrect predictions (i.e., untargeted adversarial attacks, where the goal is to deviate from the original category rather than specify a target category), resulting in adversarial examples. In practical scenarios such as autonomous driving, untargeted attacks can cause severe safety risks (e.g., misclassifying traffic signs into irrelevant categories). This study focuses on semi-white-box untargeted attacks (consistent with the explicit threat model in the Abstract, i.e., no access to the target model’s private information during training/testing), aiming to generate high-sparsity and high-transferability adversarial examples to evaluate DNN robustness. These examples pose considerable risks. For instance, in autonomous driving, minor perturbations in images may cause systems to misidentify traffic signs, leading to severe safety issues. Therefore, researching adversarial example generation methods and their impact on deep learning models has become crucial for improving model robustness and ensuring application security.
Current adversarial example generation methods can be broadly categorized into three classes. The first category, gradient-based methods [5,6,7,8], generates adversarial examples quickly by applying perturbations in the direction of the model’s loss gradient. These methods are efficient but often lack control over perturbation sparsity. The second category, optimization-based methods [9], uses optimization algorithms to produce more visually natural and sparse adversarial examples, usually with higher attack success rates. However, these methods are computationally expensive and often require optimization for individual data samples.
The third category involves GAN-based methods [10], which generate adversarial examples rapidly by learning the probability distribution of input data through adversarial training between a generator and a discriminator. Xiao et al. proposed AdvGAN [11], which employs a generator to produce perturbations and a discriminator to ensure the authenticity of the generated samples, achieving higher efficiency than traditional optimization methods like C&W [9]. However, AdvGAN primarily focuses on global image features in its generator design, leading to dense perturbations with insufficient sparsity and large perturbation magnitudes, compromising the authenticity of the samples. AdvGAN++ [12] improves on this by leveraging latent features of deep neural networks instead of direct input images for generating adversarial examples. By sampling noise vectors and generating distributions, it extracts latent features to produce adversarial images closer to the input distribution. This approach relies on the observation that latent features are more susceptible to adversarial perturbations than input images. However, the generated adversarial examples depend on the vulnerability of latent features, which may vary with model architecture and training strategies, resulting in poor transferability.
To enhance the adversarial effect on classification models, Bai et al. proposed AI-GAN [13], which introduces an attack module during training to strengthen the generator’s attack capability and stabilize GAN training, further improving the quality and sparsity of adversarial examples. Despite these improvements, limitations in enhancing sparsity remain. GE-AdvGAN [14] introduces a novel gradient editing (GE) mechanism, exploring frequency domains to determine the direction of gradient editing, thereby improving transferability and algorithm efficiency. However, G E - A d v G A N only uses frequency-domain information to guide gradient direction, lacking multi-scale frequency decomposition and targeted constraint on perturbation distribution. These methods— A d v G A N , A d v G A N ++, A I - G A N , and G E - A d v G A N —rely on the l 2 norm to assess differences between original and adversarial samples, which is insufficient to capture multi-scale structural and texture details. Moreover, they fail to synergize frequency-domain constraints with local feature enhancement: AdvGAN series focus on global features, leading to dense perturbations; AI-GAN strengthens attack capability but ignores frequency-domain structure preservation; and GE-AdvGAN lacks targeted enhancement of critical local regions. These limitations result in compromised sparsity and transferability of adversarial examples. However, the l 2 norm is less effective in capturing multi-scale features, structural information, and texture details of images. This study addresses this by leveraging local image features and details from the frequency domain to evaluate differences, thereby enhancing the sparsity of adversarial examples.
Our main contributions are as follows:
1.
We propose the Wavelet-AdvGAN adversarial example generation method, which achieves improved attack success rates, generation efficiency, transferability, and sparsity. Unlike existing frequency-domain GAN-based methods (e.g., GE-AdvGAN) that rely on single-scale frequency difference or gradient editing, Wavelet-AdvGAN innovates a ‘frequency-domain constraint + local feature enhancement’ synergistic mechanism to address the long-standing trade-off between perturbation sparsity and transferability.
2.
We introduce the Frequency Sub-band Difference (FSD) module based on wavelet transform. Unlike existing frequency-domain methods that use uniform frequency metrics (e.g., l 2 norm), this module decomposes images into multi-scale sub-bands (LL/LH/HL/HH) and assigns adaptive weights, emphasizing the preservation of global structural information (low-frequency) while constraining detailed perturbations (high-frequency), thus generating adversarial examples with higher sparsity and visual authenticity.
3.
We present the Wavelet Transform Local Feature (WTLF) module, which leverages wavelet convolution to separate global structures (low-frequency) and detailed features (high-frequency), and integrates lightweight 1D convolution to preserve channel information. Unlike AdvGAN series that focus only on global features or AI-GAN’s attack module enhancement, this module enables the model to focus on critical local regions, directly improving attack effectiveness while maintaining efficiency. This design extends the ‘divide-and-conquer’ idea from mixture-of-experts (MoE) architectures to adversarial attack’s frequency-domain processing.

2. Related Work

2.1. GAN-Based Adversarial Example Generation Methods

The basic structure of a Generative Adversarial Network (GAN), as shown in Figure 1, consists of a generator G and a discriminator D. The generator G produces data from random noise, aiming to approximate the real data distribution. Meanwhile, the discriminator D classifies the input data, determining whether it comes from the real data distribution or is generated by G. In the initial stage, the samples generated by G are of low quality, allowing D to easily distinguish between real and generated data. D uses this distinction to optimize its parameters. Concurrently, G adjusts its parameters based on the gradient feedback from D, progressively improving the quality of the generated samples to closely resemble real data. This process involves an iterative optimization where G and D are alternately trained, forming a dynamic adversarial game: G strives to “fool” D, while D continuously enhances its discriminative capability. This adversarial mechanism ultimately leads G to produce high-quality samples, while D maintains strong performance in distinguishing real from generated data.
Xiao et al. [11] combined adversarial attacks with Generative Adversarial Networks (GANs) and proposed AdvGAN to address the time-consuming nature of optimization-based attack algorithms.
This method leverages the ability of GANs [10] to learn and approximate the distribution of the original data. The model is trained on the dataset, where adversarial examples are generated through the feedforward generator network, and the discriminator encourages the adversarial examples to become more similar to the original images. A key innovation of AdvGAN is the definition of semi-white-box attacks. Wavelet-AdvGAN inherits and clarifies this semi-white-box setting (a core component of the threat model defined in the Abstract), consistent with the practical attack scenario design in related studies: (1) Training phase: The generator only requires the architecture of the target model (e.g., ResNet18, Vgg16) and public training data (CIFAR-10) to learn perturbation patterns; it does not need access to the target model’s private parameters, gradients, or intermediate feature maps. (2) Testing phase: After training, the generator can directly generate adversarial examples by inputting clean images, without any further interaction with the target model (e.g., no need to query the model’s output or gradient). This semi-white-box setting, combined with the strictly untargeted attack type, forms the complete threat model of this study. This advancement makes the process more efficient and scalable, particularly when generating adversarial examples for large datasets or for situations where multiple attack attempts are required.
The structure of AdvGAN is shown in Figure 2.
Beyond the aforementioned GAN-based methods, two recent studies are closely related to our research. Zhang et al. [15] proposed “Towards Model Resistant to Transferable Adversarial Examples via Trigger Activation”, which enhances model robustness against transferable adversarial examples by strengthening the model’s dependence on specific trigger features (e.g., task-specific local patterns). Their method constructs defense by binding model decision to fixed triggers, but overlooks the adaptability of adversarial examples to multi-scale frequency features. This study complements our work: while their goal is to defend against transferable attacks, Wavelet-AdvGAN targets generating more transferable adversarial examples that can break through such trigger-based defense mechanisms by leveraging wavelet transform to perturb multi-scale frequency components (not limited to trigger-related local features) and enhancing local critical regions via WTLF module, thus providing a more rigorous test for model robustness.
Another study, “Moe-ffd: Mixture of experts for generalized and parameter-efficient face forgery detection” [16] by Liu et al., adopts a mixture-of-experts (MoE) architecture to decompose complex face forgery detection tasks into sub-tasks handled by specialized experts (each expert focuses on a specific forgery pattern). Their core insight is that “divide-and-conquer” can improve the efficiency and targeting of complex task processing. Inspired by this idea, our WTLF module extends the MoE concept to the field of adversarial attack: we treat wavelet multi-scale sub-bands (LL/LH/HL/HH) as “frequency experts”, where each expert is responsible for perturbing a specific frequency component (low-frequency for global structure, high-frequency for local details). This design enables targeted perturbation for different frequency components, avoiding the over-perturbation of irrelevant regions and thus improving perturbation sparsity—addressing the limitation of MoE’s original application in detection tasks (not applicable to adversarial perturbation generation) by adapting it to frequency-domain feature processing.

2.2. Wavelet Transform

Wavelet Transform, known for its multi-scale analysis capabilities, is widely applied in feature extraction within machine learning and signal processing, particularly excelling at capturing local features of signals [17,18,19,20]. By decomposing a signal into the superposition of wavelet functions at different scales, the wavelet transform provides localized information in both the time and frequency domains. Different types of wavelet basis functions achieve localized feature representation across multiple scales through translation, scaling, decomposition, and reconstruction operations.
The 2D Discrete Wavelet Transform (DWT) [21] decomposes an input signal into four sub-bands: Low–Low (LL), Low–High (LH), High–Low (HL), and High–High (HH). This decomposition reveals the signal’s frequency characteristics in different directions, enabling detailed analysis of its local structures.

2.3. ELA

Efficient Layer Attention (ELA) [22] effectively enhances image classification and object detection tasks by precisely capturing both local and global features while reducing the number of parameters. Compared to traditional spatial attention methods such as Squeeze-and-Excitation (SE) [23], Coordinate Attention (CA) [24], and Convolutional Block Attention Module (CBAM) [25], ELA demonstrates significant performance improvements and lower computational costs across various visual tasks. The ELA module aims to improve the performance of Convolutional Neural Networks (CNNs) in tasks like image classification, object detection, and semantic segmentation. Its core innovation lies in using 1D convolutions [26] and Group Normalization (GN) [27], effectively encoding horizontal and vertical spatial position information. This approach avoids the dimensionality reduction issues present in existing channel-based methods, ensuring richer feature representation and more efficient processing.

2.4. Adversarial Training

Adversarial Training (AT), proposed by Goodfellow et al. [6], involves training a model using both original and adversarial samples (such as those generated by FGSM) to enhance its robustness. Existing research [28,29,30,31] highlights the importance of adversarial training in improving the resilience of deep neural networks against adversarial perturbations. These studies emphasize its effectiveness across various application scenarios, including defense against backdoor attacks, generalization performance optimization, cross-domain applications, economic considerations, and exploration of future research directions. These findings collectively indicate that AT not only strengthens model robustness but also drives the development and diversification of defense strategies and their applications.

2.5. Data Enhancement

The importance of data augmentation techniques in enhancing the adversarial robustness of deep neural networks is widely recognized. Research has shown that optimizing adversarial training can be significantly improved by increasing data diversity [32], dynamically adjusting augmentation intensity [33], introducing regularizers [34], and employing spatial combination techniques [35]. These approaches help mitigate overfitting issues and enhance model robustness. Moreover, online instance-level data augmentation strategies [36] not only reduce the cost of searching for optimal augmentation policies but also further improve model stability and reliability against adversarial attacks. These findings collectively highlight the pivotal role of data augmentation in strengthening adversarial training and offer new directions for developing more efficient defense strategies. In this paper, we will systematically evaluate the effectiveness of the Wavelet-AdvGAN method based on data augmentation techniques to validate its practical value.

3. Method

3.1. Wavelet-AdvGAN

This study proposes an improved adversarial example generation method, Wavelet-AdvGAN, with the model structure shown in Figure 3.
Wavelet-AdvGAN consists of four components: the generator G, the discriminator D, the target model f, and the Frequency Sub-band Difference ( FSD ) module. G is responsible for generating perturbations (complying with the semi-white-box constraint: no reliance on the target model’s private information) while ensuring sparsity and imperceptibility, while D distinguishes between generated samples and original samples, guiding the training of G. The target model f ensures that the adversarial examples are misclassified into any category other than the original (i.e., the core goal of untargeted attacks, consistent with the threat model). The generator’s key objective is to generate perturbations that are sparse and imperceptible, while maximizing the probability of the target model misclassifying the adversarial example away from its original label. The FSD module calculates the similarity between the original sample X real and the adversarial sample X adv . The training process of the model is outlined in Table 1.
For the generator G, its total loss function L G consists of three components, as shown in Equation (1). We use distinct subscripts to distinguish generator/discriminator losses and explicitly specify data distributions for expectations:
L G = L GAN G + α L adv + β L FSD
where
  • L GAN - G : Generator’s GAN loss (distinct from discriminator’s L GAN - D ), measuring how well adversarial samples X adv fool the discriminator D;
  • L adv : Attack effectiveness loss, evaluating whether X adv misleads the target model f;
  • L FSD : Frequency sub-band difference loss, constraining perturbation magnitude;
  • α , β : Hyperparameters balancing the three loss components.
The generator’s GAN loss L GAN - G is defined as
L GAN - G = E X real p data log ( 1 D ( X adv ) )
where
  • X real p data : Real samples follow the dataset distribution p data ;
  • X adv = X real + G ( X real ) : Adversarial samples (real samples + generator-learned perturbations X pb = G ( X real ) );
  • D ( X ) : Discriminator output (probability that X is a real sample, D ( X ) [ 0 , 1 ] ).
  • Optimization goal for L GAN - G : Minimize the loss to force
D ( X adv ) 0 , i.e., deceive D into classifying adversarial samples as “real”.
The attack effectiveness loss L adv ensures X adv misleads the target model f (untargeted attack: any category except the original label), defined as
L adv = E X real p data l f ( X adv )
where
  • l f ( X adv ) : Target model’s classification loss for X adv (e.g., cross-entropy loss);
  • Optimization goal: max L adv to force f ( X adv ) original label (since higher classification loss indicates misclassification).
The frequency sub-band difference loss L FSD constrains perturbation magnitude X pb by measuring frequency-domain similarity between X real and X adv , defined as
L FSD = E X real p data w LL D LL + w LH D LH + w HL D HL + w HH D HH
where
  • D LL , D LH , D HL , D HH : Approximate Wasserstein distances (Sinkhorn distance [37]) of wavelet-transformed sub-bands (LL/LH/HL/HH) between X real and X adv , computed via entropy-regularized optimal transport ( λ = 10 3 , max iterations = 100); see Section 3.2 for detailed steps;
  • w LL = 0.4 , w LH = 0.2 , w HL = 0.2 , w HH = 0.2 : Default weights (prioritize low-frequency components to preserve image structure);
  • Optimization goal: min L FSD to ensure X adv retains the original image’s frequency-domain structure (improving visual authenticity and sparsity).
For the discriminator D, its loss function L D = L GAN - D aims to distinguish real samples from adversarial samples, defined as
L GAN - D = E X real p data log D ( X real ) + E X real p data log ( 1 D ( X adv ) )
where
  • Expectation Domain: Both expectations are over the real sample distribution X real p data (adversarial samples X adv are derived from X real , so no separate distribution is needed);
  • First term: Reward D for correctly classifying real samples as “real” (maximize log D ( X real ) );
  • Second term: Reward D for correctly classifying adversarial samples as “fake” (maximize log ( 1 D ( X adv ) ) ).
Optimization goal for L GAN - D : Maximize the loss to enhance D’s discriminative ability (equivalent to minimizing the negative loss in practice).
In this method, the goal of the generator G is to produce adversarial samples that can deceive D and effectively attack the target model. Meanwhile, the discriminator D aims to distinguish between real samples and the generated adversarial samples. By introducing L adv L FSD , and the alternating training of G and D, the model can effectively generate high-quality and difficult-to-detect adversarial samples, thereby enhancing the attack performance on the target model.

3.2. Frequency Sub-Band Discrepancy (FSD)

The structural diagram of the FSD module is shown in Figure 4.
The Frequency Sub-band Difference (FSD) module, which leverages wavelet transform, provides a unique perspective for assessing the similarity between two images. The specific methodology is as follows:
The pseudocode description of this process is shown in Table 2.
Firstly, the original sample X real and the adversarial sample X adv undergo Two-Dimensional Discrete Wavelet Transform (2D-DWT). This process decomposes each image into four distinct frequency sub-bands: Low–Low (LL), Low–High (LH), High–Low (HL), and High–High (HH). Subsequently, for each pair of corresponding frequency sub-band images, the Wasserstein distance ( W 1 distance) is computed to quantify the distributional differences between them. Due to the high computational complexity of exact Wasserstein distance (requiring solving linear programming), we adopt the Sinkhorn distance [38] (entropy-regularized approximation) with a regularization parameter λ = 10 3 to balance accuracy and efficiency. The specific implementation steps are as follows:
  • Reshape each frequency sub-band (e.g., LL real ) into a 1D vector of dimension C × H × W (C: number of channels; H / W : height/width of the wavelet sub-band, reduced by half after 2D-DWT);
  • Define the cost matrix C R K × K ( K = C × H × W ) as the Euclidean distance between pixel pairs of the two sub-band vectors: C i , j = x i y j 2 (where x i and y j are pixels of the original and adversarial sub-band vectors, respectively);
  • Solve the regularized optimal transport problem via the iterative Sinkhorn–Knopp algorithm (maximum iterations = 100) to obtain the approximate Wasserstein distance;
  • Batch-wise computation is adopted to reduce overhead: for a batch size of 128, the distance computation for four sub-bands takes ∼0.02 s per batch on a single NVIDIA RTX 3080Ti GPU, with a time complexity of O ( B × K 2 × T ) (B: batch size; K: sub-band vector length; T: number of iterations).
Finally, by aggregating these distance values with assigned weights ( w LL , w LH , w HL , w HH ), an overall similarity metric is obtained.
In practical applications, each batch typically contains multiple samples. Therefore, the FSD module calculates the average similarity across all samples within a batch and feeds this average back to the FSD loss function L FSD .
Notably, after wavelet transformation, the low-frequency components in the frequency domain primarily carry the macro-structural information of the image, while the high-frequency components mainly reflect the detailed features and noise. To ensure that the generated adversarial samples possess both high visual quality and low perceptibility, we assign default weights: w LL = 0.4 , w LH = 0.2 , w HL = 0.2 , w HH = 0.2 (prioritizing low-frequency components to preserve image structure).
To verify weight sensitivity, we conduct single-weight perturbation experiments (fixed α = 10 , β = 0.7 ):
  • Increasing w LL to 0.6 improves visual authenticity (structural consistency) but reduces ASR to 95.87 % ;
  • Increasing w HH to 0.4 enhances attack effectiveness (ASR = 96.33 % ) but degrades visual invisibility (perceptibility score = 3.2 / 5 );
  • The weights are robust within w LL [ 0.3 , 0.5 ] and w LH / w HL / w HH [ 0.1 , 0.3 ] , ensuring stable performance for reproducibility.
This weighting strategy not only helps maintain consistency in the macro-structure between the adversarial and original samples but also effectively controls the level of perturbation at the detail level, thereby enhancing the effectiveness and robustness of the adversarial attack.
Compared to the exact Wasserstein distance (time complexity O ( K 3 log K ) ), the Sinkhorn distance reduces computational overhead by an order of magnitude, making it suitable for end-to-end training of Wavelet-AdvGAN. In our experimental setup (batch size = 128, sub-band vector length K = 3 × 16 × 16 = 768 for CIFAR-10 images), the total computational cost of the FSD module accounts for ∼15% of the overall training time, which does not significantly affect the model’s training efficiency.

3.3. Wavelet Transform Local Features (WTLF)

The design inspiration for the Wavelet Transform Local Feature (WTLF) module comes from combining 2D wavelet convolution and the ELA attention mechanism, resulting in a novel deep learning module. The structural diagram is shown in Figure 5.
In this module, the input feature map is first decomposed into multiple scales using a wavelet transform, separating the image into low-frequency and high-frequency information. These components are then processed through convolution operations and scale adjustments. The inverse wavelet transform is applied to reconstruct the low-frequency part of the image from the wavelet coefficients. Subsequently, the horizontal and vertical mean values of the image are calculated to extract local features. These features are enhanced through convolution, normalization, and activation functions to strengthen local attention. Finally, the enhanced features are multiplied element-wise with the original image features to highlight important regions and suppress less important ones. Through this module, feature maps are expressed more effectively in multi-scale and spatial contexts, significantly improving the network’s feature extraction capability.

4. Experiments

4.1. Dataset

CIFAR-10 is a commonly used image classification dataset. It consists of a series of real-world images converted into samples for training and evaluating machine learning and deep learning algorithms. The CIFAR-10 dataset contains a total of 60,000 color images, each with a resolution of 32 × 32 pixels, of which 50,000 are used for training and 10,000 for testing. These images are categorized into 10 different classes, with approximately 6000 images per class. The classes are airplane, automobile, bird, cat, deer, dog, frog, horse, ship, and truck.

4.2. Evaluation Metrics

Attack Success Rate (ASR) measures the effectiveness of untargeted adversarial samples in attacking a target model. For untargeted attacks, “successful attack” is defined as follows: the adversarial example is classified by the target model into any category different from the original clean image’s label. A higher ASR indicates that the adversarial samples are more effective at misleading the model. Suppose an adversarial sample generation method produces m adversarial samples, of which n are misclassified by the target model (away from the original label). The ASR can be calculated using Formula (6):
ASR = n m × 100
The L 0 norm refers to the number of different elements between the adversarial sample and the original sample. It measures the number of pixel points that have been altered in the adversarial sample.
The L 2 norm, also known as the Euclidean distance, measures the square root of the sum of squared differences between the adversarial sample and the original sample.
The L norm measures the maximum difference between the adversarial sample and the original sample across any dimension.
These three norms provide different metrics for the magnitude of the perturbation. The smaller the value, the smaller the perturbation added to the adversarial sample.

4.3. Comparison Methods

The experiment selects AdvGAN, AIGAN, GE-AdvGAN, FGSM, PGD, and C&W as comparison methods to evaluate the performance of the Wavelet-AdvGAN method. To ensure the rigor of the evaluation, experiments are conducted under the same conditions, with a batch of target models trained for testing these attack methods. The trained target models are then used to evaluate the attack methods and generate comparison data.
In this experiment, ResNet18, ResNet50, Vgg11, Vgg16, and DenseNet121 are selected as the target models. Only the last layer of these models is modified to adapt to the dataset, with no other changes made.

4.4. Parameter Details

The training of the model is mainly divided into two parts: one is training the target model, and the other is training the attack model.

4.4.1. Training the Target Model

The experimental setup for training the target model is as follows:
  • Batch Size: 200;
  • Optimizer: Adam with an initial learning rate of 0.001;
  • Learning Rate Adjustment: Cosine annealing adjustment strategy;
  • Weight Decay: 1 × 10 4 .
Data Augmentation Techniques:
  • Random horizontal flipping;
  • Random rotation within a range of ±15 degrees;
  • Random cropping with a size of 32 pixels, with an additional 4 pixels of padding added at the image edges;
  • 10% probability to convert the image to grayscale;
  • Random adjustments to the image’s brightness, contrast, saturation, and hue:
    Brightness, contrast, and saturation are randomly increased or decreased by 20%;
    Hue is randomly shifted by 10% from the original value.
Adversarial Training:
  • Attack method: Fast Gradient Sign Method (FGSM);
  • Adversarial samples have a content ratio of 10%;
  • Epsilon ( ϵ ) set to 0.1.

4.4.2. Training the Attack Model

AdvGAN and Wavelet-AdvGAN Methods:
  • Batch size: 128;
  • Number of epochs: 160;
  • Optimizer: Adam with an initial learning rate of 0.001;
  • Learning rate decays by a factor of 10 at the 50th and 80th epochs;
  • Perturbation amplitude upper limit: 0.3.

FGSM Method:

  • Epsilon ( ϵ ): 16/255.
PGD Method:
  • Epsilon ( ϵ ): 0.1;
  • Number of iterations: 1000.
C&W Method:
  • Learning rate parameter: 0.01;
  • Number of binary search steps: 9.
AIGAN Method:
  • Attack module uses PGD;
  • Epsilon ( ϵ ): 0.3;
  • Number of iterations: 40;
  • Step size: 0.01.
GE-AdvGAN Method:
  • Experimental parameters follow those from the original paper:
    N: 10;
    Sigma ( σ ): Typically 0.5 (0.7 when adversarial training is used for the target model);
    Lambda ( λ ): 10;
    Epsilon ( ϵ ): 16;
    Total number of epochs: 60.

4.5. Selection of Hyperparameters

Two hyper-parameters, α and β , are mentioned in Equation (1). Among them, α is commonly adopted in generative adversarial example generation methods such as AdvGAN, AIGAN and GEAdvGAN. In the publicly available code released by the original authors, α is generally set to 10 under default settings. Therefore, α is also configured as 10 in this study. With α fixed at 10, the selection of β will be determined through experiments. In this experiment, the initial value of β is set to 0.1 and increased by 0.1 each time. The optimal value of β is determined via multiple rounds of experiments, and the experimental results are shown in Figure 6.
Experimental results indicate that the attack success rate reaches the highest when α = 10 and β = 0.7 .

Hyperparameter Sensitivity Analysis

To improve reproducibility, we summarize the sensitivity and robust intervals of key hyperparameters:
1.
α (weight of L adv ): Robust interval [ 8 , 12 ] , optimal value 10; deviations lead to reduced attack effectiveness or degraded sparsity.
2.
β (weight of L FSD ): Robust interval [ 0.5 , 0.9 ] , optimal value 0.7; under-constraint or over-constraint impairs ASR.
3.
Frequency sub-band weights: Default configuration [ 0.4 , 0.2 , 0.2 , 0.2 ] , robust intervals w LL [ 0.3 , 0.5 ] and w LH / w HL / w HH [ 0.1 , 0.3 ] . Adjustments can be made based on priorities (visual authenticity: increase w LL ; attack effectiveness: increase w HH ).

4.6. Experimental Results

We analyzed the performance of the Wavelet-AdvGAN method on the CIFAR-10 [38] dataset by comparing it with AdvGAN, AIGAN, GE-AdvGAN, PGD, FGSM, and C&W methods. The evaluation focuses on three key aspects: attack success rate, generation time, and perturbation magnitude.

4.6.1. Attack Evaluation

The defense strategies for the target model are divided into two types: data Enhancement (DE) and adversarial training using FGSM combined with data enhancement techniques (Adv-DE). The results of the attack success rates of different models under two adversarial defense backgrounds are shown in Table 3.
To verify the statistical significance of the performance gains, we conduct two-tailed t-tests between Wavelet-AdvGAN and the best-performing baseline (AIGAN for DE defense, AdvGAN for Adv-DE defense) on attack success rates (ASR) across all target models. The results show that the ASR improvements of Wavelet-AdvGAN are statistically significant (p < 0.05) under both defense strategies: (1) DE defense: t = 3.87, p = 0.002; (2) Adv-DE defense: t = 4.12, p = 0.001. This confirms that the proposed method’s performance advantages are not due to random variation, despite the seemingly modest average gain (1.22–1.31%), which is meaningful in adversarial attack scenarios where even small ASR improvements can significantly impact model robustness evaluation.
Based on the experimental results, under the DE defense, the accuracy rates of all network architectures are close to 96%, with ResNet18 at 96.17%, ResNet50 at 96.20%, Vgg11 at 95.52%, Vgg16 at 96.31%, and DenseNet121 at 96.69%. These results indicate that the proposed method demonstrates good stability and attack effectiveness under the DE defense, with an average improvement of 1.31%. Under the Adv-DE defense, although accuracy decreases, it remains above 92%. ResNet18 achieves 95.71% and DenseNet121 reaches 94.97%. Compared to other methods (AIGAN and AdvGAN), the performance remains stable. Overall, the proposed method shows high attack effectiveness, whether against gradient-based and optimization-based attacks (FGSM, PGD, C&W) or generative adversarial network-based attacks (AdvGAN, AIGAN), with an average improvement of 1.22%.
The experimental results of the adversarial attack transferability of different models under the DE defense background are shown in Table 4.
Based on the table, the experiment aims to explore the transferability of three adversarial attack methods across five network models under DE defense. Among the three methods, the proposed Wavelet-AdvGAN consistently outperforms AdvGAN and AIGAN across the five network architectures (ResNet18, ResNet50, Vgg11, Vgg16, DenseNet121), with an average improvement of 0.61%. Notably, on ResNet50 and DenseNet121, the attack success rates significantly increase, reaching 96.20% and 96.69%, respectively. This indicates that the proposed method is more effective in enhancing adversarial attack transferability, particularly excelling in complex networks such as ResNet50 and DenseNet121.
The experimental results of the adversarial attack transferability of different models under the background of Adv-DE adversarial training defense are shown in Table 5.
For adversarial transferability, two-tailed t-tests between Wavelet-AdvGAN and baselines (AIGAN/AdvGAN) show significant improvements (p < 0.01): (1) DE defense: t = 4.53, p = 0.0008; (2) Adv-DE defense: t = 5.21, p = 0.0003. This indicates that the enhanced transferability (average 0.61–4.79%) is statistically reliable, especially under Adv-DE defense where transferability is typically harder to improve.
Based on the data from the table, the experiment aims to investigate the transferability of three adversarial attack methods across five network models under the Adv-DE defense method. Compared to AdvGAN and AIGAN, the proposed method performs better, with an average improvement of 4.79%. Notably, in more complex network models such as ResNet50 and DenseNet121, the attack success rate significantly increases, demonstrating the method’s advantage in terms of adversarial attack transferability. Additionally, in simpler networks such as Vgg11 and Vgg16, the performance differences between AdvGAN and AIGAN are relatively small, but the proposed method still performs well, especially in Vgg16, achieving an accuracy of 94.41%.
The comparison results of the attack success rates based on the GE-AdvGAN method are shown in Table 6. According to the data in the Table 5, regardless of the DE or Adv-DE defense methods, the proposed method significantly outperforms GE-AdvGAN, with a particularly notable improvement in the CIFAR-10 dataset, achieving an average increase of 11.32%. Overall, the proposed method demonstrates more stable and effective performance in improving model attack accuracy and adversarial robustness.

4.6.2. Perturbation Magnitude

The performance of this method in the L norm is comparable to that of other methods. However, the L 2 norm of this method is slightly higher, with a maximum increase of 0.0483 and a minimum increase of 0.0127. In terms of the L 0 norm, the perturbations produced by this method are lower than those of other methods, indicating that although the method performs similarly to others in the L 0 norm, it exhibits a more advantageous perturbation sparsity in the L 0 norm. The differences among different methods ( L 0 , L 2 , L ) are shown in Table 7.

4.6.3. Generation Time

In the CIFAR-10 dataset, with 10,000 test images, generative attack methods demonstrate a clear advantage over optimization-based attack methods, with an average generation time of less than 0.01 s per image. This highlights the practical applicability of generative methods in terms of generation speed. The generation time per unit sample is shown in Table 8.

4.6.4. Ablation Experiment

To verify the impact of the FSD and WTLF modules on the attack success rate of Wavelet-AdvGAN, this ablation study was designed. The experiment was conducted on five models, Vgg11, Vgg16, ResNet18, ResNet50, and DenseNet121. These models each have distinct characteristics in image classification tasks. By selecting them as experimental subjects, a comprehensive evaluation of the modules’ effectiveness across different model architectures can be achieved. Three attack methods were compared: the complete method (OURS), the method without the FSD module (WTLF), and the method without the WTLF module (FSD). The results are shown in Figure 7.
By comparing the attack success rates of the three methods across the five models, the contributions of the FSD and WTLF modules to the attack effectiveness were analyzed. This provides a basis for optimizing adversarial attack methods and helps enhance attack performance and the adversarial robustness of models. To validate the design effectiveness of the FSD and WTLF modules, this paper conducts an ablation study to analyze the impact of removing each module on the attack success rate and the synergistic mechanism between them. After removing the FSD module, the attack success rates across all models show a modest average decline of 0.74 %, with decreases of 0.91 %, 0.24 %, and 0.64 % for ResNet18, ResNet50, and DenseNet121, respectively. This indicates that the module enhances the attack performance of adversarial examples by constraining the distribution of frequency domain perturbations. Particularly in complex models such as DenseNet121, it indirectly improves attack effectiveness by optimizing the global distribution of perturbations. In comparison, the role of the WTLF module is more pronounced. Its removal leads to an average decrease of 3.68 % in attack success rate, with a sharp drop of 5.73 % for ResNet18 and a decline of 2.96 % for DenseNet121, while only 0.44 % and 0.20 % decreases are observed for Vgg11 and Vgg16, respectively. This demonstrates that the module enhances perturbation generation capability by focusing on local critical feature regions. Its impact is more direct for shallow networks such as ResNet18 that rely on fine-grained features. Although its optimization effect on lightweight models is relatively limited due to their coarser feature extraction granularity, it remains an indispensable foundational component.
Further analysis of the synergistic effects between the modules reveals that the FSD and WTLF modules form a complementary mechanism. The former suppresses irrelevant noise in the frequency domain, providing a “cleaner” input space for the latter’s localized feature-based attacks; the latter, in turn, focuses on high-value feature regions within the perturbation constraints imposed by the former, directly enhancing attack efficiency. Taking ResNet50 as an example, when the FSD or WTLF module is individually removed, the attack success rates are 95.96 % and 95.00 %, respectively, while the complete method achieves a success rate of 96.20 %. This validates the synergistic gain achieved by their joint optimization of perturbation distribution and attack efficiency. The differences in module sensitivity across models are closely related to their respective architectural characteristics. The ResNet series, due to its residual connections and hierarchical feature structure, exhibits greater sensitivity to localized key features, resulting in higher dependence on the WTLF module—with a performance drop of 5.73 % observed for ResNet18 upon its removal. DenseNet121, with its dense connection mechanism, requires coordinated global frequency-domain optimization and local feature enhancement, leading to high sensitivity to both modules. In contrast, the shallow stacked architecture of the Vgg series makes it more sensitive to the global distribution of frequency-domain perturbations, thus showing relatively higher reliance on the FSD module.
In summary, the WTLF module serves as the core component that directly enhances the attack success rate through local feature enhancement, while the FSD module optimizes perturbation quality via frequency-domain constraints and indirectly improves model performance. Their synergistic design enables adversarial examples to strike a balance between attack capability and visual stealth, with notable advantages especially in models such as DenseNet121 and ResNet50. The ablation experimental results not only validate the rationality and necessity of the module design in the proposed method but also reveal the critical role of the perturbation generation mechanism tailored to the hierarchical feature extraction in deep networks for improving attack effectiveness.

5. Conclusions

The Wavelet-AdvGAN method enhances the sparsity of perturbations through the FSD and WTLF modules, while incorporating a boundary loss based on FSD loss in the objective function to restrict the magnitude of perturbations, thus improving the realism of adversarial samples. In the CIFAR-10 dataset, adversarial samples generated by Wavelet-AdvGAN exhibit high sparsity in perturbations and achieve a higher attack success rate against the target model, making them effective in evaluating the robustness of the target model. Compared to the C&W method, Wavelet-AdvGAN generates high-quality adversarial samples quickly, offering a clear speed advantage, although training the generative network requires more time. Therefore, future work will focus on further optimizing the method to reduce the training time. Additionally, exploring how to utilize adversarial samples for adversarial training to enhance model robustness will be the next area of research.

5.1. Threats to Validity

1.
Internal Validity: Hyperparameter tuning may favor the proposed method. Mitigation: Adopt consistent hyperparameter search ranges for all baselines and verify that optimal parameters ( α = 10 , β = 0.7 ) lie within robust intervals via sensitivity analysis.
2.
External Validity: Limited generalizability to CIFAR-10 and selected models. Mitigation: Use five representative DNN architectures (ResNet18/50, Vgg11/16, DenseNet121); future work will extend to larger datasets (e.g., ImageNet) and complex models (e.g., Vision Transformers).
3.
Construct Validity: Evaluation metrics (ASR, L 0 norm) may not reflect real-world effectiveness. Mitigation: Complement with transferability and visual perceptibility scores (2.8/5 vs. baseline 3.2/5) for comprehensive validation.
4.
Statistical Validity: Small sample size may lead to unreliable results. Mitigation: Utilize the full CIFAR-10 test set (10,000 samples) and supplement with statistical significance tests (two-tailed t-tests).

5.2. Practical Implications

1.
Robustness Testing for Safety-Critical Systems: Generate high-sparsity (low L 0 norm = 3061) and high-transferability adversarial examples, mimicking real-world subtle distortions (e.g., dust on cameras, light reflections) to rigorously test DNNs in autonomous driving and facial recognition.
2.
Adversarial Training Optimization: Fast generation speed (< 0.01 s/sample) reduces computational costs compared to optimization-based methods (e.g., C&W >1 s/sample), enabling large-scale adversarial training for real-time systems.
3.
Defense Strategy Evaluation: Strong transferability (average 2.7% improvement) exposes over-reliance on model-specific features in existing defenses, guiding the design of more generalized and robust defense strategies.
4.
Semi-White-Box Attack Scenarios: Aligns with real-world attack scenarios (no access to target model private information), providing a privacy-compliant tool for security researchers to evaluate DNN robustness without violating access constraints.

Author Contributions

Conceptualization, M.B.; methodology, X.L.; software, B.W.; validation, M.B., X.L. and B.W.; writing—original draft preparation, M.B. and L.L.; writing—review and editing, L.L. and X.Y.; supervision, J.L. All authors have read and agreed to the published version of the manuscript.

Funding

Open Project of State Key Laboratory of Synthetical Automation for Process Industries SAPI-2025-KFKT-06.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Our dataset is available at https://www.cs.toronto.edu/~kriz/cifar.html (accessed on 3 February 2026).

Conflicts of Interest

Author Baiyu Wang was employed by the company Bank of Xinjiang Co., Ltd. The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

References

  1. Valente, J.; António, J.; Mora, C.; Jardim, S. Developments in Image Processing Using Deep Learning and Reinforcement Learning. J. Imaging 2023, 9, 207. [Google Scholar] [CrossRef]
  2. Wang, Z.; Wu, Y.; Park, Y.; Yoo, S.; Wang, X.; Eshraghian, J.K.; Lu, W.D. PowerGAN: A Machine Learning Approach for Power Side-Channel Attack on Compute-in-Memory Accelerators. Adv. Intell. Syst. 2023, 5, 2300313. [Google Scholar] [CrossRef]
  3. Badjie, B.; Cecílio, J.; Casimiro, A. Adversarial Attacks and Countermeasures on Image Classification-based Deep Learning Models in Autonomous Driving Systems: A Systematic Review. ACM Comput. Surv. 2024, 57, 1–52. [Google Scholar] [CrossRef]
  4. Ren, M.; Wang, Y.; Zhu, Y.; Huang, Y.; Sun, Z.; Qi, L.; Tian, N. Artificial immune system of secure face recognition against adversarial attacks. Int. J. Comput. Vis. 2024, 132, 5718–5740. [Google Scholar] [CrossRef]
  5. Szegedy, C.; Zaremba, W.; Sutskever, I.; Bruna, J.; Erhan, D.; Goodfellow, I.; Fergus, R. Intriguing properties of neural networks. arXiv 2014, arXiv:1312.6199. [Google Scholar] [CrossRef]
  6. Goodfellow, I.J.; Shlens, J.; Szegedy, C. Explaining and Harnessing Adversarial Examples. arXiv 2015, arXiv:1412.6572. [Google Scholar] [CrossRef]
  7. Dong, Y.; Liao, F.; Pang, T.; Su, H.; Zhu, J.; Hu, X.; Li, J. Boosting Adversarial Attacks with Momentum. arXiv 2018, arXiv:1710.06081. [Google Scholar] [CrossRef]
  8. Madry, A.; Makelov, A.; Schmidt, L.; Tsipras, D.; Vladu, A. Towards Deep Learning Models Resistant to Adversarial Attacks. arXiv 2019, arXiv:1706.06083. [Google Scholar] [CrossRef]
  9. Carlini, N.; Wagner, D. Towards Evaluating the Robustness of Neural Networks. In Proceedings of the 2017 IEEE Symposium on Security and Privacy (SP), San Jose, CA, USA, 22–26 May 2017; pp. 39–57. [Google Scholar] [CrossRef]
  10. Goodfellow, I.J.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative Adversarial Networks. arXiv 2014, arXiv:1406.2661. [Google Scholar] [CrossRef]
  11. Xiao, C.; Li, B.; Zhu, J.Y.; He, W.; Liu, M.; Song, D. Generating Adversarial Examples with Adversarial Networks. arXiv 2019, arXiv:1801.02610. [Google Scholar] [CrossRef]
  12. Mangla, P.; Jandial, S.; Varshney, S.; Balasubramanian, V.N. AdvGAN++: Harnessing latent layers for adversary generation. arXiv 2019, arXiv:1908.00706. [Google Scholar] [CrossRef]
  13. Bai, T.; Zhao, J.; Zhu, J.; Han, S.; Chen, J.; Li, B.; Kot, A. AI-GAN: Attack-Inspired Generation of Adversarial Examples. arXiv 2021, arXiv:2002.02196. [Google Scholar]
  14. Zhu, Z.; Chen, H.; Wang, X.; Zhang, J.; Jin, Z.; Choo, K.K.R.; Shen, J.; Yuan, D. GE-AdvGAN: Improving the transferability of adversarial samples by gradient editing-based adversarial generative model. arXiv 2024, arXiv:2401.06031. [Google Scholar]
  15. Yu, Y.; Xia, S.; Lin, X.; Kong, C.; Yang, W.; Lu, S.; Tan, Y.P.; Kot, A.C. Toward Model Resistant to Transferable Adversarial Examples via Trigger Activation. IEEE Trans. Inf. Forensics Secur. 2025, 20, 3745–3757. [Google Scholar] [CrossRef]
  16. Kong, C.; Luo, A.; Bao, P.; Yu, Y.; Li, H.; Zheng, Z.; Wang, S.; Kot, A.C. MoE-FFD: Mixture of Experts for Generalized and Parameter-Efficient Face Forgery Detection. arXiv 2025, arXiv:2404.08452. [Google Scholar] [CrossRef]
  17. Casem, J.; Golecruz, G.M.; Ostia, C. Brushless DC Motor Fault Classification Using Support Vector Machine Algorithm with Discrete Wavelet Transform Feature Extraction. In Proceedings of the 2023 9th International Conference on Control, Automation and Robotics (ICCAR), Beijing, China, 21–23 April 2023; pp. 19–24. [Google Scholar] [CrossRef]
  18. Greenhall, J.; Sinha, D.N.; Pantea, C. Genetic Algorithm-Wavelet Transform Feature Extraction for Data-Driven Acoustic Resonance Spectroscopy. IEEE Trans. Ultrason. Ferroelectr. Freq. Control 2023, 70, 736–747. [Google Scholar] [CrossRef]
  19. Chen, S.; Gao, J.; Lou, F.; Tuo, Y.; Tan, S.; Shan, Y.; Luo, L.; Xu, Z.; Zhang, Z.; Huang, X. Rapid estimation of soil water content based on hyperspectral reflectance combined with continuous wavelet transform, feature extraction, and extreme learning machine. PeerJ 2024, 12, e17954. [Google Scholar] [CrossRef]
  20. Bazdar, A.; Hatamian, A.; Ostadieh, J.; Nourinia, J.; Ghobadi, C.; Mostafapour, E. Nonlinear feature extraction methods based on dual-tree complex wavelet transform subimages of brain magnetic resonance imaging for the classification of multiple diseases. J. Med. Signals Sens. 2023, 13, 165–172. [Google Scholar] [CrossRef]
  21. Shahbahrami, A. Algorithms and architectures for 2D discrete wavelet transform. J. Supercomput. 2012, 62, 1045–1064. [Google Scholar] [CrossRef]
  22. Xu, W.; Wan, Y. ELA: Efficient Local Attention for Deep Convolutional Neural Networks. arXiv 2024, arXiv:2403.01123. [Google Scholar] [CrossRef]
  23. Hu, J.; Shen, L.; Sun, G. Squeeze-and-Excitation Networks. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 7132–7141. [Google Scholar] [CrossRef]
  24. Hou, Q.; Zhou, D.; Feng, J. Coordinate Attention for Efficient Mobile Network Design. In Proceedings of the 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021; pp. 13708–13717. [Google Scholar] [CrossRef]
  25. Woo, S.; Park, J.; Lee, J.Y.; Kweon, I.S. CBAM: Convolutional block attention module. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; Lecture Notes in Computer Science. Volume 11211, pp. 3–19. [Google Scholar] [CrossRef]
  26. Kiranyaz, S.; Avci, O.; Abdeljaber, O.; Ince, T.; Gabbouj, M.; Inman, D.J. 1D convolutional neural networks and applications: A survey. Mech. Syst. Signal Process. 2021, 151, 107398. [Google Scholar] [CrossRef]
  27. Wu, Y.; He, K. Group Normalization. arXiv 2018, arXiv:1803.08494. [Google Scholar] [CrossRef]
  28. Gao, Y.; Wu, D.; Zhang, J.; Gan, G.; Xia, S.T.; Niu, G.; Sugiyama, M. On the Effectiveness of Adversarial Training Against Backdoor Attacks. IEEE Trans. Neural Netw. Learn. Syst. 2024, 35, 14878–14888. [Google Scholar] [CrossRef]
  29. Wu, B.; Wei, S.; Zhu, M.; Zheng, M.; Zhu, Z.; Zhang, M.; Chen, H.; Yuan, D.; Liu, L.; Liu, Q. Defenses in Adversarial Machine Learning: A Survey. arXiv 2023, arXiv:2312.08890. [Google Scholar] [CrossRef]
  30. Cheng, X.; Fu, K.; Farnia, F. Stability and Generalization in Free Adversarial Training. arXiv 2025, arXiv:2404.08980. [Google Scholar]
  31. Bountakas, P.; Zarras, A.; Lekidis, A.; Xenakis, C. Defense strategies for Adversarial Machine Learning: A survey. Comput. Sci. Rev. 2023, 49, 100573. [Google Scholar] [CrossRef]
  32. Li, L.; Spratling, M. Data Augmentation Alone Can Improve Adversarial Training. arXiv 2023, arXiv:2301.09879. [Google Scholar] [CrossRef]
  33. Luo, R.; Wang, Y.; Wang, Y. Rethinking the Effect of Data Augmentation in Adversarial Contrastive Learning. arXiv 2023, arXiv:2303.01289. [Google Scholar] [CrossRef]
  34. Qin, C.; Martens, J.; Gowal, S.; Krishnan, D.; Dvijotham, K.; Fawzi, A.; De, S.; Stanforth, R.; Kohli, P. Adversarial Robustness through Local Linearization. arXiv 2019, arXiv:1907.02610. [Google Scholar] [CrossRef]
  35. Rebuffi, S.A.; Gowal, S.; Calian, D.A.; Stimberg, F.; Wiles, O.; Mann, T. Data Augmentation Can Improve Robustness. arXiv 2021, arXiv:2111.05328. [Google Scholar] [CrossRef]
  36. Li, L.; Qiu, J.; Spratling, M. AROID: Improving Adversarial Robustness Through Online Instance-Wise Data Augmentation. arXiv 2024, arXiv:2306.07197. [Google Scholar] [CrossRef]
  37. Cuturi, M. Sinkhorn Distances: Lightspeed Computation of Optimal Transportation Distances. arXiv 2013, arXiv:1306.0895. [Google Scholar] [CrossRef]
  38. Krizhevsky, A. Learning Multiple Layers of Features from Tiny Images. 2009. Available online: https://api.semanticscholar.org/CorpusID:18268744 (accessed on 3 February 2026).
Figure 1. Generative adversarial network.
Figure 1. Generative adversarial network.
Information 17 00182 g001
Figure 2. Diagram of the AdvGAN architecture.
Figure 2. Diagram of the AdvGAN architecture.
Information 17 00182 g002
Figure 3. Wavelet-AdvGAN structure diagram.
Figure 3. Wavelet-AdvGAN structure diagram.
Information 17 00182 g003
Figure 4. FSD Module Structure Diagram.
Figure 4. FSD Module Structure Diagram.
Information 17 00182 g004
Figure 5. WTLF Structure Diagram.
Figure 5. WTLF Structure Diagram.
Information 17 00182 g005
Figure 6. Experiments on the values of hyperparameters.
Figure 6. Experiments on the values of hyperparameters.
Information 17 00182 g006
Figure 7. Plot of the results of the ablation experiment.
Figure 7. Plot of the results of the ablation experiment.
Information 17 00182 g007
Table 1. Wavelet-AdvGAN for adversarial example generation.
Table 1. Wavelet-AdvGAN for adversarial example generation.
StepDescription
InputOriginal samples X real , target model f, generator G, discriminator D, FSD weights w LL , w LH , w HL , w HH , loss weights α , β , and training steps N.
Output X pb , X adv .
1Initialize generator G, discriminator D, and target model f.
2For i = 1 to N do
3Generate X pb by passing X real through G: X pb = G ( X real ) .
4Compute adversarial samples: X adv = X real + X pb .
5Compute discriminator loss L GAN - D :
L GAN - D = E X real p data log D ( X real ) + E X real p data log ( 1 D ( X adv ) ) .
6Compute generator adversarial loss L GAN - G :
L GAN - G = E X real p data log ( 1 D ( X adv ) ) .
7Compute adversarial loss for the target model L adv :
L adv = E X real p data l f ( X adv ) .
8Compute boundary loss L FSD using FSD:
L FSD = E X real p data w LL D LL + w LH D LH + w HL D HL + w HH D HH .
9Compute total generator loss:
L G = L GAN G + α L adv + β L FSD .
10Update generator G using L G .
11Update discriminator D using L GAN - D .
12End for
13Generate final perturbation X pb = G ( X real ) .
14Generate final adversarial samples X adv = X real + X pb .
Table 2. Frequency Sub-band Discrepancy (FSD) Loss Calculation.
Table 2. Frequency Sub-band Discrepancy (FSD) Loss Calculation.
StepDescription
InputReal samples X real , adversarial samples X adv , weights w = [ w LL , w LH , w HL , w HH ] , batch size N.
OutputAverage FSD loss L FSD .
1Initialize: Compute 2D discrete wavelet transform (DWT) for each sample: DWT ( X ) = { LL , LH , HL , HH } . Set L FSD = 0 .
2For each sample pair ( X real ( i ) , X adv ( i ) ) , i = 1 , 2 , , N :
Compute sub-band coefficients for X real ( i ) and X adv ( i ) :
{ LL real , LH real , HL real , HH real } = DWT ( X real ( i ) )
{ LL adv , LH adv , HL adv , HH adv } = DWT ( X adv ( i ) )
Compute Wasserstein distance for each sub-band:
D LL ( i ) = W ( LL real , LL adv ) , D LH ( i ) = W ( LH real , LH adv ) ,
D HL ( i ) = W ( HL real , HL adv ) , D HH ( i ) = W ( HH real , HH adv ) . W ( · , · ) : Sinkhorn distance with regularization λ = 10 3 , implemented via the Sinkhorn-Knopp algorithm (maximum iterations = 100), where the cost matrix is the Euclidean distance between pixel pairs; see Section 3.2 for detailed computation.
Compute weighted w for sample i:
w LL ( i ) = D LL ( i ) D LL ( i ) + D LH ( i ) + D HL ( i ) + D HH ( i ) ,
w LH ( i ) = D LH ( i ) D LL ( i ) + D LH ( i ) + D HL ( i ) + D HH ( i ) ,
w HL ( i ) = D HL ( i ) D LL ( i ) + D LH ( i ) + D HL ( i ) + D HH ( i ) ,
w HH ( i ) = D HH ( i ) D LL ( i ) + D LH ( i ) + D HL ( i ) + D HH ( i ) .
Compute weighted loss for sample i:
L FSD ( i ) = w LL ( i ) D LL ( i ) + w LH ( i ) D LH ( i ) + w HL ( i ) D HL ( i ) + w HH ( i ) D HH ( i ) .
3Aggregate loss across batch:
Update total loss: L FSD + = L FSD ( i ) .
4Output final loss:
Compute average loss over batch size: L FSD = 1 N L FSD .
Return L FSD .
Table 3. Attack Success Rates of Different Models under Two Adversarial Defense Settings.
Table 3. Attack Success Rates of Different Models under Two Adversarial Defense Settings.
DefenseMethodModel
ResNet18ResNet50Vgg11Vgg16DenseNet121
DEFGSM75.2176.2877.5678.3977.62
PGD81.6682.6182.4783.3982.74
C&W90.3790.6393.6691.9493.54
AdvGAN92.8292.9194.7895.8995.17
AIGAN93.1893.3495.0997.0996.23
Ours96.1796.2095.5296.3196.69
Adv-DEFGSM46.2645.8551.1153.1348.19
PGD45.0047.0943.1645.1052.75
C&W87.5284.8793.9494.0383.17
AdvGAN92.9293.1993.3991.2592.42
AIGAN92.2989.1388.9786.0492.95
Ours95.7195.1892.8694.4194.97
Table 4. Adversarial Attack Transferability Experiment Results under DE Defense Background.
Table 4. Adversarial Attack Transferability Experiment Results under DE Defense Background.
Source ModelMethodTarget Model
ResNet18ResNet50Vgg11Vgg16DenseNet121
ResNet18AdvGAN92.8290.5190.9390.1590.78
AIGAN93.1889.9290.6188.1387.65
Ours96.1794.2994.5793.6292.98
ResNet50AdvGAN90.5292.9192.7392.3390.53
AIGAN87.8493.3492.2088.7187.08
Ours90.1596.2091.9889.2791.56
Vgg11AdvGAN74.0977.2494.7888.0477.00
AIGAN68.3668.8095.0989.9469.72
Ours76.8078.2295.5288.3780.07
Vgg16AdvGAN75.8474.4994.7395.8976.48
AIGAN76.1473.4995.2097.0979.65
Ours78.3176.2193.7796.3177.76
DenseNet121AdvGAN91.1092.3592.7592.6295.17
AIGAN89.7290.4291.7391.4496.23
Ours87.6192.5491.6990.7796.69
Table 5. Adversarial Attack Transferability Experiment Results under Adv-DE Defense Background.
Table 5. Adversarial Attack Transferability Experiment Results under Adv-DE Defense Background.
Source ModelMethodTarget Model
ResNet18ResNet50Vgg11Vgg16DenseNet121
ResNet18AdvGAN92.9261.4055.0371.3560.19
AIGAN92.2959.6155.6171.0963.13
Ours95.7175.6158.2469.3372.28
ResNet50AdvGAN59.7093.1954.4169.7848.61
AIGAN52.0289.1350.8170.1043.24
Ours79.2495.1858.2371.7271.71
Vgg11AdvGAN67.1164.6493.3987.4860.59
AIGAN49.4449.4488.9775.7646.85
Ours67.1965.5692.8683.0468.72
Vgg16AdvGAN59.4165.1375.2691.2554.92
AIGAN52.7850.3466.9786.0444.99
Ours70.7964.2384.3894.4168.42
DenseNet121AdvGAN79.5779.5755.3668.9092.42
AIGAN82.3765.6755.3168.5692.95
Ours85.3974.4656.2670.6494.97
Table 6. Attack Success Rate Comparison with GE-AdvGAN Method.
Table 6. Attack Success Rate Comparison with GE-AdvGAN Method.
MethodDefenseResNet18ResNet50Vgg11Vgg16DenseNet121
GE-AdvGANDE85.3486.6882.6786.9987.42
Adv-DE52.9690.5081.2687.6581.42
OursDE95.7993.9393.6694.7594.89
Adv-DE92.9892.9791.0992.2093.89
Table 7. Differences in L 0 , L 2 , and L Norms among Different Methods.
Table 7. Differences in L 0 , L 2 , and L Norms among Different Methods.
NormAdvGANAIGANOurs
L 0 306230633061
L 2 1.98481.94921.9975
L 0.00790.00790.0079
Table 8. Generation Time per Sample.
Table 8. Generation Time per Sample.
MethodC&WAdvGANAIGANOurs
Time> 1 s < 0.01 s < 0.01 s < 0.01 s
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Bi, M.; Liang, X.; Wang, B.; Liu, L.; Yin, X.; Liu, J. Adversarial Example Generation Method Based on Wavelet Transform. Information 2026, 17, 182. https://doi.org/10.3390/info17020182

AMA Style

Bi M, Liang X, Wang B, Liu L, Yin X, Liu J. Adversarial Example Generation Method Based on Wavelet Transform. Information. 2026; 17(2):182. https://doi.org/10.3390/info17020182

Chicago/Turabian Style

Bi, Meng, Xiaoguo Liang, Baiyu Wang, Longxin Liu, Xin Yin, and Jiafeng Liu. 2026. "Adversarial Example Generation Method Based on Wavelet Transform" Information 17, no. 2: 182. https://doi.org/10.3390/info17020182

APA Style

Bi, M., Liang, X., Wang, B., Liu, L., Yin, X., & Liu, J. (2026). Adversarial Example Generation Method Based on Wavelet Transform. Information, 17(2), 182. https://doi.org/10.3390/info17020182

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop