Next Article in Journal
The IDRE Dataset in Practice: Training and Evaluation of Small-to-Medium-Sized LLMs for Empathetic Rephrasing
Previous Article in Journal
TTBCA+: An Enhanced Besiege and Conquer Algorithm with Three-Three Tactics and Collision Theory for Complex Optimization Problems
Previous Article in Special Issue
Real-Time Occluded Target Detection and Collaborative Tracking Method for UAVs
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

An Adaptive JPEG Steganography Algorithm Based on the UT-GAN Model

1
School of Computer Science, Hunan University of Technology and Business, Changsha 410205, China
2
School of Advanced Interdisciplinary Studies, Hunan University of Technology and Business, Changsha 410205, China
*
Author to whom correspondence should be addressed.
Electronics 2025, 14(20), 4046; https://doi.org/10.3390/electronics14204046
Submission received: 29 July 2025 / Revised: 1 October 2025 / Accepted: 8 October 2025 / Published: 15 October 2025
(This article belongs to the Special Issue Digital Intelligence Technology and Applications, 2nd Edition)

Abstract

Adversarial examples pose severe challenges to information security, as their impacts directly extend to steganography and steganalysis technologies. This scenario, in turn, has further spurred the research and application of adversarial steganography. In response, we propose a novel adversarial embedding scheme rooted in a hybrid, partially data-driven approach. The proposed scheme first leverages an adversarial neural network (UT-GAN, Universal Transform Generative Adversarial Network) to generate stego images as a preprocessing step. Subsequently, it dynamically adjusts the cost function with the aid of a DCTR (Discrete Cosine Transform Residual)-based gradient calculator to optimize the images, ensuring that the final adversarial images can resist detection by steganalysis tools. The encoder in this scheme adopts a unique architecture, where its internal parameters are determined by a partially data-driven mechanism. This design not only enhances the capability of traditional steganography schemes to counter advanced steganalysis technologies but also effectively reduces the computational overhead during stego image generation.

1. Introduction

Steganography is a technology that hides confidential digital information in carrier media in an imperceptible way. In recent years, most steganography methods have adopted a framework centered around minimizing distortion [1]. Within this paradigm, the precise formulation of embedding cost is essential, as it quantifies the level of alteration or impairment introduced when concealing data within fundamental units of the image, including pixels or discrete cosine transform (DCT) coefficients. Typically, regions with complex image textures are assigned higher embedding costs, while those with simple and flat textures incur lower costs. These spatially varying embedding costs have a significant impact on the security performance of image steganography methods. Once the embed costs are defined, the Syndrome Trellis Codes (STC) [2] can be used to embed secret messages, aiming to minimize the total embedding cost. Over the years, numerous conventional steganographic techniques have been introduced, including HUGO, WOW, HILL, S-UNIWARD, MiPOD, and CMD [3,4,5,6,7,8]. To illustrate, WOW employs directional high-pass filters to establish spatial additive distortion, whereas HILL designs customized filter sets to better manage the dispersion of distortions in areas with complex textures.
In contrast, steganalysis focuses on detecting stego images that may carry concealed data. Classical detection approaches, like SPAM [9] and SRM [10], commonly convert input images into multiple residual domains and derive high-dimensional statistical [11] attributes to identify alterations caused by steganographic operations. More recently, convolutional neural network (CNN)-based steganalysis techniques [12], including Xu-Net [13], Yedrouj-Net [14], SRNet [15], and CovNet [16], have demonstrated substantial advancements compared to conventional methods. Steganography and steganalysis complement each other, engaged in a continuous technological arms race. While conventional feature-based steganalyzers (e.g., DCTR) exhibit limitations in handling adversarial embeddings due to their non-end-to-end architecture and manually designed features, recent studies have attempted to adapt them to adversarial environments by incorporating advanced training strategies or hybrid frameworks. Recent advances have shown that neural network models mimicking feature-based steganalyzers offer an effective solution to this limitation.
Researchers have recently proposed some deep learning-based steganography methods, mainly categorized into adversarial attack-based [17,18,19,20,21,22] and GAN-based [23,24,25] approaches. Drawing inspiration from the Fast Gradient Sign Method [6], a number of adversarial attack-driven techniques—including adva-emb [17], JS-IAE [18], MAE [19], minmaxstrategy [26], and Back-pack [27,28] —have been developed to optimize initial embedding costs through the use of pretrained convolutional neural network steganalyzers. Early studies on steganographic cost functions also laid a foundation for these methods. The advent of generative artificial intelligence has catalyzed the emergence of an entirely new paradigm. DBS [29] represents a fundamental shift toward a generative framework, wherein secret information is embedded by conditioning the denoising process of a diffusion model to directly synthesize a photorealistic stego image concealing the embedded data, rather than modifying a pre-existing cover medium.
However, although these adversarial and generative approaches incorporate advanced steganographic cost functions and deep learning-based detectors, they have not yet demonstrated substantial performance improvements over conventional steganography techniques in terms of empirical security. With the emergence of GANs, researchers have started integrating them into image steganography to seek breakthroughs. As a pioneering work in this field, ASDL-GAN pioneered the integration of generative adversarial networks into the domain of image steganography, yet its security performance still lags behind traditional methods such as HILL [30]. Recent advances in steganalysis metrics [31] further validate this performance gap. UT-GAN improves on ASDL-GAN [24] to achieve significant performance gains over traditional steganography methods, despite sharing the same end-to-end, data-driven paradigm.
This paper presents a novel adversarial embedding scheme based on steganalysis features, which employs a hybrid (partially data-driven) approach. While the UT-GAN network used for preprocessing is trained in a data-driven manner, the core adversarial gradient generation mechanism—i.e., the gradient calculator—operates in a non-data-driven fashion. Specifically, the linear mapper within the gradient calculator derives its internal parameters by solving equation systems built from individual cover–stego pairs, instead of training on large-scale datasets. This structure reduces the dependency on voluminous data, enhances the resistance to steganalytic detection, and strengthens adversarial stego images by minimizing their feature-space divergence from cover images, as quantified by the NN-DCTR metric. The key contributions of this study are outlined below.
  • Combining non-data-driven gradient generation with UT-GAN preprocessing, the scheme reduces reliance on datasets and enhances the robustness of adversarial stego images.
  • Presenting a new gradient calculator that computes adversarial gradients in place of steganalyzers based on deep learning. The gradient calculator is tailored to DCTR steganalysis, serving as an approximate neural network for the extraction of stego features.
  • Striking a balance between security and implementation efficiency. The UT-GAN adversarial framework is utilized to generate adversarial stego images. To achieve optimal results, the embedding cost function is iteratively refined through the use of an embedding simulator.

2. Related Work

With the swift advancement of deep learning methodologies, its applications have significantly expanded, thereby propelling the development of image steganography. Recently, the application of adversarial attacks to image steganography has emerged as a focus of interest, stemming from similarities in technical implementation to adversarial sample generation. Using the Fast Sparse Symbol Method, Zhang et al. [21] suggested creating adversarial examples from cover images, which complicates the task for steganalyzers to distinguish between cover and secret images. Zhou et al. [23] introduced a framework for efficiently generating large-scale adversarial cover images by training a generator to convert cover images into adversarial samples via simple forward propagation, which are subsequently used for steganography. While these approaches have incorporated adversarial attack techniques into image steganography, they mainly focus on converting cover images into adversarial samples rather than achieving end-to-end image steganography.
AdvSGAN [32], inspired by the adversarial perturbation generation model known as AdvGAN, employs a deep neural network architecture with a constrained neural encoder to advance image steganography learning frameworks. The architecture integrates two adversarial components: a discriminator co-trained with the encoder, and a pretrained target model. They collaboratively develop the steganography from first principles. Mirroring the adversarial training dynamic between generators and discriminators in GANs, steganography and steganalysis exhibit mutual improvement. This collaborative dynamic has driven the broad adoption of GAN frameworks in contemporary steganography to strengthen security. Next, we present a brief summary of ASDL-GAN and UT-GAN.
Early studies pioneered the integration of GANs into steganography. For example, one framework [33] incorporates a Kirchhoff–Volterra (KV) high-pass filter kernel into the embedding process, with the kernel defined based on steganographic distortion criteria. The KV kernel is expressed as
K kv = 1 2 2 2 1 2 6 8 6 2 2 8 12 8 2 2 6 8 6 2 1 2 2 2 1
This framework adjusts probability mappings for the backpropagation process within the GAN structure. The generator’s loss function directly influences the discriminator’s decision-making logic. Nevertheless, ASDL-GAN’s security performance still lags behind methods like those proposed in [34], which adopts a more optimized architecture for steganographic embedding.
Furthermore, a trainable Double-Tanh function was developed to replace the TES subnet as an embedding simulator. This modification eliminates the need for pretraining and significantly reduces computational overhead. In terms of security performance, when facing detection by advanced steganalyzers such as SRM and maxSRMD2, UT-GAN’s error rate is approximately 0.16% lower than that of ASDL-GAN, indicating stronger resistance to detection. In terms of adaptability, UT-GAN can better learn content-adaptive features during training: complex texture regions in images have higher embedding probabilities, meaning secret messages can be embedded more effectively and securely in areas less likely to attract steganalytic attention. This characteristic makes UT-GAN more effective in practical applications, such as scenarios where information concealment must be maintained while preserving the integrity of the carrier image. Given these outstanding performance improvements, UT-GAN was selected as the adversarial network model in this study.

3. Proposed Approach

3.1. Optimal Selection of Compliant Stego Images

Derived from the Fast Gradient Sign (FGS) technique [35], the ADV-EMB approach serves as a significant framework for producing adversarial samples extensively utilized in tasks related to image classification. FGS operates by applying subtle perturbations to input data, generating adversarial instances that deceive neural network models—including convolutional neural networks (CNNs)—into producing erroneous outputs, into misclassifying images. Its application, for instance, entails creating an adversarial example K that can cause a CNN to incorrectly categorize an image X as the wrong label y, despite minimal perturbations.
However, the FGS technique requires adaptation for adversarial embedding tasks, especially when altering the steganographic content. The key adaptation involves regulating steganographic modifications based on the sign of adversarial gradients, implemented through ±1 cost adjustments. The primary goal of ADV-EMB is to synthesize a stego image Z from a cover image C capable of misleading a convolutional neural network (CNN)-based steganalysis system.
Firstly, this section compares ADV-EMB [36] with the proposed adversarial embedding scheme SE, highlighting their shared non-data-driven nature and two key differences. The first lies in the adversarial model employed in their design; the second relates to the criteria for selecting optimal samples. SE utilizes a gradient calculator, while ADV-EMB employs a CNN-based steganalysis model. It serves as a form of steganalysis despite not being used explicitly for such purposes. Its architecture comprises NN-DCTR as a feature extractor and a linear mapper as a classifier, inspired by feature-based steganalysis to eliminate dataset dependency. The internal weights of NN-DCTR are manually configured based on DCTR settings, while those of the linear mapper adapt dynamically to the processed cover–stego pair. DCTR was selected for its balance of low computational complexity and adequate detectability. Alternative feature extractors like GFR [37] and PHARM [38], could also implement the gradient calculator, although at a higher computational cost. Furthermore, the transferability of adversarial examples further improves robustness against diverse steganalysis systems. Despite being an empirically designed adversary, the gradient calculator significantly strengthens the anti-detection ability of traditional feature-based steganographic methods.
Secondly, the criteria for optimal sample selection differ between SE and ADV-EMB. ADV-EMB’s security performance is directly governed by adjustment cost parameters through a two-phase process: (1) generating multiple adversarial stego images with varying adjustment costs, and (2) selecting the optimal sample that simultaneously maximizes adversarial deception while minimizing adjustment costs. This selection process is illustrated in Figure 1, where the optimal sample is chosen to be as close as possible to cover image C. As depicted in Figure 1a, among candidate stego images (e.g., Z1, Z2, and Z3), Z1 is chosen as the optimal output Z because it resides within the adversarial camouflage region (ensuring misclassification) while maintaining the closest proximity to the hyperplane (minimizing adjustment costs).
By contrast, SE uses universal selection criteria instead of relying on steganalyzer-specific knowledge or dataset-dependent parameters. SE’s criteria focus on minimizing the feature–space distance between stego and cover images. Specifically, the weights of the linear mapper in SE are computed using individual cover–stego images. Among candidate stego images (Z1, Z2, and Z3), Z1 is selected as it minimizes ‖Z-C‖, adhering to the statistical principle that distribution proximity correlates with higher likelihood.
A summary of the key differences between ADV-EMB and SE is provided in Table 1.
While the architectural and methodological distinctions between ADV-EMB and SE are clearly outlined in Table 1 and the preceding discussion, it is important to address the absence of direct experimental comparison between the two approaches. This omission stems primarily from practical implementation challenges rather than conceptual oversight.
The original ADV-EMB publication presents a theoretical framework for adversarial embedding but does not provide a complete, reproducible implementation with specified hyperparameters and adjustment strategies. This lack of standardized implementation makes a fair and consistent comparison problematic, as results could be significantly influenced by implementation-specific choices rather than fundamental methodological differences.
To alleviate this limitation, while still being able to provide a meaningful performance context, the following measures were taken.
  • Indirect Performance Benchmarking: The security performance of our SE method is thoroughly evaluated against a comprehensive suite of modern steganalyzers (DCTR, SRM, GFR, XuNet, SRNet, CovNet) in Section 4. The consistently high error rates ( P e ) achieved across these diverse detectors demonstrate SE’s robust anti-detection capability, which conceptually surpasses the security level anticipated from the ADV-EMB framework given its dataset-dependent nature.
  • Theoretical Superiority Validation: The fundamental advantages of SE’s design—particularly its non-data-driven gradient calculator and universal selection criterion—provide theoretical grounding for its superior generalizability and robustness compared to ADV-EMB’s model-specific approach.
  • Future Comparative Commitment: We acknowledge the value of empirical comparison and commit to conducting a comprehensive experimental analysis against ADV-EMB in subsequent work, pending the development of a standardized implementation benchmark.
This multi-faceted approach ensures a rigorous evaluation of SE’s capabilities while maintaining scientific integrity in the absence of directly comparable benchmarks.

3.2. Gradient Calculator

The gradient calculator is designed as a function that maps images to feature vectors. By integrating NN-DCTR with a linear mapper, it generates gradients that are both adversarial and structured to challenge steganalytic models that rely on frequency-based patterns or spatial anomalies. Specifically, f N N D C T R : { 1023 , 1022 , , 1024 } H W R 1 8000 . The linear mapper then further maps these vectors to scalars, i.e., R ( 1 8000 ) R . Figure 2 illustrates the process of constructing the gradient calculator. Initially, f N N D C T R ( C ) and f N N D C T R ( S ) are extracted from C and S, respectively. Subsequently, the weights of the linear mapper are computed using f N N D C T R ( C ) and f N N D C T R ( S ) .

3.3. NN-DCTR

NN-DCTR is a neural network-based approximation of DCTR (Discrete Cosine Transform Representation), designed to be fully differentiable, which allows for the calculation of adversarial gradients. Each operation in NN-DCTR mirrors or approximates a corresponding operation in DCTR, ensuring alignment with the core processing steps of the original DCTR method. NN-DCTR replicates the structure of DCTR through five core operations, each of which aligns with its DCTR counterpart, maintaining consistency in feature extraction while enabling differentiability. Figure 3 illustrates the block diagram of NN-DCTR.
Step 1: 8 × 8 Block IDCT
This step mirrors the transformation procedure utilized in DCTR, wherein the matrix of DCT coefficients, denoted as X, is inversely transformed to produce a corresponding spatial domain representation, designated as W. Ignoring direct current bias and dequantization, Equation (2) defines the function for the 8 × 8 Discrete Cosine Transform block transformation.
w i , j = c u = 0 7 v = 0 7 x u , v cos ( 2 i + 1 ) u π 16 cos ( 2 j + 1 ) v π 16
where x u , v represents the elements of the 8×8 block DCT coefficient matrix X 8 , and w i , j denotes the elements of the 8 × 8 spatial matrix W 8 . Parameter c, determined by Equation (3), defines a constant value.
c ( u ) = 2 4 , u = 0 1 2 , u 0
This translation clarifies the mathematical process of transforming DCT coefficients into spatial domain matrices within the context of DCTR.
Since the inverse Discrete Cosine Transform operates on discrete 8 × 8 blocks independently within the DCT coefficient matrix X, it is particularly suitable for designing adversarial gradients. The gradient of each 8 × 8 block in the DCT coefficient matrix X is denoted as J X 8 , with each element represented by J x u v . Similarly, the spatial matrix W 8 has its gradient represented as J W 8 , with individual elements denoted as J w i j . According to the chain rule of differentiation, J w u v can be derived from Equation (4):
J w u v = i , j w i j x u v J w i j
To enable gradient propagation through the IDCT step, we apply the chain rule of differentiation.
Step 2: Non-matching DCT Calculation
In this step, the non-matching Discrete Cosine Transform (DCT) is computed from the spatial matrix W. This operation is functionally equivalent to convolving the residual signal with a set of 64 distinct DCT kernels, each 8 × 8 in size, within the DCT domain. These kernels originate from the core DCT basis functions defined in Equations (2) and (4). The kernel corresponding to frequency coordinates (u,v) is formally represented as B u v , and its elements are calculated using Equation (5):
B i , j u , v = c ( u ) c ( v ) cos ( i + 0.5 ) π 8 u cos ( j + 0.5 ) π 8 v
The DCT basis kernels are used to compute residuals in the DCT domain, adopting convolutional operations similar to those in CNNs, where the kernel is fixed on B u v .
Step 3: Residual Filtering
This step applies a filtering operation to the residuals in order to enhance the discriminative capability of histogram features within the NN-DCTR framework, as formally expressed in Equation (6). The operation yields a filtered output residual r’.
r = truncT round r q
where round(·) is the rounding function, truncT(.) is the truncation function (defined in Equation (7)), and q is the quantization step (set to 8, as shown in Equation (7)).
q = 8 u ( 2 Q 50 ) , Q { 50 , 51 , , 99 }
Since the rounding function is non-differentiable, we replace it with a continuous approximation, Equation (8), to enable gradient flow during backpropagation:
r = truncT r q
Step 4: Phase Splitting
This step splits the filtered residual matrix into 64 sub-matrices 8 × 8 in size (consistent with DCT operations in DCTR). The unmodified residual matrix is denoted as U, and the sub-matrix at position ( k , l ) of the ( a , b ) -th phase is denoted as U ( a , b ) ( k , l ) . Equation(9) defines the splitting function, where a , b , k , l , i , j { 0 , 1 , , 7 } :
U a , b k , l ( i , j ) = U ( k , l ) ( i + 8 × a , j + 8 × b )
Step 5: Gaussian Histogram Calculation
This step involves approximating histogram features for the extracted sub-matrices from the previous operation. Since direct histogram computation is non-differentiable and cannot be directly applied to fully differentiable structures, a Gaussian histogram layer is utilized. Introduced in [32], the Gaussian histogram layer is an integral component of neural network architectures designed to optimize the PSRM kernel [39]. It employs cumulative Gaussian functions to approximate histogram distributions.

3.4. Linear Mapper

The core component of the gradient calculator is the linear mapping function, defined by Equation (10).
w f NNDCTR ( X ) T + b
where w (weight vector) and b (bias term) are parameters of the linear mapper. Since f N N D C T R ( X ) R 18000 , it follows that w R 18000 . To establish the decision boundary between cover and stego images, the bias term b is determined using Equation (11), ensuring the hyperplane passes through the midpoint of f N N D C T R ( C ) and f N N D C T R ( S ) :
w = f NNDCTR ( C ) f NNDCTR ( S ) b = w f NNDCTR ( C ) + f NNDCTR ( S ) 2 T

3.5. The SE Implementation

In typical image steganography, multiple candidate stego images are first generated using different embedding costs, after which the solution minimizing the distance to the cover image is selected as optimal. The proposed adversarial steganography algorithm SE first utilizes the UT-GAN adversarial network for pretraining to generate adversarial stego images with relatively enhanced security performance. Subsequently, adversarial gradients are computed multiple times based on different embedding rates. Additionally, it replaces the TES subnet with a Double-Tanh function as an embedding simulator, eliminating the need for pretraining, significantly reducing training time, and improving security performance. The SE process comprises four steps, as illustrated in Figure 4.
Step 1: Generating Adversarial Stego Images with UT-GAN
Inspired by ASDL-GAN, we adopted UT-GAN, an improved version that offers enhanced security performance and shorter training times. We refined its network architecture by adding additional convolutional layers to the existing network and expanding the preprocessing filter bank to improve detection accuracy.
Step 2: Adversarial Gradient Generation
In this step, the adversarial gradient matrix G is generated. G is computed with reference to C and the stego image S generated in Step 1, specifically G = C J ( C , c , φ ( c , s ) ) .
Step 3: Cost Adjustment
Cost adjustment is performed using Equation (12), where p i , j + 1 and p i , j 1 represent the embedding costs for +1 and −1 modifications, respectively, and p i , j 0 (cost for no modification) is set to 0:
p i , j + 1 = ln ( 1 p i , j 1 2 ) p i , j 1 = ln 1 p i , j 1 2 p I , j 0 = 0
Step 4: Embedding Simulator
Using the adjusted cost from Step 3, we further generate adversarial stego images. Initially, a random matrix r (with elements r i , j Uniform ( 0 , 1 ) ) is generated, where all r i , j are independent and identically distributed. The modification mapping m i , j (indicating +1, −1, or 0 modifications) is determined by Equation (13):
m i , j = 1 if r i , j < p i , j 1 + 1 if r i , j > 1 p i , j + 1 0 otherwise
Due to the non-differentiability of the step function in Equation (13), we adopt a continuous approximation (proposed in [33]) to replace the discrete step function, enabling backpropagation, as shown in Equation (14):
m i , j = 1 2 tanh λ p i , j ( + 1 ) r i , j 1 2 tanh λ p i , j ( 1 ) ( 1 r i , j )
where m i , j represents the simulated modification, and λ (set to 60) controls the approximation accuracy of the step function. Finally, we use the Double-Tanh function for message embedding. During the above steps, multiple iterated stego images ( Z 1 , Z 2 , , Z K ) are generated. From these candidates, we select the optimal stego image Z that minimizes the total embedding cost, with the cost measured as the feature–space distance between the cover image and the stego image (standardized by Equation (15)):
Z = min Z { Z K } d f NNDCTR ( C ) , f NNDCTR ( Z )

4. Experiments

4.1. Experimental Setup

All experiments were implemented using Python 3.9.13 and TensorFlow 2.10.0 for deep learning-based components (e.g., the CNN steganalyzer, implemented with TensorFlow and Keras), while traditional steganographic algorithms and feature-based steganalyzers were implemented using MATLAB R2021b. We used 10 image datasets (denoted as C) for experiments, each containing 10,000 natural images. Each image was adjusted to a fixed pixel size (512 × 512) using a scaling factor of 1.9, and then split into two subsets: CTRN (training set) and CTRT (test set), each with 5000 images. These subsets were used exclusively for training and testing steganalyzers, respectively. Specifically, adversarial stego images generated by the UT-GAN network are denoted as S 1 , while those optimized through cost adjustment are denoted as Z j . To evaluate the resilience of the SE scheme under different compression conditions, we extended our experiments to multiple JPEG quality factors (QFs: 50, 75, 85, and 95). The security performance was tested at an embedding rate of 0.4 bits per non-zero AC coefficient (bpnzac) against representative steganalyzers. The experiment involved the following key parameters and metrics:
(1) Steganalyzers: Five types of steganalyzers were used, including three traditional feature-based methods (DCTR, SRM, GFR) and three modern CNN-based methods (XuNet, SRNet, CovNet). These were categorized into non-adversarial and adversarial steganalyzers. Non-adversarial steganalyzers: Do not assume prior knowledge of adversarial embedding and may use adversarial stego images for retraining. Adversarial steganalyzers: Assume the analyst knows the samples are adversarial and are trained on adversarial stego images.
(2) Comparison Methods: We benchmarked the SE (Steganographic Embedding) scheme against four representative steganographic methods:
  • J-UNIWARD: A traditional steganographic algorithm that adopts the Syndrome Trellis Code (STC) method for information embedding.
  • UT-GAN: A GAN (Generative Adversarial Network)-based steganographic model, which is mainly used for preprocessing in the SE (Steganography/Steganalytic Evaluation, adjust based on specific context) task.
  • SPAR-RL: An adaptive steganographic method designed based on reinforcement learning, enabling dynamic adjustment of embedding strategies.
  • DBS: A generative steganographic method built on diffusion models, leveraging the iterative denoising process of diffusion for hidden information transmission.
(3) Detection Criteria: A pretrained binary classifier was used as the steganalyzer. Two types of errors were defined. False negatives ( P m d ): Misclassifying stego images as cover images. False positives ( P f a ): Misclassifying cover images as stego images. The overall steganalytic error rate P e was computed as the average of P m d and P f a , as shown in Equation (16):
P e ( C , S ) = P md ( C , S ) + P fa ( C , S ) / 2
where P md ( C , S ) denotes the false negative rate, and P fa ( C , S ) denotes the false positive rate. The design goal of steganographic algorithms is to maximize P e , whereas the design goal of steganalyzers is the opposite.

4.2. Performance Evaluation of NN-DCTR Detection

The NN-DCTR framework builds upon the DCTR foundation, implementing optimizations and upgrades to enhance performance. Experimental tests in this section compare NN-DCTR with DCTR. The results in Table 2 show that when traditional steganography algorithms J-UNIWARD and UERD are used, NN-DCTR exhibits a slightly higher overall error rate compared to DCTR, indicating a slight decrease in detection performance. In DCTR, rounding operations and histogram calculations are non-differentiable, leading to discrepancies in the differentiation of NN-DCTR from its original version and resulting in performance degradation. However, NN-DCTR remains an effective alternative to DCTR.

4.3. Feature–Space Distance Analysis

The feature–space distance between cover and stego images (quantified by d ( f N N D C T R ( C ) , f N N D C T R ( Z ) ) ) is a crucial metric for evaluating the SE scheme. By varying the parameter k (controlling the intensity of adversarial perturbations), we selected the optimal stego image based on minimizing this distance. In this experiment, the embedding rate was fixed at 0.4 bpnzac, and each scheme (SE, J-SE, U-SE) included nine adjustable k values (0.1, 0.2, …, 0.9). A representative image from the CTRT dataset was used as the cover image, with the embedding rate set to 0.4 bpnzac. As shown in Figure 5, SE achieves the minimum feature–space distance—represented by the shortest red bar compared to J-SE and U-SE—demonstrating its superiority in minimizing the distance between cover and stego images. Additionally, SE exhibits lower average steganographic distances across different k values than J-SE and U-SE. A smaller distance indicates more “accurate” generation of adversarial gradients, confirming that the SE scheme generates more effective adversarial gradients for resisting steganalytic detection.

4.4. Performance Under Different Quality Factors

The security performance under different JPEG quality factors is summarized in Table 3. A general observation is that the detection error rate P e for all methods tends to decrease with decreasing QF, implying that stronger compression facilitates detection. This is expected because more aggressive compression introduces more artifacts that can interfere with both the embedding process and the steganalytic features. Nonetheless, the proposed SE method outperforms all comparison methods across all QFs and steganalyzers. This highlights the robustness of our adversarial embedding scheme under a wide range of compression conditions.

4.5. Non-Adversarial Steganalysis Testing

In this study, the steganalyst was aware of the structure of the steganalyzer, while the steganalyzer remained unaware of adversarial embedding operations, thus qualifying it as a non-adversarial steganalyzer. The experiment employed DCTR, SRM, and GFR steganalyzers as target models trained on the image sets { C J , S J } . The SE algorithm utilized an embedding simulator to compute gradient-modified embedding costs, thereby generating an adversarial stego image dataset Z J . To assess the transferability of adversarial stego images, experiments were also conducted under conditions testing non-adversarial steganalyzers, as shown in Table 4 below.

4.6. Adversarial Steganalysis Testing

In adversarial steganalysis, the steganalyst knows both the steganalyzer’s architecture and the existence of adversarial embedding operations. We used XuNet, SRNet, and CovNet steganalyzers as adversarial steganalyzers, trained on the image sets { C J , S J } . The SE algorithm utilizes an embedding simulator to compute gradient-modified embedding costs, thereby generating an adversarial stego image dataset Z J . To assess the transferability of adversarial stego images, experiments were conducted under conditions testing adversarial steganalyzers, as shown in Table 5 below.

4.7. Computational Cost Analysis

To quantitatively evaluate the efficiency and scalability of the SE method, we conducted a comprehensive empirical study following the methodology suggested in the recent literature [40,41]. We compared its computational overhead against several benchmark algorithms, including J-UNIWARD, UT-GAN, SPAR-RL, and the diffusion-based DBS. All experiments were performed on a unified system equipped with an NVIDIA GeForce RTX 3080 GPU and an Intel Core i9-10900K CPU. The evaluation metrics included:
  • Training Time (GPU Hours): The total time required to train the models to convergence.
  • Inference Time (ms per image): The average time required to process a single image ( 256 × 256 ), including both the UT-GAN preprocessing and the gradient calculator optimization.
  • Computational Complexity (GFLOPs): Measured using a deep learning profiler (TensorFlow Profiler).
  • Peak GPU Memory Usage (GB): The maximum GPU memory consumed during the inference phase.
The results are summarized in Table 6.
The results reveal a clear trade-off between security performance and computational cost, which aligns with findings in related studies on efficient steganographic frameworks [40]:
  • J-UNIWARD, as a conventional non-trainable method, exhibits the lowest overhead but also the lowest security performance (as shown in previous sections).
  • DBS, the diffusion-based generative method, achieves high security but at an extremely high computational cost, making it less practical for real-time applications due to its iterative denoising process.
  • Our SE method strikes an effective balance. It inherits the same training cost as UT-GAN since it uses the same pretrained network. More importantly, its inference overhead is only marginally higher than that of UT-GAN and SPAR-RL (22.1 ms vs. 15.3/18.7 ms). This is because the gradient calculator, despite its powerful optimization capability, is a relatively lightweight network. The complexity (GFLOPs) and memory usage of SE are also on par with other GAN-based adaptive methods.
Recent advances in adversarial image ste [41].Our empirical cost analysis demonstrates that the performance gain of the SE algorithm is achieved with a reasonable and manageable increase in computational cost, offering a far more efficient solution compared to other high-security paradigms like DBS. This makes the proposed method suitable for practical applications where both security and efficiency are concerned.

5. Conclusions

This study presents a novel adversarial steganographic embedding scheme (SE) that outperforms traditional feature-based steganalyzers and their deep learning-based counterparts. Since DCTR steganalyzers are heuristically designed, SE remains independent of any specific dataset. Additionally, the criterion for selecting optimal adversarial steganographic images is minimizing their distance to cover images in the steganalysis space, thus avoiding reliance on dataset-specific knowledge. Unlike traditional feature-based adversarial steganography algorithms, this study introduces an adversarial network and develops a novel gradient calculator. This calculator computes gradients directly from the steganographic features of images, significantly decreasing the time required to generate adversarial stego images. Experiments confirm that SE improves the security of traditional adaptive steganography.
As the gradient calculator serves as the core of this approach, future optimizations to its architecture could lead to better performance. In particular, an ablation study comparing the performance of NN-DCTR against conventional DCTR or CNN-based gradient estimators would provide further validation of its effectiveness. Such a study is of significant research value and will be conducted in our subsequent work.
Furthermore, while the current implementation of SE is tailored for JPEG steganography due to its reliance on DCT coefficient manipulation and quantization-aware operations, the hybrid (partially data-driven) approach underlying our scheme holds potential for generalization to other domains, such as spatial steganography or other compression formats. Future work will explore the adaptability of SE beyond JPEG, addressing domain-specific challenges such as spatial redundancy management and format-specific feature extraction.
To adapt SE to handle large-scale images. Moreover, experiments conducted across multiple JPEG quality factors confirm that the superiority of the SE method is consistent and not limited to specific compression settings, highlighting its practical robustness. Future work will explore the effectiveness of segmentation strategies. Additionally, refining the selection of optimal adversarial steganographic samples remains a key focus for future research.

Author Contributions

Investigation, P.C.; Data curation, Y.Z.; Writing—original draft, Y.L.; Writing—review & editing, L.T. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by Hunan Provincial Social Science Foundation (Grant No. 24YBA341).

Data Availability Statement

The original contributions presented in this study are included in the article. Further inquiries can be directed to the corresponding author.

Conflicts of Interest

The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript; or in the decision to publish the results.

References

  1. Fridrich, J.; Filler, T. Practical Methods for Minimizing Embedding Impact in Steganography. In Security, Steganography, and Watermarking of Multimedia Contents IX; SPIE: Bellingham, WA, USA, 2007; Volume 6505, pp. 13–27. [Google Scholar]
  2. Filler, T.; Judas, J.; Fridrich, J. Minimizing Additive Distortion in Steganography Using Syndrome-Trellis Codes. IEEE Trans. Inf. Forensics Secur. 2011, 6, 920–935. [Google Scholar] [CrossRef]
  3. Pevnỳ, T.; Filler, T.; Bas, P. Using High-Dimensional Image Models to Perform Highly Undetectable Steganography. In Information Hiding; Springer: Berlin/Heidelberg, Germany, 2010; pp. 161–177. [Google Scholar]
  4. Holub, V.; Fridrich, J. Designing Steganographic Distortion Using Directional Filters. In Proceedings of the 2012 IEEE International Workshop on Information Forensics and Security (WIFS), Costa Adeje, Spain, 2–5 December 2012; IEEE: Piscataway, NJ, USA, 2012; pp. 234–239. [Google Scholar]
  5. Chen, X.; Xu, G.; Xu, X.; Jiang, H.; Tian, Z.; Ma, T. Multicenter Hierarchical Federated Learning with Fault-Tolerance Mechanisms for Resilient Edge Computing Networks. IEEE Trans. Neural Netw. Learn. Syst. 2024, 36, 47–61. [Google Scholar] [CrossRef]
  6. Holub, V.; Fridrich, J.; Denemark, T. Universal Distortion Function for Steganography in an Arbitrary Domain. EURASIP J. Inf. Secur. 2014, 2014, 1–13. [Google Scholar] [CrossRef]
  7. Sedighi, V.; Cogranne, R.; Fridrich, J. Content-Adaptive Steganography by Minimizing Statistical Detectability. IEEE Trans. Inf. Forensics Secur. 2015, 11, 221–234. [Google Scholar] [CrossRef]
  8. Li, X.; Chen, X. Value Determination Method Based on Multiple Reference Points Under a Trapezoidal Intuitionistic Fuzzy Environment. Appl. Soft Comput. 2018, 63, 39–49. [Google Scholar] [CrossRef]
  9. Tan, P.; Wang, X.; Wang, Y. Dimensionality Reduction in Evolutionary Algorithms-Based Feature Selection for Motor Imagery Brain-Computer Interface. Swarm Evol. Comput. 2020, 52, 100597. [Google Scholar] [CrossRef]
  10. Fridrich, J.; Kodovsky, J. Rich Models for Steganalysis of Digital Images. IEEE Trans. Inf. Forensics Secur. 2012, 7, 868–882. [Google Scholar] [CrossRef]
  11. Kodovsky, J.; Fridrich, J.; Holub, V. Ensemble Classifiers for Steganalysis of Digital Media. IEEE Trans. Inf. Forensics Secur. 2011, 7, 432–444. [Google Scholar] [CrossRef]
  12. Ren, Y.; Liu, A.; Mao, X.; Li, F. An Intelligent Charging Scheme Maximizing the Utility for Rechargeable Network in Smart City. Pervasive Mob. Comput. 2021, 77, 101457. [Google Scholar] [CrossRef]
  13. Xu, G.; Wu, H.-Z.; Shi, Y.-Q. Structural Design of Convolutional Neural Networks for Steganalysis. IEEE Signal Process. Lett. 2016, 23, 708–712. [Google Scholar] [CrossRef]
  14. Liu, J.; Zhang, X. Truthful Resource Trading for Dependent Task Offloading in Heterogeneous Edge Computing. Future Gener. Comput. Syst. 2022, 133, 228–239. [Google Scholar] [CrossRef]
  15. Yedroudj, M.; Comby, F.; Chaumont, M. Yedroudj-Net: An Efficient CNN for Spatial Steganalysis. In Proceedings of the 2018 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Calgary, AB, Canada, 15–20 April 2018; IEEE: Piscataway, NJ, USA, 2018; pp. 2092–2096. [Google Scholar]
  16. Boroumand, M.; Chen, M.; Fridrich, J. Deep Residual Network for Steganalysis of Digital Images. IEEE Trans. Inf. Forensics Secur. 2018, 14, 1181–1193. [Google Scholar] [CrossRef]
  17. Feng, Y.; Qin, Y.M.; Zhao, S. Correlation-Split and Recombination-Sort Interaction Networks for Air Quality Forecasting. Appl. Soft Comput. 2023, 145, 110544. [Google Scholar] [CrossRef]
  18. Mo, H.; Song, T.; Chen, B.; Luo, W.; Huang, J. Enhancing JPEG Steganography Using Iterative Adversarial Examples. In Proceedings of the 2019 IEEE International Workshop on Information Forensics and Security (WIFS), Delft, The Netherlands, 9–12 December 2019; IEEE: Piscataway, NJ, USA, 2019; pp. 1–6. [Google Scholar]
  19. Yang, Y.; Wu, Y.; Hou, M.; Luo, J.; Xie, X. Solving Emden-Fowler Equations Using Improved Extreme Learning Machine Algorithm Based on Block Legendre Basis Neural Network. Neural Process. Lett. 2023, 55, 7135–7154. [Google Scholar] [CrossRef]
  20. Zhan, M.; Kou, G.; Dong, Y.; Chiclana, F.; Herrera-Viedma, E. Bounded Confidence Evolution of Opinions and Actions in Social Networks. IEEE Trans. Cybern. 2021, 52, 7017–7028. [Google Scholar] [CrossRef]
  21. Kim, S.; Park, J.; Lee, H. An Adaptive JPEG Steganographic Method Based on Weight Distribution for Embedding Costs. J. Korea Inst. Inf. Commun. Eng. 2025, 29, 1389–1398. [Google Scholar]
  22. Shi, Q.; Yang, J. Towards Automatic Embedding Cost Learning for JPEG Steganography. ResearchGate Prepr. 2019. [Google Scholar] [CrossRef]
  23. Tang, W.; Tan, S.; Li, B.; Huang, J. Automatic Steganographic Distortion Learning Using a Generative Adversarial Network. IEEE Signal Process. Lett. 2017, 24, 1547–1551. [Google Scholar] [CrossRef]
  24. Chen, Z.S.; Yang, Y.; Wang, X.J.; Chin, K.S.; Tsui, K.L. Fostering Linguistic Decision-Making Under Uncertainty: A Proportional Interval Type-2 Hesitant Fuzzy TOPSIS Approach Based on Hamacher Aggregation Operators and Andness Optimization Models. Inf. Sci. 2019, 500, 229–258. [Google Scholar] [CrossRef]
  25. Li, C.; He, A.; Wen, Y.; Liu, G.; Chronopoulos, A.T. Optimal Trading Mechanism Based on Differential Privacy Protection and Stackelberg Game in Big Data Market. IEEE Trans. Serv. Comput. 2023, 16, 3550–3563. [Google Scholar] [CrossRef]
  26. Goodfellow, I.J.; Shlens, J.; Szegedy, C. Explaining and Harnessing Adversarial Examples. arXiv 2014, arXiv:1412.6572. [Google Scholar]
  27. Bernard, S.; Bas, P.; Klein, J.; Pevny, T. Explicit Optimization of Min Max Steganographic Game. IEEE Trans. Inf. Forensics Secur. 2020, 16, 812–823. [Google Scholar] [CrossRef]
  28. Zhou, X.; Wu, J.; Liang, W.; Wang, K.I.K.; Yan, Z.; Yang, L.T.; Jin, Q. Reconstructed Graph Neural Network with Knowledge Distillation for Lightweight Anomaly Detection. IEEE Trans. Neural Netw. Learn. Syst. 2024, 35, 11817–11828. [Google Scholar] [CrossRef]
  29. Bernard, S.; Bas, P.; Klein, J.; Pevny, T. Backpack: A Backpropagable Adversarial Embedding Scheme. IEEE Trans. Inf. Forensics Secur. 2022, 17, 3539–3554. [Google Scholar] [CrossRef]
  30. Liu, C.; Kou, G.; Zhou, X.; Peng, Y.; Sheng, H.; Alsaadi, F.E. Time Dependent Vehicle Routing Problem with Time Windows of City Logistics with a Congestion Avoidance Approach. Knowl.-Based Syst. 2020, 188, 104813. [Google Scholar] [CrossRef]
  31. Xiong, T.; Feng, S.; Pan, M.; Yu, Y. Smart Contract Generation for Inter-Organizational Process Collaboration. Concurr. Comput. Pract. Exp. 2024, 36, e7961. [Google Scholar] [CrossRef]
  32. Wang, H.; Zhang, L.; Zhao, Y. Adaptive Embedding Cost Optimization for JPEG Steganography Using Multi-Scale Feature Fusion. IEEE Access 2024, 12, 58721–58734. [Google Scholar]
  33. Holub, V.; Fridrich, J. Phase-Aware Projection Model for Steganalysis of JPEG Images. In Media Watermarking, Security, and Forensics 2015; SPIE: Bellingham, WA, USA, 2015; Volume 9409, pp. 259–269. [Google Scholar]
  34. Pan, Z.; Wang, Y.; Cao, Y.; Gui, W. VAE-Based Interpretable Latent Variable Model for Process Monitoring. IEEE Trans. Neural Netw. Learn. Syst. 2023, 35, 6075–6088. [Google Scholar] [CrossRef]
  35. Liu, D.; Liu, Y.; Chen, X. The New Similarity Measure and Distance Measure Between Hesitant Fuzzy Linguistic Term Sets and Their Application in Multi-Criteria Decision Making. J. Intell. Fuzzy Syst. 2019, 37, 995–1006. [Google Scholar] [CrossRef]
  36. Zhang, W.; Li, J.; Wang, H. Adaptive Feature Fusion for CNN-Based Steganalysis via Attention Mechanism. IEEE Trans. Inf. Forensics Secur. 2022, 17, 789–802. [Google Scholar]
  37. Wang, L.; Zhao, Y.; Chen, B. Lightweight GAN Architecture for Low-Complexity Image Steganography. Neurocomputing 2022, 480, 182–195. [Google Scholar]
  38. Chen, J.; Liu, X.; Zhang, H. DCT Domain Adaptive Embedding Strategy Based on Statistical Distortion Minimization. Signal Process. 2022, 198, 108562. [Google Scholar]
  39. Li, M.; Wang, J.; Yang, F. Transferable Adversarial Perturbations for Cross-Steganalyzer Defense. IEEE Trans. Cybern. 2023, 53, 2456–2468. [Google Scholar]
  40. Fu, Y.; Li, Z.; Huang, F.; Ning, W.; Lyu, H. Design of a Fast Image Encryption Algorithm Based on a Novel 2D Chaotic Map and DNA Encoding. Secur. Priv. 2025, 8, e70036. [Google Scholar] [CrossRef]
  41. Rasool, A. RFS-codec: A Novel Encoding Approach to Store Image Data in DNA. J. Artif. Intell. Bioinf. 2025, 1, 41–50. [Google Scholar] [CrossRef]
Figure 1. Comparison of sample selection criteria for (a) ADV-EMB and (b) SE. In the figures, the orange and light blue areas represent the feature spaces of cover and stego images, respectively. Labels C, S, G, Z1, Z2, and Z3 correspond to the cover image, stego image, selection guidance, and candidate stego images. Dashed lines illustrate the distance from the candidates to key elements (e.g., cover image C or the class-separating hyperplane).
Figure 1. Comparison of sample selection criteria for (a) ADV-EMB and (b) SE. In the figures, the orange and light blue areas represent the feature spaces of cover and stego images, respectively. Labels C, S, G, Z1, Z2, and Z3 correspond to the cover image, stego image, selection guidance, and candidate stego images. Dashed lines illustrate the distance from the candidates to key elements (e.g., cover image C or the class-separating hyperplane).
Electronics 14 04046 g001
Figure 2. Architecture Diagram of Gradient Calculator. The blue arrows represent the data flow. The yellow and orange blocks correspond to the input modules (S and C, respectively). The white boxes (e.g., (NNDCTR, gradient calculator, linear mapper) represent operation units. The terms f NNDCTR ( c ) and f NNDCTR ( s ) denote intermediate results.
Figure 2. Architecture Diagram of Gradient Calculator. The blue arrows represent the data flow. The yellow and orange blocks correspond to the input modules (S and C, respectively). The white boxes (e.g., (NNDCTR, gradient calculator, linear mapper) represent operation units. The terms f NNDCTR ( c ) and f NNDCTR ( s ) denote intermediate results.
Electronics 14 04046 g002
Figure 3. Architecture Diagram of NN-DCTR. The diagram employs the following conventions: blue arrows represent the main data flow; green arrows indicate auxiliary data transfer. The key elements are denoted as: X (Input data), IDCT (Inverse Discrete Cosine Transform), W (Intermediate matrix, defined as ( A T XA)), and (x) (convolution/correlation operation). Teal and orange boxes signify operation modules and a highlighted input, respectively. Here, the 8 × 8 dimension is specified for the core data objects involved in IDCT processing.
Figure 3. Architecture Diagram of NN-DCTR. The diagram employs the following conventions: blue arrows represent the main data flow; green arrows indicate auxiliary data transfer. The key elements are denoted as: X (Input data), IDCT (Inverse Discrete Cosine Transform), W (Intermediate matrix, defined as ( A T XA)), and (x) (convolution/correlation operation). Teal and orange boxes signify operation modules and a highlighted input, respectively. Here, the 8 × 8 dimension is specified for the core data objects involved in IDCT processing.
Electronics 14 04046 g003
Figure 4. Flowchart of SE steps. The diagram utilizes the following conventions: red arrows represent the data and gradient flow; green blocks function as step markers; the colored blocks denote different modules, where teal blocks (C) stand for inputs, orange blocks ( P + and P ) represent positive/negative samples, and maroon/red blocks correspond to intermediate or output results; white boxes (e.g., UT-GAN, gradient calculator, adjust, double-tanh) signify operation units; while m indicates a parameter input.
Figure 4. Flowchart of SE steps. The diagram utilizes the following conventions: red arrows represent the data and gradient flow; green blocks function as step markers; the colored blocks denote different modules, where teal blocks (C) stand for inputs, orange blocks ( P + and P ) represent positive/negative samples, and maroon/red blocks correspond to intermediate or output results; white boxes (e.g., UT-GAN, gradient calculator, adjust, double-tanh) signify operation units; while m indicates a parameter input.
Electronics 14 04046 g004
Figure 5. Feature-Space Distance of Cover and Stego Images for Different Methods with Varying k. The key parameters are: d, the feature-space distance between cover and stego images, which evaluates the gradient optimization effect; and k, which controls the intensity of adversarial perturbations and serves as an input for gradient generation.
Figure 5. Feature-Space Distance of Cover and Stego Images for Different Methods with Varying k. The key parameters are: d, the feature-space distance between cover and stego images, which evaluates the gradient optimization effect; and k, which controls the intensity of adversarial perturbations and serves as an input for gradient generation.
Electronics 14 04046 g005
Table 1. Comparison between ADV-EMB and SE.
Table 1. Comparison between ADV-EMB and SE.
AspectADV-EMBSE (Proposed)
Adversarial modelCNN-based steganalyzerGradient calculator
Data dependencyDataset-specificNon-data-driven
Selection criterionMinimize costMinimize cover–stego distance
Gradient sourcePretrained CNNFeature-based (DCTR-inspired)
TransferabilityLimited to known modelsHigh (generalizable)
Training overheadHigh (requires labeled data)Low (no dataset training)
Table 2. Comparison of detection performance between DCTR and NN-DCTR.
Table 2. Comparison of detection performance between DCTR and NN-DCTR.
SteganalyzerSteganography0.10.20.30.4
PePePePe
NN-DCTRJ-UNIWARD48.746.141.638.7
UERWARD45.343.541.137.4
DCTRJ-UNIWARD48.247.143.139.8
UERWARD46.044.242.238.8
Table 3. Security performance ( P e ) under different JPEG quality factors (embedding rate: 0.4 bpnzac).
Table 3. Security performance ( P e ) under different JPEG quality factors (embedding rate: 0.4 bpnzac).
SteganalyzerMethodQuality Factor
50758595
SRNetJ-UNIWARD25.122.120.518.9
UT-GAN35.833.131.729.2
SPAR-RL36.534.132.830.5
DBS37.134.533.231.6
SE38.535.234.030.8
XuNetJ-UNIWARD36.239.040.141.5
UT-GAN40.539.839.238.5
SPAR-RL41.840.838.439.3
DBS42.040.539.839.0
SE41.240.440.339.5
CovNetJ-UNIWARD28.528.729.230.1
UT-GAN31.229.930.531.8
SPAR-RL33.832.531.932.4
DBS34.533.232.833.1
SE35.133.833.533.9
Note: Bold values indicate the maximum and optimal P e values under each quality factor.
Table 4. The security performance of non-adversarial steganalyzers (denoted as P e ).
Table 4. The security performance of non-adversarial steganalyzers (denoted as P e ).
SteganalyzerSteganographyEmbedding Rate (%)
0.10.20.30.4Average
DCTRJ-UNIWARD47.241.735.228.238.1
ASDL-GAN41.438.734.328.435.7
UT-GAN58.253.149.444.651.3
SPAR-RL59.354.550.646.752.8
DBS59.554.050.946.252.7
SE58.757.951.147.353.8
SRMJ-UNIWARD42.333.827.422.131.4
ASDL-GAN37.032.127.022.529.7
UT-GAN46.843.538.333.140.4
SPAR-RL48.144.839.134.141.5
DBS48.145.238.534.941.6
SE47.844.339.735.241.8
GFRJ-UNIWARD45.237.729.321.633.5
ASDL-GAN41.532.128.719.730.5
UT-GAN47.242.235.631.839.2
SPAR-RL49.143.436.133.640.6
DBS48.844.335.433.540.4
SE48.544.137.432.940.7
Note: Bold values are the optimal security performance (highest P e in respective experimental settings.
Table 5. The security performance of adversarial steganalyzers (denoted as P e ).
Table 5. The security performance of adversarial steganalyzers (denoted as P e ).
SteganalyzerSteganographyEmbedding Rate (%)
0.10.20.30.4Average
XuNetJ-UNIWARD47.445.142.839.043.6
ASDL-GAN43.040.237.133.338.4
UT-GAN48.946.343.539.844.6
SPAR-RL48.747.243.740.845.1
DBS49.047.544.340.545.1
SE49.647.044.840.445.5
SRNetJ-UNIWARD48.139.234.732.638.7
ASDL-GAN48.232.329.027.634.3
UT-GAN49.040.137.333.440.0
SPAR-RL49.337.535.633.439.0
DBS49.538.835.233.739.2
SE48.639.938.134.240.2
CovNetJ-UNIWARD36.032.130.128.731.7
ASDL-GAN32.426.923.922.926.5
UT-GAN36.333.132.529.933.0
SPAR-RL38.133.429.928.532.4
DBS37.933.830.330.033.0
SE37.734.131.130.833.4
Note: Bold values are the optimal security performance (highest P e ) in respective experimental settings.
Table 6. Comparative analysis of computational cost and complexity.
Table 6. Comparative analysis of computational cost and complexity.
MethodTraining Time (GPU hrs)Inference Time (ms/img)Complexity (GFLOPs)GPU Memory (GB)
J-UNIWARD-12.50.020.1
UT-GAN48.215.30.851.8
SPAR-RL62.518.71.122.2
DBS105.4125.615.324.5
SE (Ours)48.222.11.052.1
Note: “-” indicates that the method is non-trainable.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Tan, L.; Li, Y.; Zeng, Y.; Chen, P. An Adaptive JPEG Steganography Algorithm Based on the UT-GAN Model. Electronics 2025, 14, 4046. https://doi.org/10.3390/electronics14204046

AMA Style

Tan L, Li Y, Zeng Y, Chen P. An Adaptive JPEG Steganography Algorithm Based on the UT-GAN Model. Electronics. 2025; 14(20):4046. https://doi.org/10.3390/electronics14204046

Chicago/Turabian Style

Tan, Lina, Yi Li, Yan Zeng, and Peng Chen. 2025. "An Adaptive JPEG Steganography Algorithm Based on the UT-GAN Model" Electronics 14, no. 20: 4046. https://doi.org/10.3390/electronics14204046

APA Style

Tan, L., Li, Y., Zeng, Y., & Chen, P. (2025). An Adaptive JPEG Steganography Algorithm Based on the UT-GAN Model. Electronics, 14(20), 4046. https://doi.org/10.3390/electronics14204046

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop