Figure 1.
Framework of AGASI: The encoder embeds the secret grayscale image () into a color cover image (), generating a stego-image (). Next, the stego-image () undergoes adversarial perturbation attacks, and through adversarial training, an adversarial stego-image () is generated. Subsequently, the encoder and discriminator form a generative adversarial network (GAN), where the discriminator’s task is to distinguish between the cover image () and the adversarial stego-image (), ultimately generating the adversarial stego-image (). Finally, after processing through the decoder, the secret grayscale image () is successfully extracted from ().
Figure 1.
Framework of AGASI: The encoder embeds the secret grayscale image () into a color cover image (), generating a stego-image (). Next, the stego-image () undergoes adversarial perturbation attacks, and through adversarial training, an adversarial stego-image () is generated. Subsequently, the encoder and discriminator form a generative adversarial network (GAN), where the discriminator’s task is to distinguish between the cover image () and the adversarial stego-image (), ultimately generating the adversarial stego-image (). Finally, after processing through the decoder, the secret grayscale image () is successfully extracted from ().
Figure 2.
The Inception module processes the input image in parallel using convolutional kernels and pooling operations of various sizes, enabling it to capture features at different scales, thereby enhancing the network’s expressiveness and efficiency.
Figure 2.
The Inception module processes the input image in parallel using convolutional kernels and pooling operations of various sizes, enabling it to capture features at different scales, thereby enhancing the network’s expressiveness and efficiency.
Figure 3.
The embedding and extraction of the secret grayscale image , where represents the secret information extracted by the decoder without any added adversarial perturbations.
Figure 3.
The embedding and extraction of the secret grayscale image , where represents the secret information extracted by the decoder without any added adversarial perturbations.
Figure 4.
In this process, the stego-image undergoes adversarial attacks to generate an adversarial stego-image (), which deceives the neural network analyzer, causing misclassification and enhancing the security of the stego-image.
Figure 4.
In this process, the stego-image undergoes adversarial attacks to generate an adversarial stego-image (), which deceives the neural network analyzer, causing misclassification and enhancing the security of the stego-image.
Figure 5.
The encoder replaces the generator. Together with the discriminator, they form a GAN. The generated makes it difficult for third parties to accurately identify the secret information in the image through deep learning models, while the secret information () extracted by the decoder maintains high quality.
Figure 5.
The encoder replaces the generator. Together with the discriminator, they form a GAN. The generated makes it difficult for third parties to accurately identify the secret information in the image through deep learning models, while the secret information () extracted by the decoder maintains high quality.
Figure 6.
Success rate curves of different neural network models under FGSM and PGD attacks. The horizontal axis represents the perturbation strength, while the vertical axis shows the attack success rate. The blue line corresponds to the ResNet50 model, the green line to the VGG16 model, and the red line to the DenseNet121 model. Solid lines indicate the success rate under the FGSM attack, while dashed lines represent the success rate under the PGD attack. These curves offer a clear comparison of the degrees of robustness of different models across varying attack strengths.
Figure 6.
Success rate curves of different neural network models under FGSM and PGD attacks. The horizontal axis represents the perturbation strength, while the vertical axis shows the attack success rate. The blue line corresponds to the ResNet50 model, the green line to the VGG16 model, and the red line to the DenseNet121 model. Solid lines indicate the success rate under the FGSM attack, while dashed lines represent the success rate under the PGD attack. These curves offer a clear comparison of the degrees of robustness of different models across varying attack strengths.
Figure 7.
As the perturbation strength increases, the quality of the steganographic image () deteriorates.
Figure 7.
As the perturbation strength increases, the quality of the steganographic image () deteriorates.
Figure 8.
The first column represents the secret image to be hidden (); the second column is the color cover image (); the third column is the steganographic image generated by the encoder (); the fourth column shows the adversarial steganographic images generated at PGD attack perturbation strengths of 0.01, 0.03, and 0.05, from top to bottom (); the last column shows the final adversarial steganographic image generated after processing through the GAN with a PGD perturbation strength of 0.10 (). It can be observed that the GAN effectively improves the quality of the images formed after adversarial attacks.
Figure 8.
The first column represents the secret image to be hidden (); the second column is the color cover image (); the third column is the steganographic image generated by the encoder (); the fourth column shows the adversarial steganographic images generated at PGD attack perturbation strengths of 0.01, 0.03, and 0.05, from top to bottom (); the last column shows the final adversarial steganographic image generated after processing through the GAN with a PGD perturbation strength of 0.10 (). It can be observed that the GAN effectively improves the quality of the images formed after adversarial attacks.
Figure 9.
The confusion matrix heatmap is used to evaluate the classification performance of the ResNet50 model on . The confusion matrix is a tool for assessing the performance of a classification model, showing the relationship between the predicted and true labels of the model for each category. The X-axis represents the predicted classes (Predicted Class), which are the output predictions of the ResNet50 model for the images, while the Y-axis represents the true classes (True Class), which are the actual labels of the images. The value in each cell indicates the number of images that belong to the true class but are predicted as a specific class. Darker colors indicate a higher number of predictions for that category, while lighter colors suggest fewer predictions.
Figure 9.
The confusion matrix heatmap is used to evaluate the classification performance of the ResNet50 model on . The confusion matrix is a tool for assessing the performance of a classification model, showing the relationship between the predicted and true labels of the model for each category. The X-axis represents the predicted classes (Predicted Class), which are the output predictions of the ResNet50 model for the images, while the Y-axis represents the true classes (True Class), which are the actual labels of the images. The value in each cell indicates the number of images that belong to the true class but are predicted as a specific class. Darker colors indicate a higher number of predictions for that category, while lighter colors suggest fewer predictions.
Figure 10.
The results under PGD attack are shown. The first column displays the secret information (). The second column shows the extracted secret information () from the decoder without any perturbation. The third column shows the extracted secret image () from the adversarial stego-image with a perturbation strength of 0.04. The fourth column shows the extracted secret image () from the adversarial stego-image with a perturbation strength of 0.04. The last column displays the extracted secret image () from the adversarial stego-image with a perturbation strength of 0.10.
Figure 10.
The results under PGD attack are shown. The first column displays the secret information (). The second column shows the extracted secret information () from the decoder without any perturbation. The third column shows the extracted secret image () from the adversarial stego-image with a perturbation strength of 0.04. The fourth column shows the extracted secret image () from the adversarial stego-image with a perturbation strength of 0.04. The last column displays the extracted secret image () from the adversarial stego-image with a perturbation strength of 0.10.
Figure 11.
(a,b) shows Variations in steganalysis performance of the DenseNet-40 and Wide ResNet models under different perturbation strengths. The x-axis represents the perturbation strength of FGSM and PGD attacks, ranging from 0.002 to 0.012. The blue line represents the PGD attack, while the red line represents the FGSM attack. The y-axis represents the steganalysis performance of the neural network, which indicates the probability that it correctly classifies the images.
Figure 11.
(a,b) shows Variations in steganalysis performance of the DenseNet-40 and Wide ResNet models under different perturbation strengths. The x-axis represents the perturbation strength of FGSM and PGD attacks, ranging from 0.002 to 0.012. The blue line represents the PGD attack, while the red line represents the FGSM attack. The y-axis represents the steganalysis performance of the neural network, which indicates the probability that it correctly classifies the images.
Figure 12.
In figures (a,b), the horizontal axis represents the False Positive Rate (FPR), while the vertical axis represents the True Positive Rate (TPR). The area in the bottom-right corner of the curve in the graph corresponds to the AUC value, which indicates the detection accuracy of the Wide ResNet classifier.
Figure 12.
In figures (a,b), the horizontal axis represents the False Positive Rate (FPR), while the vertical axis represents the True Positive Rate (TPR). The area in the bottom-right corner of the curve in the graph corresponds to the AUC value, which indicates the detection accuracy of the Wide ResNet classifier.
Figure 13.
The first column displays the original images, the second column shows the images and their extractions generated in the first stage, the third column presents the images and their extractions generated in the second stage, and the fourth, fifth, and sixth columns illustrate the images and their extractions generated in the third stage.
Figure 13.
The first column displays the original images, the second column shows the images and their extractions generated in the first stage, the third column presents the images and their extractions generated in the second stage, and the fourth, fifth, and sixth columns illustrate the images and their extractions generated in the third stage.
Table 1.
The qualities of C, C′, and C′adv under different perturbation intensities.
Table 1.
The qualities of C, C′, and C′adv under different perturbation intensities.
Image | Attack | Perturbation
| PSNR | SSIM | SME |
---|
| None | None | 32.924 | 0.944 | 0.047 |
| FGSM | 0.01 | 31.548 | 0.922 | 0.056 |
| FGSM | 0.02 | 28.984 | 0.876 | 0.063 |
| FGSM | 0.03 | 26.749 | 0.767 | 0.067 |
| FGSM | 0.04 | 24.451 | 0.731 | 0.071 |
| FGSM | 0.05 | 22.124 | 0.656 | 0.074 |
| PGD | 0.01 | 30.942 | 0.918 | 0.058 |
| PGD | 0.02 | 28.413 | 0.862 | 0.061 |
| PGD | 0.03 | 25.984 | 0.774 | 0.067 |
| PGD | 0.04 | 24.947 | 0.747 | 0.073 |
| PGD | 0.05 | 22.721 | 0.683 | 0.076 |
Table 2.
The qualities of C, C′, and C′adv’ under different disturbance intensities.
Table 2.
The qualities of C, C′, and C′adv’ under different disturbance intensities.
Image | Attack | Perturbation | PSNR | SSIM | SME |
---|
| None | None | 32.924 | 0.944 | 0.047 |
| FGSM | 0.04 | 29.863 | 0.881 | 0.057 |
| FGSM | 0.08 | 28.718 | 0.854 | 0.063 |
| FGSM | 0.10 | 27.515 | 0.796 | 0.069 |
| PGD | 0.04 | 29.762 | 0.874 | 0.058 |
| PGD | 0.08 | 28.732 | 0.862 | 0.061 |
| PGD | 0.10 | 27.523 | 0.783 | 0.069 |
Table 3.
The qualities of , and under different disturbance intensities.
Table 3.
The qualities of , and under different disturbance intensities.
Image | Attack | Perturbation
| PSNR | SSIM |
---|
| None | None | 30.932 | 0.911 |
| PGD | 0.04 | 26.443 | 0.764 |
| PGD | 0.04 | 29.146 | 0.887 |
| PGD | 0.08 | 27.813 | 0.831 |
| PGD | 0.10 | 27.237 | 0.804 |
Table 4.
The effectiveness of AGASI in reducing the accuracy of Wide ResNet detection under different adversarial environments.
Table 4.
The effectiveness of AGASI in reducing the accuracy of Wide ResNet detection under different adversarial environments.
Perturbation | Wide ResNet (FGM) | Wide ResNet (PGD) |
---|
0.000 | 85.347 | 85.347 |
0.001 | 83.142 | 81.151 |
0.003 | 77.154 | 75.612 |
0.005 | 75.689 | 73.154 |
0.007 | 67.154 | 63.157 |
0.009 | 64.841 | 56.745 |
0.011 | 62.164 | 50.168 |
0.013 | 63.145 | 40.681 |
0.015 | 65.154 | 39.641 |
0.017 | 62.653 | 40.952 |
0.020 | 61.315 | 41.214 |
0.024 | 58.154 | 34.154 |
0.026 | 56.214 | 31.541 |
0.030 | 53.754 | 28.021 |
Table 5.
Performance of AGASI secret information extraction with PGD attack.
Table 5.
Performance of AGASI secret information extraction with PGD attack.
Image | Attack | Perturbation
| PSNR | SSIM |
---|
| None | None | 31.022 | 0.951 |
| PGD | 0.004 | 26.043 | 0.764 |
| PGD | 0.004 | 28.746 | 0.887 |
| PGD | 0.006 | 27.813 | 0.843 |
| PGD | 0.010 | 27.137 | 0.834 |
Table 6.
Embedding capability of AGASI.
Table 6.
Embedding capability of AGASI.
Method | Embedding Capacity (bpp) |
---|
LSB Matching Revisited [36] | 1–2 |
Adaptive Steganography [37] | 1–3 |
QIM [38] | 2–4 |
AGASI | 8 |