Next Article in Journal
Some Results on the Erdős–Faber–Lovász Conjecture
Previous Article in Journal
Cosmic Neutrinos as a Window to Departures from Special Relativity
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Imperceptible Image Steganography Using Symmetry-Adapted Deep Learning Techniques

Computer Science Department, Purdue University Fort Wayne, Fort Wayne, IN 46805, USA
*
Author to whom correspondence should be addressed.
Symmetry 2022, 14(7), 1325; https://doi.org/10.3390/sym14071325
Submission received: 30 May 2022 / Revised: 10 June 2022 / Accepted: 18 June 2022 / Published: 27 June 2022

Abstract

:
Digital image steganography is the process of embedding information within a cover image in a secure, imperceptible, and recoverable way. This research extends a symmetry-adapted deep-learning approach to identify hidden patterns of images using two-dimensional convolutional neural networks (CNN). The proposed method (SteGuz) uses three CNNs to implement the different phases of the steganography process on digital image covers. SteGuz introduced a gain function, based on several image similarity metrics, to maximize the imperceptibility of the hiding process. Using 10 different pairs of cover-secret images, the performance of the proposed method was measured in terms of standard metrics such as peak signal to noise ratio (PSNR) and structured similarity index measurement (SSIM). The results showed that the proposed methodology outperformed the original method in terms of both imperceptibility and recoverability. In addition, when compared with some existing methods, SteGuz proved the outstanding performance of achieving a very high PSNR value while maintaining high accuracy of the extracted image.

1. Introduction

Steganography is the process of hiding a message by placing it within another, innocuous medium. The literal meaning of the word, derived from the Greek language, means “covered writing” [1]. The creator of the word, Trithemius, used the Greek words “steganos” and “graphia”, which mean “covered” and “writing”, respectively. One notable example of steganography is found in a story by Greek historian Herodotus. In this story, a Milesian tyrant ordered a slave be sent to the city of Miletus with a secret message tattooed onto his scalp. The slave allowed for time to pass such that his hair grew back, fully obscuring the tattoo. Once he arrived at Miletus, his head was shaved to reveal the tattooed message to the city advisor, Aristagoras. Upon discovery of the message, Aristagoras started a revolution against the Persian king [2]. In this scenario, the slave was the carrier of the message, the tattooed scalp was the hidden information, the tyrant was the sender, and Aristagoras was the receiver.
More recently, many methods exist for embedding hidden information inside digital mediums such as images, audio files, hypertext transport protocol (HTTP) headers, and even plain text files [3]. Nearly any type of digital data, like text and audio files, can be embedded into a cover digital medium. This research focuses on the process of embedding a digital image within another. In fact, digital images provide an excellent medium for information hiding due to several reasons. The first reason is their popularity due to the ease of capture, storage, and exchange. Secondly, images contain many different characteristics, such as bit depth, colors, edges, corners, dimensions, and metadata [4] offering a multiplicity of redundant data that can be manipulated to embed information. Furthermore, the visual integrity of an image is resistant to slight changes in pixel values that may be imperceptible to the human eye. This reduces the likelihood that someone would be able to discern the presence of hidden information.
The Least Significant Bit Substitution (LSB) steganographic method is one of the classical approaches for image steganography. It modifies the least significant bits of the cover pixels to embed the bits of a hidden message. In the case of colored images, the three 8-bit wide color channels (Red, Green, and Blue) provide plenty of space for embedding information [5]. The LSB ++ method [6] is a variation of the LSB algorithm that combines encryption with hiding to increase the level of security against steganalysis. The method uses a key to lock the frequency difference of any two adjacent image bins to maintain sensitive pixels in the cover image, hence reducing the distortion introduced by the embedding process. In [7], the authors proposed another LSB substitution method using pixel value differencing (PVD) and exploiting modification directions (EMD).
Another category of image steganographic methods embed the information in the transform domain of an image. A transform is a mathematical function that is used to alter the structure of the data into waves, such as sine and cosine waves. One example uses used hierarchical trees coding and digital Fourier transform [8] to hide data in images. The Fourier transform (FT) is a transform function that is used to decompose a function into waves of different amplitudes and frequencies [9]. The Discrete Cosine transform (DCT) [10] and The Discrete Wavelet Transform (DWT) [11] are other transform functions that are typically utilized for image steganography. In fact, WLT decomposes a function into a series of wavelets that are wave-like oscillations which are localized in time [12]. Some methods may combine more than one transform function to enhance the hiding process. One novel image steganography methodology used the Finite Ridgelet Transform (FRT), Discrete Wavelet Transform (DWT), and Arnold Scrambling to provide high imperceptibility and hiding capacity [13]. Another technique relied on DWT in conjunction with all phase discrete cosine biorthogonal transform (APDCBT) and singular value decomposition (SVD) [14]. APDCBT is a transform that improves upon the discrete cosine transform (DCT) by offering greater performance in high-frequency and low-frequency aggregation.
With the increased interest in deep learning applications, research efforts were directed to find effective ways to train symmetry-adapted neural networks to identify hidden patterns of image data and deal with the dynamic nature of steganography. The authors in [15] used three Convolutional Neural Networks (CNNs) that were trained together to perform the hiding and the extraction process. Using a similar network structure, the technique proposed in [16] combined the Image Elliptic Curve Cryptography (ECC) and the Discrete Cosine Transform (DCT) as part of the pre-processing network to normalize the secret image and extract its important features. Deep convolutional generative adversarial networks (DCGANs) have also been used in the field of steganography, as seen in [17]. The generative model of the GAS is used to create steganographic images that look like samples from the dataset, while the discriminatory model is used to detect if the steganographic image contains a hidden message [18].
In this research, we introduce an extension of the work introduced in [15] in which three CNNs are built to embed a full-size image into a cover image. In the original work the retrieval of the secret image was approximate yet recognizable. However, in this research three modifications are presented: (1) Removing the normalization and denormalization steps to eliminate the reliance of training on the ImageNet dataset. (2) The noise layer was removed allowing for LSB hiding and hence a higher imperceptibility. (3) The loss function used to update the network weights was replaced with a “gain function”.
The rest of the paper is organized as follows: Section 2 provides a quick overview on the principles of CNN. Section 3 describes the details of the steganographic method (called SteGuz) including the network structure, the training process, as well as the model hyperparameters. Experimental results are discussed in Section 4 and the performance of the proposed method is compared with some existing techniques in Section 5. Finally, Section 6 concludes the research.

2. Convolutional Neural Networks

Deep learning is an approach to artificial intelligence that uses neural networks to achieve a task or goal [19]. Neural networks are loosely modeled after the human brain in the sense that a neural network consists of many interconnected nodes (also known as perceptrons) that communicate with each other. To this end, they are aptly named ‘neural networks.’ The implementation and architecture of neural networks can vary greatly, but the foundation is largely the same. Neural networks are comprised of several layers, with the first layer being the input layer and the last layer being the output layer [20]. In between the input layer and output layer are the hidden layers where the number of hidden layers will vary depending on the application and domain. Each perceptron in the network has one or more weighted inputs that are used in conjunction with an activation function to produce an output variable [21].
Convolutional networks are special types of neural networks designed to work well with two-dimensional data, such as images and media [22]. As shown in Figure 1, a CNN is structured into several convolutional layers that process an image or other two-dimensional input to discover important information such as edges and corners [23]. The distinction between important and unimportant information is decided by the activation function. That information is provided to the pooling layer, where the spatial size of the convolution is reduced while summarizing and maintaining the important details about the convolution. This, in turn, reduces the computational power needed to process the data because there is less information to process. This is achieved by one or more of several types of pooling functions, such as average and max pooling.
When the network is initialized and the weights of the perceptrons are randomized, the gradient descent method [24] or other optimization algorithms are used to update the weights and bias of the perceptron after every step is taken in the training process. As the network continues to update its weights and biases and reach the minimum of the loss function, diminishing results will be returned from the training process [25]. One method of accelerating the training process involves feeding the input data of a network in batches, as opposed to individual samples. The batch size of a network refers to the number of samples passed into the neural network at a time. For this research, the batch size refers to the number of pairs of cover and hidden images processed by the network for each step of the training process.

3. The SteGuz Approach

The goal of the proposed steganographic method is to embed a colored image into another one with the same size. The resultant stego-image should be visually indistinguishable from the original cover. To achieve this, a CNN system is built using the same structure proposed in [15]. Three individual networks: preparatory, hiding, and reveal network, are trained together as one cohesive unit.

3.1. The Network Architecture

As shown in Figure 2, the preparatory network preprocesses the secret image before the embedding takes place. The preparatory network consists of five two-dimensional convolution layers, with each convolution layer having fifty filters each of 3 × 3, 4 × 4, and 5 × 5 patches. Each patch has a kernel size of five and has a ReLU activation function. The output of the preparatory network is a secret tensor with the dimensions of 224 × 224 × 150. By convolving the secret image with the operations in the preparatory network, the dimensionality of the secret image is expanded, thereby allowing the hiding network to select the most important features for the hiding process. As illustrated in Figure 3, the secret tensor is fed into the hiding network along with the cover image to produce the stego-image, which is then sent to the reveal network that reproduces the secret image. Both the hiding network and reveal network have the same structure as the preparatory network. Before the images are provided to the network, they are first rescaled to 224 × 224 pixels. Each image is full-color, with each color channel occupying eight bits per pixel.
It is worth noting that in [15] a noise layer was added to avoid embedding the secret data inside the LSB of the cover image’s pixels. Although this step increases the robustness against steganalysis, it also increases the distortion level in the stego-images. Therefore, in this research, we chose to remove this layer to further improve the imperceptibility of the hiding results. Furthermore, in their original implementation [15], the input images are normalized using the mean and standard deviation values that were calculated from the images in ImageNet dataset. Those hard-coded values are also used by the denormalization process that takes place after the stego-image is generated. In this research, however, we decided to eliminate those steps to decouple the network structure from the training dataset.

3.2. The Gain Function

The learning process of the method proposed in [15] uses a loss function that is presented in (1) where c represents the cover image, c represents the stego-image, s represents the secret image, and s represents the recovered image. The c + c part of the loss function represents the error between the cover and steganographic image and is applied to the preparatory and hiding networks. On the other hand, the β s s part of the loss function represents the error between the hidden and extracted image and is applied to all three networks. The parameter β represents the prescriptive weight given to the extraction error which directly affects the accuracy of the retrieved image.
L ( c ,   c ,   s ,   s ) = c c + β s s
The loss function defined in (1) is based on the pixel level error of two images which ignores the strong correlations between the pixels that carry important information about the structure of the images. The Structured Similarity Index Measure (SSIM), on the other hand, has been efficiently used to measure the image quality as perceived by the human visual system [10]. In the pursuit of a better visual imperceptibility, the learning goal of our network was altered to focus on maximizing image quality instead of minimizing losses. Therefore, we decided to define a “gain function” as a combination of two image quality measures to consider both the image structure and the similarity of image content. The gain function is expressed as follows:
G ( c ,   c ,   s ,   s ) = P S N R ( c , c ) + P S N R ( s , s ) + S S I M ( c , c ) + S S I M ( s , s )
The Peak-to-Signal Ratio (PSNR) is computed as the difference between two images and is calculated in (3) which is expressed in terms of the Mean-Square-Error (MSE) as given in Equation (4). Notice that M represents the maximum pixel value, m and n represent the dimensions of the images, I ( i , j ) represents the original image pixel values at ( i , j ) , and I ( i , j ) represents the corresponding pixel in the modified image.
M S E = i = 1 m j = 1 n ( I ( i , j ) I ( i , j ) ) 2 m n
P S N R = 10 ( l o g 10 ( M 2 M S E ) )
On the other hand, the Structured Similarity Index Measure (SSIM) [16] can be measured using Equation (5), where μ I represents the mean of the original image I while μ I represents the mean of modified image (I′), σ I I 2 represents the covariance of I and I′, σ I 2 represents the mean variance of I, σ I 2 represents the mean variance of I′, and c 1 and c 2 represent two variables to stabilize division, in the event that the denominator is weak [26]. The value of SSIM ranges between 0 and 1, where a higher value reflects a greater similarity between the two images I and I′.
S S I M ( I , I ) = ( 2 μ I μ I + c 1 ) ( 2 σ I I + c 2 ) ( μ I 2 + μ I 2 + c 1 ) ( σ I 2 + σ I 2 + c 2 )
Obviously, the higher the PSNR and SSIM coefficients, the better is the image quality. Furthermore, to optimize the gain/loss function, Adam optimizer was used to train the three networks. The Adam optimizer works well with networks that have large input requirements and has been shown to work well with CNNs [27]. Therefore, the Adam optimizer was used to maximize the gain function described in Equation (2).

3.3. The Model Hyperparameters

The network used in this research was built on an unofficial Python implementation [28] of the method in [15]. The code was revised using the first major version of the Python TensorFlow library. Additional libraries were used in this experiment to include scikit-image for gathering the image metrics [29] and Matplotlib for displaying the images inside the Jupyter Notebook [30]. The hyperparameters were adjusted from the original network as well. The batch size was increased from 8 to 16. The motivation for increasing the batch size was to produce more variation in image details during the training process. The three networks were trained together for ten epochs. However, since the steps taken per epoch were directly correlated to the batch sizes, the original network’s number of steps per epoch was 1784. Since the batch size was doubled in the proposed network, the number of steps per epoch was reduced to 891. The learning rate was slightly increased from 0.0001 to 0.0002. The motivation for increasing the learning rate by a small amount was to speed up the training process, but not so much that the network converges on the wrong solution. In addition, the momentum was decreased from 0.75 to 0.25 to reduce the speed of the training process. In fact, we have noticed that a large momentum value contributed to the “ghosting” effect in the resulting images where the hidden image is slightly visible in the resultant stego-image.

3.4. The Training Dataset

To train and test both the original network and the enhanced network, the Linnaeus 5 [31] and Imagenette [32] datasets were used. Linnaeus 5 is a dataset that contains 6000 square images of five different types of objects. The images in the Linnaeus 5 dataset are saved in a JPG format, with each image having a height and width of 128 pixels. The Imagenette, on the other hand, is derived from the original ImageNet dataset and consists of 9468 images of ten different types of objects. The images provided by the Imagenette dataset were also saved in JPG format but varied widely in height and width.
When the images are supplied to the preparatory and hiding networks, they are first resized to 224 pixels square with the goal of hiding one full-size color image inside another color image of equal size. To enhance the diversity of data during the training process, a randomly selected pair of cover and secret images are used to produce a steganographic image. The generated steganographic image is then used by the reveal network to extract the hidden image and update the network weights. It is important to highlight that the hiding neural network efficiently scattered the secret information throughout the bits of the stego-image which has been simultaneously trained with the reveal network to extract the hidden image.

4. Results

4.1. The Experimental Setup

In our experiments, Google Colaboratory, also referred to as Colab, was used as the code editor and execution environment for training and testing the network. Colab works by hosting the interpretation and execution of Jupyter notebooks on their cloud service, where the Jupyter notebooks are Python-based computational documents. To accelerate the training process, the Pro version of Colab was used, which enabled the use of an Nvidia Tesla T4 or Nvidia Tesla P100 GPU within a high-RAM runtime environment. The space occupied by our model parameters is 34.6 MB.
To showcase the performance of the implemented approach, we choose to use a wide variety of images exhibiting different color and structural details. Table 1 shows the 20 images used as covers and secret images throughout the experiments. The images were collected from various sources. Baboon and Lena, for example, are popular choices in the field of image processing. Other images, such as the Karaoke and French Horn, are samples from the datasets used for training the network. There are also a few original images, such as Graffiti and Lotus. For all test cases, the secret image has the same size as the cover image.

4.2. Objective Performance Measures

Although the visual inspection of the resultant images can be used as a basis for evaluating the efficacy of a given steganography technique, it is still a subjective measure that may differ from one person to the other. Hence, a set of standardized metrics are used here to quantitatively measure and compare the performance of the proposed method with several existing methods. The used metrics covers three aspects: The embedding capacity, the hiding effectiveness, and retrieval similarity.
The hiding capacity of an image steganographic technique is usually measured in terms of the bit per pixel (bpp) value. Obviously, a greater bpp means that the technique can hide more information inside the pixels of the cover image. The following equation can be used to compute the bpp where E represents the maximum embedding capacity of the given technique (in bits) and H and W represent the height and width of the cover image (in pixels), respectively [33]:
H i d i n g   c a p a c i t y = E H W
Throughout the experiments, the Peak-to-Signal Ratio (PSNR) is used to measure the distortion introduced to the resultant stego-image in comparison to the original one. PSNR is computed as the difference between the cover and steganographic images (refer to Equations (3) and (4)) and is measured in decibels (dB). Of course, the higher PSNR value the less the distortion between the two input images which reflects the imperceptibility of the steganographic method. In addition, we choose to measure the accuracy of the extraction process using the Structured Similarity Index Measure (SSIM). Using Equation (5), the value of SSIM ranges between 0 and 1, where a higher value reflects a greater similarity between the secret image and the retrieved one and hence a higher extraction accuracy.

4.3. Experiemental Results

The following set of experiments focus on evaluating the performance of the SteGuz approach in terms of both invisibility and accuracy of extraction. PSNR between the cover and the stego-image is used as a measure of the invisibility and SSIM is used to measure the similarity between the secret image and the retrieved one. Table 2 shows the PSNR and SSIM values for the 10 test image pairs. In addition, the table shown the same metrics evaluated on the results of the original method in [15].
The results in Table 2 shows that the changes proposed in the network structure implemented in SteGuz has a strong impact not only on imperceptibility but also on recoverability. When compared side-by-side, the stego-images generated by SteGuz were visually indistinguishable, while the steganographic images generated by the original network clearly exposed details of the hidden image. A close examination reveals that some details of the secret image, such as edges and corners, are visually perceptible in the stego-image produced by the original method. In most of the test cases, the PSNR values were significantly improved in SteGuz when compared to the results from the original method. The average value of PSNR for SteGuz were 43.04027 dB. On the other hand, the experiments on the original method resulted in the average of 28.69137 dB for PSNR, based on an example figure provided by the author of the original implementation. When trained on the Imagenette and Linnaeus 5 datasets, the averages of PSNR computed for the original method slightly dropped to 27.69137 dB. The reason behind this is the normalization step that is hard-coded based on the parameters collected from the ImageNet dataset.
The recoverability of the proposed network was improved as well in SteGuz when compared with the original network. Comparing the extracted images from the original and proposed networks revealed that the original method’s extraction process lost some of the sharpness and minute details of the image. Additionally, extracted images contained a distorted, wave-like pattern. In Table 2, the sample extracted image generated by the original method resulted in an average of 0.84422 for the SSIM values. The proposed StegGuz approach improved slightly on the similarity between the hidden and extracted image with a SSIM value of 0.87402. Visually, the extraction process of the proposed network succeeded in maintaining the sharpness and the fine details of the recovered image. However, the contrast of the extracted images was noticeably reduced, and the extracted images were slightly grainy.
In conclusion, the results of the experiments strongly support that the changes introduced by the proposed methodology greatly improved the quality of the imperceptibility of the hiding process without sacrificing much in terms the quality of the extracted image. In fact, this trade-off reflects the choice we made when setting the hyperparameters of the networks which highly valued the security of the transmission process over the precise extraction of the hidden information.

5. Discussion

To further discuss the performance of the proposed method, we compare it with several existing methods. Comparison of their performance is carried out from two aspects: objective quantitative metrics (PSNR and SSIM) and the hiding capacity. In the case of image covers, the hiding capacity is measured in bit-per-pixel (bpp) which reflects the maximum message size that might be embedded into an image.
Table 3 shows the methods as named by their authors in their respective publications. The table also shows the images used in their experiments and, whenever possible, the dataset where those images were collected. An exception is the LSB method which is a classical image steganographic approach. The results given here are based on the authors’ implementation of the algorithm. Four bits per each pixel color are utilized for hiding a secret image. If the image doesn’t undergo any compression or noise addition, the extracted image should be identical to the originally embedded one. This means 100% similarity or SSIM of 1.
The results show that SteGuz offers the highest imperceptibility among all the listed methods while maintaining a high accuracy in retrieving the hidden secret image. Except for the LSB method, all methods retrieve an approximation of the secret image. The highest SSIM value between the secret and the extracted images was achieved by FCDNet [34] and yet is less than our method. Furthermore, SteGuz outperformed all other methods in terms of the hiding capacity except for SteganoCNN which offers double the capacity achieved by our method. To achieve this, however, SteganoCNN sacrifices the hiding imperceptibility which is obviously the lowest among the listed methods.
There are several ways in which this work can be improved. The Python code for the proposed network was written using the first version of TensorFlow. Created by Google Brain, TensorFlow 1.0 was released by in 2017 and its successive major revision, TensorFlow 2.0, was released in 2019 [35]. Converting the existing code to TensorFlow 2.0 will benefit from the changes made to the library and improve the accuracy of the results. To reduce the risk of overfitting, one can try different activation functions or use a larger dataset with a greater variety of images for training the network. Another avenue for future work would be testing the proposed methodology using standard steganalysis techniques to quantify the possibility of detecting the presence of a hidden message.

6. Conclusions

In this research, the fields of digital image steganography and convolutional neural networks were explored to develop an enhanced version of the method in [15]. The original method is considered the state-of-art deep-learning-based image steganography. The proposed method (named SteGuz) used the same network structure which consists of three networks that are trained as one unit. However, significant changes have been made to improve the imperceptibility of the hiding process. The main contributions are: (1) the removal of the image normalization/denormalization steps, (2) careful tuning of the network hyperparameters, (3) removing the noise layer, and (4) replacing the loss function with a gain function that prioritizes the similarity of the cover and hidden images over the reduction of the error between the input and generated images. Experimental results showed that the proposed method succeeded in providing higher impressibility as well as better retrieval accuracy than the original one. When compared with several existing methods, SteGuz outperformed their imperceptibility results without sacrificing the accuracy of the extracted image.

Author Contributions

Conceptualization, A.K.; methodology, A.G. and A.K.; software programming, A.G.; supervision, A.K.; formal analysis, A.G. and A.K.; writing—original draft preparation, A.K.; writing—review and editing, A.G. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no fund.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Acknowledgments

We would like to thank the department of Computer Science at Purdue University Fort Wayne for the support with the CoLab pro subscription needed for network training and testing.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Jamil, T. Steganography: The art of hiding information in plain sight. IEEE Potentials 1999, 18, 10–12. [Google Scholar] [CrossRef]
  2. Cox, I.; Miller, M.; Bloom, J.; Fridrich, J.; Kalker, T. Digital Watermarking and Steganography; Elsevier Science & Technology: San Francisco, CA, USA, 2007. [Google Scholar]
  3. Artz, D. Digital steganography: Hiding data within data. IEEE Internet Comput. 2001, 5, 75–80. [Google Scholar] [CrossRef]
  4. What Is Bit Depth? Available online: https://etc.usf.edu/techease/win/images/what-is-bit-depth/ (accessed on 3 April 2022).
  5. Kasapbaşi, M.C.; Elmasry, W. New LSB-based colour image steganography method to enhance the efficiency in payload capacity, security and integrity check. Sādhanā 2018, 43, 68. [Google Scholar] [CrossRef] [Green Version]
  6. Ghazanfari, K.; Ghaemmaghami, S.; Khosravi, S.R. LSB++: An Improvement to LSB+ Steganography. In Proceedings of the IEEE Region 10 International Conference, Bali, Indonesia, 21–24 November 2011. [Google Scholar]
  7. Pradhan, A.; Sekhar, K.R.; Swain, G. Digital Image Steganography Using LSB Substitution, PVD, and EMD. Math. Probl. Eng. 2018, 2018, 1804953. [Google Scholar] [CrossRef]
  8. Chen, W.-Y. Color image steganography scheme using set partitioning in hierarchical trees coding, digital Fourier transform and adaptive phase modulation. Appl. Math. Comput. 2007, 185, 432–448. [Google Scholar] [CrossRef]
  9. Time Series, Signals, & the Fourier Transform, towards Data Science. Available online: https://towardsdatascience.com/time-series-signals-the-fourier-transform-f68e8a97c1c2 (accessed on 3 October 2021).
  10. Lin, C.-C.; Shiu, P.-F. High capacity data hiding scheme for DCT-based images. J. Inf. Hiding Multimed. Signal Process. 2010, 1, 220–240. [Google Scholar]
  11. Hamad, S.; Khalifa, A.; Elhadad, A. A Blind High-Capacity Wavelet-Based Steganography Technique for Hiding Images into other Images. Adv. Electr. Comput. Eng. 2014, 14, 35–42. [Google Scholar] [CrossRef]
  12. The Wavelet Transform. An Introduction and Example, Towards Data Science. Available online: https://towardsdatascience.com/the-wavelet-transform-e9cfa85d7b34 (accessed on 3 October 2021).
  13. Thanki, R.; Borra, S. A color image steganography in hybrid FRT–DWT domain. J. Inf. Secur. Appl. 2018, 40, 92–102. [Google Scholar] [CrossRef]
  14. Zhou, X.; Zhang, H.; Wang, C. A Robust Image Watermarking Technique Based on DWT, APDCBT, and SVD. Symmetry 2018, 10, 77. [Google Scholar] [CrossRef] [Green Version]
  15. Baluja, S. Hiding Images in Plain Sight: Deep Steganography. In Proceedings of the 31st Conference on Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017. [Google Scholar]
  16. Duan, X.; Guo, D.; Liu, N.; Li, B.; Gou, M.; Qin, C. A New High Capacity Image Steganography Method Combined With Image Elliptic Curve Cryptography and Deep Neural Network. IEEE Access 2020, 8, 25777–25788. [Google Scholar] [CrossRef]
  17. Volkhonskiy, D.; Nazarov, I.; Evgeny, B. Steganographic Generative Adversarial Networks. Int. Soc. Opt. Photonics 2020, 11433, 114333M. [Google Scholar] [CrossRef]
  18. Liu, M.-Y.; Tuzel, O. Coupled generative adversarial networks. In Proceedings of the 2016 Conference on Neural Information Processing Systems (NIPS 2016), Barcelona, Spain, 5–10 December 2016. [Google Scholar]
  19. Explained: Neural Networks. Available online: https://news.mit.edu/2017/explained-neural-networks-deep-learning-0414 (accessed on 19 September 2021).
  20. Everything You Need to Know about Neural Networks and Backpropagation—Machine Learning Easy and Fun. Available online: https://towardsdatascience.com/everything-you-need-to-know-about-neural-networks-and-backpropagation-machine-learning-made-easy-e5285bc2be3a (accessed on 19 September 2021).
  21. Rosenblatt, F. The perceptron: A probabilistic model for information storage and organization in the brain. Psychol. Rev. 1958, 65, 386–408. [Google Scholar] [CrossRef] [Green Version]
  22. How Do Convolutional Layers Work in Deep Learning Neural Networks. Available online: https://machinelearningmastery.com/convolutional-layers-for-deep-learning-neural-networks/ (accessed on 19 September 2021).
  23. Towards Data Science. Available online: https://towardsdatascience.com/a-comprehensive-guide-to-convolutional-neural-networks-the-eli5-way-3bd2b1164a53 (accessed on 26 February 2022).
  24. Everything You Need to Know about Gradient Descent Applied to Neural Networks. Available online: https://medium.com/yottabytes/everything-you-need-to-know-about-gradient-descent-applied-to-neural-networks-d70f85e0cc14 (accessed on 26 March 2022).
  25. IBM. What Is Gradient Descent? 27 October 2020. Available online: https://www.ibm.com/cloud/learn/gradient-descent (accessed on 27 March 2022).
  26. Wang, Z.; Simoncelli, E.P.; Bovik, A.C. Multiscale structural similarity for image quality assessment. In Proceedings of the Thrity-Seventh Asilomar Conference on Signals, Systems & Computers, Pacific Grove, CA, USA, 9–12 November 2003. [Google Scholar]
  27. Kingma, D.P.; Ba, J. Adam: A Method for Stochastic Optimization. In Proceedings of the International Conference for Learning Representations (ICLR 2015), San Diego, CA, USA, 7–9 May 2015. [Google Scholar]
  28. Hiding Images Using AI—Deep Steganography. Available online: https://medium.com/buzzrobot/hiding-images-using-ai-deep-steganography-b7726bd58b06 (accessed on 10 January 2021).
  29. van der Walt, S.; Schönberger, J.L.; Nunez-Iglesias, J.D.; Boulogne, F.; Warner, J.; Yager, N.; Gouillart, E.; Yu, T. scikit-image: Image processing in Python. PeerJ 2014, 2, e453. [Google Scholar] [CrossRef]
  30. Hunter, J.D. Matplotlib: A 2D graphics environment. Comput. Sci. Eng. 2007, 9, 90–95. [Google Scholar] [CrossRef]
  31. Chaladze, G.; Kalatozishvili, L. Linnaeus 5 Dataset for Machine Learning. 2017. Available online: http://chaladze.com/l5/ (accessed on 12 April 2022).
  32. Howard, J. Imagenette. Available online: https://github.com/fastai/imagenette (accessed on 12 April 2022).
  33. Zakaria, A.A.; Hussain, M.; Wahab, A.W.A.; Idris, M.Y.I.; Abdullah, N.A.; Jung, K.-H. High-Capacity Image Steganography with Minimum Modified Bits Based on Data Mapping and LSB Substitution. Appl. Sci. 2018, 8, 2199. [Google Scholar] [CrossRef] [Green Version]
  34. Duan, X.; Nao, L.; Mengxiao, G.; Yue, D.; Xie, Z.; Ma, Y.; Qin, C. High-Capacity Image Steganography Based on Improved FC-DenseNet. IEEE Access 2020, 8, 170174–170182. [Google Scholar] [CrossRef]
  35. TensorFlow 2.0 Is Now Available! Available online: https://blog.tensorflow.org/2019/09/tensorflow-20-is-now-available.html (accessed on 10 June 2022).
  36. Chen, F.; Xing, Q.; Liu, F. Technology of Hiding and Protecting the Secret Image Based on Two-Channel Deep Hiding Network. IEEE Access 2020, 8, 21966–21979. [Google Scholar] [CrossRef]
  37. Duan, X.; Liu, N.; Gou, M.; Wang, W.; Qin, C. SteganoCNN: Image Steganography with Generalization Ability Based on Convolutional Neural Network. Entropy 2020, 22, 1140. [Google Scholar] [CrossRef] [PubMed]
Figure 1. An illustration of a CNN with several convolutional layers to extract the features of an input image (Image source: https://towardsdatascience.com/a-comprehensive-guide-to-convolutional-neural-networks-the-eli5-way-3bd2b1164a53, last accessed on 12 April 2022).
Figure 1. An illustration of a CNN with several convolutional layers to extract the features of an input image (Image source: https://towardsdatascience.com/a-comprehensive-guide-to-convolutional-neural-networks-the-eli5-way-3bd2b1164a53, last accessed on 12 April 2022).
Symmetry 14 01325 g001
Figure 2. The details of the preparatory network of the SteGuz method.
Figure 2. The details of the preparatory network of the SteGuz method.
Symmetry 14 01325 g002
Figure 3. The hiding and reveal network system of the SteGuz method.
Figure 3. The hiding and reveal network system of the SteGuz method.
Symmetry 14 01325 g003
Table 1. The images used in experiments.
Table 1. The images used in experiments.
Cover ImagesSecret Images
Symmetry 14 01325 i001
Baboon
Symmetry 14 01325 i002
Berries
Symmetry 14 01325 i003
Graffiti
Symmetry 14 01325 i004
Karaoke
Symmetry 14 01325 i005
Chainsaws
Symmetry 14 01325 i006
Church
Symmetry 14 01325 i007
Lena
Symmetry 14 01325 i008
Lotus
Symmetry 14 01325 i009
Dog
Symmetry 14 01325 i010
Fish
Symmetry 14 01325 i011
Parachute
Symmetry 14 01325 i012
Parrot
Symmetry 14 01325 i013
French Horn
Symmetry 14 01325 i014
Garbage Truck
Symmetry 14 01325 i015
Pens
Symmetry 14 01325 i016
Peppers
Symmetry 14 01325 i017
Gas Pump
Symmetry 14 01325 i018
Golf Balls
Symmetry 14 01325 i019
Stained Glass
Symmetry 14 01325 i020
Thistle
Table 2. The hiding/extraction results of StegGuz vs. the deep Steganography method.
Table 2. The hiding/extraction results of StegGuz vs. the deep Steganography method.
ExperimentSteGuzDeep Steganography [15]
Graffiti in Baboon Symmetry 14 01325 i021
PSNR: 42.85748
Symmetry 14 01325 i022
SSIM: 0.88449
Symmetry 14 01325 i023
PSNR: 27.96642
Symmetry 14 01325 i024
SSIM: 0.87001
Karaoke in Berries Symmetry 14 01325 i025
PSNR: 44.33152
Symmetry 14 01325 i026
SSIM: 0.93213
Symmetry 14 01325 i027
PSNR: 25.62975
Symmetry 14 01325 i028
SSIM: 0.85424
Lena in Chainsaws Symmetry 14 01325 i029
PSNR: 42.4834
Symmetry 14 01325 i030
SSIM: 0.91373
Symmetry 14 01325 i031
PSNR: 29.35
Symmetry 14 01325 i032
SSIM: 0.91303
Lotus in Church Symmetry 14 01325 i033
PSNR: 44.17786
Symmetry 14 01325 i034
SSIM: 0.90252
Symmetry 14 01325 i035
PSNR: 29.59204
Symmetry 14 01325 i036
SSIM: 0.86413
Parachute in Dog Symmetry 14 01325 i037
PSNR: 44.72544
Symmetry 14 01325 i038
SSIM: 0.91306
Symmetry 14 01325 i039
PSNR: 25.7312
Symmetry 14 01325 i040
SSIM: 0.93006
Parrot in Fish Symmetry 14 01325 i041
PSNR: 44.487
Symmetry 14 01325 i042
SSIM: 0.92739
Symmetry 14 01325 i043
PSNR: 25.77007
Symmetry 14 01325 i044
SSIM: 0.92166
Pens in French Horn Symmetry 14 01325 i045
PSNR: 42.45224
Symmetry 14 01325 i046
SSIM: 0.83423
Symmetry 14 01325 i047
PSNR: 27.37058
Symmetry 14 01325 i048
SSIM: 0.77593
Peppers in Garbage Truck Symmetry 14 01325 i049
PSNR: 43.06884
Symmetry 14 01325 i050
SSIM: 0.89151
Symmetry 14 01325 i051
PSNR: 27.89983
Symmetry 14 01325 i052
SSIM: 0.89549
Stained Glass in Gas Pump Symmetry 14 01325 i053
PSNR: 41.65722
Symmetry 14 01325 i054
SSIM: 0.81930
Symmetry 14 01325 i055
PSNR: 25.3442
Symmetry 14 01325 i056
SSIM: 0.83418
Thistle in Golf Balls Symmetry 14 01325 i057
PSNR: 40.16172
Symmetry 14 01325 i058
SSIM: 0.92120
Symmetry 14 01325 i059
PSNR: 25.99714
Symmetry 14 01325 i060
SSIM: 0.87267
Table 3. A performance comparison with some existing techniques.
Table 3. A performance comparison with some existing techniques.
MethodApproachCover/Secret ImagesImage DatasetCover vs. Stego-Image PSNR (db)Secret vs. Retrieved Image SimilarityHiding Capacity (bpp)
LSBSpatial Symmetry 14 01325 i061N/A35.28112
TDHN [36]CNN Symmetry 14 01325 i062NWPU-RESISC4540.30.9824
FCDNet [34]FC-DenseNet Symmetry 14 01325 i063ImageNet40.1960.98423.96
SteganoCNN [37]CNN Symmetry 14 01325 i064ImageNet21.6140.85548
Deep Stego [15]CNN Symmetry 14 01325 i065ImageNet28.690.7324
SteGuzCNN Symmetry 14 01325 i066Linnaeus 5/Imagenette44.330.9324
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Khalifa, A.; Guzman, A. Imperceptible Image Steganography Using Symmetry-Adapted Deep Learning Techniques. Symmetry 2022, 14, 1325. https://doi.org/10.3390/sym14071325

AMA Style

Khalifa A, Guzman A. Imperceptible Image Steganography Using Symmetry-Adapted Deep Learning Techniques. Symmetry. 2022; 14(7):1325. https://doi.org/10.3390/sym14071325

Chicago/Turabian Style

Khalifa, Amal, and Anthony Guzman. 2022. "Imperceptible Image Steganography Using Symmetry-Adapted Deep Learning Techniques" Symmetry 14, no. 7: 1325. https://doi.org/10.3390/sym14071325

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop