Next Article in Journal
Performance Improvement for Discretely Modulated Continuous-Variable Measurement-Device-Independent Quantum Key Distribution with Imbalanced Modulation
Previous Article in Journal
Solutions of a Class of Switch Dynamical Systems
Previous Article in Special Issue
Image Forensics in the Encrypted Domain
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Improved CNN Prediction Based Reversible Data Hiding for Images

by
Yingqiang Qiu
1,*,
Wanli Peng
2 and
Xiaodan Lin
1
1
College of Information Science & Engineering, Huaqiao University, Xiamen 361021, China
2
School of Computer Science, Fudan University, Shanghai 200433, China
*
Author to whom correspondence should be addressed.
Entropy 2025, 27(2), 159; https://doi.org/10.3390/e27020159
Submission received: 7 January 2025 / Revised: 28 January 2025 / Accepted: 31 January 2025 / Published: 3 February 2025

Abstract

:
This paper proposes a reversible data hiding (RDH) scheme for images with an improved convolutional neural network (CNN) predictor (ICNNP) that consists of three modules for feature extraction, pixel prediction, and complexity prediction, respectively. Due to predicting the complexity of each pixel with the ICNNP during the embedding process, the proposed scheme can achieve superior performance compared to a CNNP-based scheme. Specifically, an input image is first split into two sub-images, i.e., a “Circle” sub-image and a “Square” sub-image. Meanwhile, each sub-image is applied to predict another one with the ICNNP. Then, the prediction errors of pixels are sorted based on the predicted pixel complexities. In light of this, some sorted prediction errors with less complexity are selected to be efficiently applied for low-distortion data embedding with a traditional histogram-shifting technique. Experimental results show that the proposed ICNNP can achieve better rate-distortion performance than the CNNP, demonstrating its effectiveness.

1. Introduction

Reversible data hiding (RDH) can extract embedded data correctly and recover cover media without any loss [1,2]. Due to these traits, RDH has become a research focus in the community of information hiding and has been widely applied in several realistic scenarios, including medical, military, and law forensics. Based on the domain hiding additional data, RDH can be categorized into two main categories: spatial domain-based RDH [3,4,5,6,7,8,9,10,11,12,13,14,15,16,17,18,19,20,21,22,23,24,25,26] and JPEG domain-based RDH [27,28,29,30,31,32]. Spatial domain-based RDH generally utilizes three classic techniques and their extensions, i.e., lossless compression (LC) [3,4,5], difference expansion (DE) [6,7,8,9,10,11,12,13,14,15,16,17,18], and histogram shifting (HS) [19,20,21,22,23,24,25,26]. In contrast, JPEG domain-based RDH is mainly based on DCT coefficient modification [27,28,29,30] or Huffman table modification [31,32].
Currently, in the RDH field, pixel prediction has become a key point, which dramatically affects the performance of RDH algorithms [15]. Traditional predictors include the median edge direction predictor (MEDP) [7], interpolation predictor [8], gradient-adjusted predictor (GAP) [9], pixel value ordering (PVO) predictor [10,13,25], linear predictor [11], rhombus predictor [20,21,22,23], ridge regression predictor [26], etc. Although these predictors have achieved visual improvement, there is still a notable weakness, which is that few adjacent pixels are applied for pixel prediction [15]. If more adjacent pixels serve as reference pixels, a higher prediction performance can be achieved. Due to its strong capabilities of fusing different receptive fields and whole optimization, a convolutional neural network (CNN) can be constructed and trained to precisely predict pixels by building a nonlinear mapping for pixel prediction. In light of this, Luo et al. [14] presented a CNN-based stereo image RDH scheme leveraging the correlations between right and left views. Hu et al. [15] proposed a CNN predictor (CNNP)-based RDH scheme, where a grayscale image is split into two sub-images, and each one is predicted with another one alternatively using the CNNP. After that, Hu et al. [16] divided a cover image into four parts, and each part was predicted with the other three parts in turn using a CNNP for a better prediction performance. Furthermore, superior performance is attained through prediction-error-ordering (PEO)-based adaptive embedding. Overall, the prediction performance of CNN predictors can be better than that of traditional predictors. In [17], Yang et al. introduced a novel RDH approach that strategically segments a cover image into four distinct regions. Within each region, each pixel is then predicted using the surrounding eight neighbor pixels through a custom CNN predictor. This CNN predictor was designed to enhance the precision of the prediction errors, thereby facilitating a more efficient data embedding process with the classical prediction error expansion (PEE) embedding strategy. Zhou et al. [18] presented a new RDH method that seamlessly integrates a transformer predictor with an improved PEO-based adaptive embedding strategy. This method, characterized by its multiple embedding rules, is adept at significantly diminishing embedding distortions and elevating the visual quality of the resultant marked images.
From the above discussion, in order to improve performance, existing methods conduct pixel prediction by leveraging adjacent pixels, but these methods [14,15,16,17] ignore the complexity of each pixel with deep learning, which limits the performance of RDH. To tackle the above limitation, in this letter, we improve the CNNP presented in [15] by adding a complexity prediction part to precisely predict pixels’ complexities, which is called the improved CNNP (ICNNP) in the rest of this letter. Specifically, during data embedding, we first split a grayscale image into two sub-images, where one sub-image is predicted by another one. Then, we sort the prediction errors of the predicted pixels based on their complexities, and the prediction errors with less complexity are used for data embedding with a classical HS strategy. Finally, experimental results show that the achieved performance of the proposed ICNNP-based scheme is better than that of the CNNP-based scheme presented in [15].

2. Proposed Improved Scheme

2.1. Network Architecture

As shown in Figure 1, according to the checkerboard context model [20], we split the cover image into two sub-images, which consist of “Circle” and “Square” pixels, respectively. For the “Circle” sub-image, the values of the “Circle” pixels are maintained, while those of the “Square” pixels are set to 0. Meanwhile, just the values of the “Circle” pixels are set to 0 for the “Square” sub-image. Due to the pixel correlation of the two sub-images, each sub-image is used to predict the pixel values and complexities of the other sub-image.
Figure 2 depicts the architecture of the proposed ICNNP. The architecture of the ICNNP consists of three parts, i.e., feature extraction, pixel prediction, and complexity prediction. “Square” sub-image I2 is fed to the network to predict the values and complexities of the “Circle” pixels, where the values of complexity are adjusted to [0, 255] for better visualization. The lower the value, the lower the complexity. The feature extraction consists of some convolution layers with different filter sizes (3 × 3, 5 × 5, 7 × 7, ), which are parallelized and appended with a 3 × 3 convolution layer, respectively, to extract features from different receptive fields. A residual block is then applied to further aggregate and refine the learnt features from different branches. With the extracted feature, the pixel prediction yields the predicted “Circle” sub-image I ~ 1 , and the complexity prediction yields the predicted complexity C ~ 1 of the “Circle” sub-image I1. “Conv” stands for the convolution unit with kernel size S × S, and the number of channels is output × input. A LeakyReLU activation function [33] is located between each two convolution layers.
It is worthy to note that the complexity prediction is similar to the pixel prediction, i.e., instead of orthogonal adjacent pixels [15,20], more adjacent pixels are used to nonlinearly predict the complexity of the pixel area, improving the performance of RDH.

2.2. Training

In the ICNNP, the well-trained parameters of the CNNP [15] are loaded into the feature extraction and pixel prediction modules. Note that these parameters are fixed, and the complexity prediction parameters are updated during the training of the ICNNP. During the training, the input of the ICNNP is the “Square” sub-image I2, while the outputs are the predicted “Circle” sub-image I ~ 1 and the predicted complexity C ~ 1 of the “Circle” sub-image I1. Since the filter parameters of the feature extraction and the pixel prediction are fixed, the training objective is no longer the “Circle” sub-image I1 but the referenced complexity C 1 of I1. The definition of C 1 is described as follows:
(1)
For “Square” pixels, C 1 i , j is set to 0.
(2)
For “Circle” pixels, if i = 1 , i = M , j = 1 , or j = N , C 1 i , j is set to 0; otherwise, C 1 i , j ( 2 i M 1 , 2 j N 1 ) is calculated as
C 1 i , j = 1 R · k = k 1 k 2 l = l 1 l 2 ( I i + k , j + l I i , j ) 2 ,
where
k 1 = 1 , k 2 = 2      ,   i = 2                   k 1 = 2 , k 2 = 1      ,   i = M 1          k 1 = 2 , k 2 = 2   ,   2 < i < M 1 ,
l 1 = 1 , l 2 = 2      ,   j = 2                   l 1 = 2 , l 2 = 1      ,   j = N 1          l 1 = 2 , l 2 = 2   ,   2 < j < N 1 ,
R = k 2 k 1 + 1 × l 2 l 1 + 1 1 .
In the proposed scheme, the max predicted pixel area is 5 × 5, which can accurately calculate the pixel complexity.
As with the CNNP in [15], we leverage back-propagation [34] and the Adam algorithm [35] to optimize the objective function defined below:
L o s s = 1 N i = 1 N ( C ~ 1 , i C 1 , i ) 2 + λ ω 2 2 ,
where N is the number of training images, and C ~ 1 , i and C 1 , i represent the predicted and referenced complexities of the “Circle” sub-image in the i-th training image, respectively. ω stands for all weights of the network, and λ denotes the weight decay.

2.3. Data Embedding of ICNNP-Based RDH

Figure 3 depicts the data embedding architecture of the ICNNP-based RDH scheme. The adopted double embedding strategy [20] with the HS technique [7] involves the successive usage of the “Circle” sub-image embedding and the “Square” sub-image embedding, and the “Square” sub-image embedding is performed after the “Circle” sub-image embedding.
Cover image I is firstly separated into two sub-images, i.e., a “Circle” sub-image I1 and a “Square” sub-image I2. Next, the predicted “Circle” sub-image I ~ 1 and the predicted complexity C ~ 1 of I1 are predicted with I2 as follows:
I ~ 1 , C ~ 1 = I C N N P I 2 .
Then, the prediction errors of I1 are calculated as
e 1 i , j = I 1 i , j I ~ 1 i , j ,   i + j m o d 2 0 ,
where “ ” represents modular congruence. According to the magnitude of predicted complexities and the size of the additional data S1, we select the predicted errors with less complexity and determine two thresholds Tn1 (Tn1 < 0) and Tp1 (Tp1 ≥ 0) for HS-based data embedding, which is achieved as
E 1 i , j = 2 e 1 i , j + b             ,   i f   e 1 i , j [ T n 1 , T p 1 ]   e 1 i , j + T p 1 + 1   ,   i f   e 1 i , j > T p 1             e 1 i , j + T n 1          ,   i f   e 1 i , j < T n 1            ,
where b 0 ,   1 is the data to be embedded, including the encrypted additional data and some auxiliary data [20]. Therefore, the marked “Circle” sub-image MI1 is generated as
M I 1 i , j = I ~ 1 i , j + E 1 i , j .
During the “Square” embedding process, due to the requirement that the pattern of the input data must match the “Square” sub-image I2, the marked “Circle” sub-image MI1 cannot be directly fed into the network to predict the “Square” sub-image I2. As illustrated in Figure 4, if the image’s height is even, rotating it clockwise by 90 degrees results in the image M I 1 matching the pattern of the ’Square’ sub-image I2. With the same ICNNP, as shown in Equation (10), we feed the rotated marked “Circle” sub-image M I 1 into the network, and then we obtain the predicted rotated “Square” sub-image I ~ 2 and its complexity C ~ 2 .
I ~ 2 , C ~ 2 = I C N N P M I 1 .
Then, I ~ 2 and C ~ 2 are rotated counterclockwise by 90 degrees to obtain the predicted “Square” sub-image I ~ 2 and the predicted complexity C ~ 2 of I2. Similarly to the “Circle” embedding, another part of additional data S2 is encrypted with K and then embedded into I2 to obtain the marked “Square” sub-image MI2.
Finally, we combine the marked “Circle” sub-image MI1 and the marked “Square” sub-image MI2 to obtain the marked image MI.

2.4. Extraction and Image Recovery of ICNNP-Based RDH

Figure 5 describes the architecture of data extraction and image recovery, which are the reverse procedures of data embedding, so we perform the “Circle” extraction/recovery ahead of the “Square” extraction/recovery. The marked image MI is firstly divided into two sub-images, i.e., the marked “Circle” sub-image MI1 and the marked “Square” sub-image MI2. With the rotated marked “Circle” sub-image M I 1 , I ~ 2 and C ~ 2 are predicted using the ICNNP as in Equation (10). I ~ 2 and C ~ 2 are then obtained by rotating I ~ 2 and C ~ 2 respectively. Next, the marked prediction errors of I2 are calculated as
E 2 i , j = M I 2 i , j I ~ 2 i , j ,   i + j m o d 2 1 .
According to the sorted magnitude of C ~ 2 and the extracted auxiliary data Tn2 (Tn2 < 0) and Tp2 (Tp2 ≥ 0) as thresholds, the data extraction is operated as
b = E 2 i , j   m o d   2   ,   E 2 i , j [ 2 T n 2 , 2 T p 2 + 1 ] ,
and the original prediction errors of I2 are recovered as
e 2 i , j = E 2 i , j / 2              , i f   E 2 i , j [ 2 T n 2 , 2 T p 2 + 1 ]   E 2 i , j T p 2 1 , i f   E 2 i , j > 2 T p 2 + 1               E 2 i , j T n 2          , i f   E 2 i , j < 2 T n 2                       ,
where · is the floor function. We decrypt the extracted bits to obtain S1 with K, and recover the cover “Square” sub-image I2 as
I 2 i , j = e 2 i , j + I ~ 2 i , j ,   i + j m o d 2 1
Similarly, S1 is extracted correctly, and the cover “Circle” sub-image I1 is recovered losslessly. Finally, we combine the recovered “Square” sub-image I2 and “Circle” sub-image I1 to obtain I.

3. Experimental Results

To assess the efficiency of the proposed ICNNP-based scheme, the parameters of its complexity prediction model were trained with 1000 grayscale images of size 512 × 512, which were selected from BOWS-2 [36] randomly. The ICNNP was trained on an Intel Core i10 CPU (3.6 GHz) with 16 GB of RAM and an NVIDIA GeForce RTX 2060. The equipment was sourced from Lenovo, located in Beijing, China. The weight decay λ was 1 × 10−3, the batch size was set to 4, and the initial learning rate was set to 1 × 10−3. The number of training epochs was 20, and the optimizer used for training the ICNNP was Adam. In [15], the prediction performance of the CNNP is proved to be better than some traditional linear predictors, such as MEDP, GAP, etc., and the achieved rate-distortion performance is better than that of the above predictor with the expansion embedding scheme. It is also better than that of BIP with the HS scheme, and the performance of HS is far better than that of expansion embedding with the CNNP. Therefore, we just evaluate the performance with the ICNNP by comparing it with that of the CNNP [15] with the same HS technique.
With the four 512 × 512 grayscale images shown in Figure 6 as cover images with the same embedding capacity (EC), we employed a PSNR (peak signal-to-noise ratio) between the marked image and the original image as the metric for objective image quality evaluation. In addition, we chose 100 grayscale images of size 512 × 512 randomly that were different from the training images from BOWS-2 [36], and we tested them with different embedding capacities (ECs) to evaluate the universality of the ICNNP.
Figure 7 shows the PSNR values of four test images with different Ecs, which indicates that the PSNRs of the ICNNP-based RDH scheme are larger than those of the CNNP-based RDH scheme. Figure 8 shows the experimental results when embedding 10,000 bits of data, corresponding to an embedding rate of approximately 0.038 bpp. The PSNRs between the stego images and the original images are 57.02 dB, 60.73 dB, 51.81 dB, and 56.60 dB, respectively. The stego images are visually indistinguishable from the original images shown in Figure 6. From Table 1, we can see the average PSNRs of 100 test images for different ECs. When the EC is 10,000 bits, the mean PSNR achieved by the ICNNP-based RDH scheme is 62.37 dB, while that of the CNNP-based RDH scheme is 61.31 dB, which is 1.06 dB lower than that of the proposed scheme. Along with the EC increases from 20,000 bits to 150,000 bits, the mean PSNRs of the ICNNP-based RDH scheme are 0.81 dB, 0.60 dB, 0.55 dB, 0.56 dB, 0.55 dB, 0.54 dB, 0.53 dB, 0.51 dB, 0.47 dB, 0.41 dB, 0.37 dB, 0.33 dB, 0.26 dB, and 0.20 dB higher than those of the CNNP-based RDH scheme with the same EC, respectively.

4. Conclusions

The improved CNN predictor for RDH presented in this paper extracts features from different receptive fields with whole optimization and uses more neighboring pixels to precisely predict the pixel value and its complexity. During data embedding, a grayscale image is split into two sub-images, one sub-image is applied to predict the other sub-image alternately using the ICNNP. Then, the pixels’ prediction errors are sorted according to the pixels’ prediction complexities, and the prediction errors with less complexity are chosen for data embedding with the classical HS strategy. The original image is recovered losslessly, and the embedded data are extracted correctly. The data extraction and image recovery are separable. Experimental results demonstrate that the proposed ICNNP, when combined with the classical histogram shifting (HS) strategy, achieves superior performance compared to the CNNP presented in [15] using the same HS strategy, thereby proving its effectiveness.
The proposed ICNNP is the first to predict pixel errors and their complexity for RDH. Employing joint training for pixel prediction and complexity prediction holds the promise of achieving superior RDH performance. In future work, we plan to contemplate the joint optimization of pixel error prediction and pixel complexity prediction using advanced deep learning methods, as well as the integration of various embedding strategies, such as PEE, PEO, and multi-histogram shifting (MHS).

Author Contributions

Conceptualization, Y.Q.; methodology, Y.Q. and W.P.; software, Y.Q. and X.L.; validation, W.P. and X.L.; formal analysis, W.P.; investigation, W.P.; resources, Y.Q. and X.L.; data curation, X.L.; writing—original draft preparation, Y.Q.; writing—review and editing, W.P.; visualization, W.P.; supervision, Y.Q.; project administration, Y.Q.; funding acquisition, Y.Q. and X.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Natural Science Foundation of Xiamen, China, grant number 3502Z20227192, and the Natural Science Foundation of China, grant number 62002124.

Institutional Review Board Statement

Not applicable.

Data Availability Statement

The data presented in this study are available on request from the corresponding author.

Acknowledgments

The authors would like to thank all anonymous reviewers and editors for their helpful suggestions for the improvement in this paper.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Shi, Y.-Q.; Li, X.; Zhang, X.; Wu, H.; Ma, B. Reversible data hiding: Advances in the past two decades. IEEE Access 2016, 4, 3210–3237. [Google Scholar] [CrossRef]
  2. Zhang, C.; Ou, B.; Peng, F.; Zhao, Y.; Li, K. A survey on reversible data hiding for uncompressed images. ACM Comput. Surv. 2024, 56, 1–33. [Google Scholar] [CrossRef]
  3. Celik, M.U.; Sharma, G.; Tekalp, A.M.; Saber, E. Lossless generalized-LSB data embedding. IEEE Trans. Image Process. 2005, 14, 253–266. [Google Scholar] [CrossRef] [PubMed]
  4. Zhang, W.; Hu, X.; Li, X.; Yu, N. Optimal transition probability of reversible data hiding for general distortion metrics and its applications. IEEE Trans. Image Process. 2015, 24, 294–304. [Google Scholar] [CrossRef]
  5. Hou, D.; Zhang, W.; Yang, Y.; Yu, N. Reversible data hiding under inconsistent distortion metrics. IEEE Trans. Image Process. 2018, 27, 5087–5099. [Google Scholar] [CrossRef]
  6. Tian, J. Reversible data embedding using a difference expansion. IEEE Trans. Circuits Syst. Video Technol. 2003, 13, 890–896. [Google Scholar] [CrossRef]
  7. Thodi, D.M.; Rodriguez, J.J. Expansion embedding techniques for reversible watermarking. IEEE Trans. Image Process. 2007, 16, 721–730. [Google Scholar] [CrossRef]
  8. Luo, L.; Chen, Z.; Chen, M.; Zeng, X.; Xiong, Z. Reversible image watermarking using interpolation technique. IEEE Trans. Inf. Forensics Secur. 2010, 5, 187–193. [Google Scholar]
  9. Coltuc, D. Low distortion transform for reversible watermarking. IEEE Trans. Image Process. 2012, 21, 412–417. [Google Scholar] [CrossRef] [PubMed]
  10. Li, X.; Li, J.; Li, B.; Yang, B. High-fidelity reversible data hiding scheme based on pixel-value-ordering and prediction-error expansion. Signal Process. 2013, 93, 198–205. [Google Scholar] [CrossRef]
  11. Dragoi, I.-C.; Coltuc, D. Local-prediction-based difference expansion reversible watermarking. IEEE Trans. Image Process. 2014, 23, 1779–1790. [Google Scholar] [CrossRef] [PubMed]
  12. Qiu, Y.; Qian, Z.; Yu, L. Adaptive reversible data hiding by extending the generalized integer transformation. IEEE Signal Process. Lett. 2016, 23, 130–134. [Google Scholar]
  13. He, W.; Cai, Z. An insight into pixel value ordering prediction-based prediction-error expansion. IEEE Trans. Circuits Syst. Video Technol. 2020, 15, 3859–3871. [Google Scholar] [CrossRef]
  14. Luo, T.; Jiang, G.; Yu, M.; Zhong, C.; Xu, H.; Pan, Z. Convolutional neural networks-based stereo image reversible data hiding method. J. Vis. Commun. Image Represent. 2019, 61, 61–73. [Google Scholar] [CrossRef]
  15. Hu, R.; Xiang, S. CNN Prediction Based Reversible Data Hiding. IEEE Signal Process. Lett. 2021, 28, 464–468. [Google Scholar] [CrossRef]
  16. Hu, R.; Xiang, S. Reversible Data Hiding by Using CNN Prediction and Adaptive Embedding. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 44, 464–468. [Google Scholar] [CrossRef] [PubMed]
  17. Yang, X.; Huang, F. New CNN-Based Predictor for Reversible Data Hiding. IEEE Signal Process. Lett. 2022, 29, 2627–2631. [Google Scholar] [CrossRef]
  18. Zhou, L.; Lu, Z.; You, W. Reversible data hiding using a transformer predictor and an adaptive embedding strategy. Front. Inf. Technol. Electron. Eng. 2023, 24, 1143. [Google Scholar] [CrossRef]
  19. Ni, Z.; Shi, Y.; Ansari, N.; Su, W. Reversible data hiding. IEEE Trans. Circuits Syst. Video Technol. 2006, 16, 354–362. [Google Scholar]
  20. Sachnev, V.; Kim, H.J.; Nam, J.; Suresh, S.; Shi, Y.-Q. Reversible watermarking algorithm using sorting and prediction. IEEE Trans. Circuits Syst. Video Technol. 2009, 19, 989–999. [Google Scholar] [CrossRef]
  21. Li, X.; Zhang, W.; Gui, X.; Yang, B. Efficient reversible data hiding based on multiple histograms modification. IEEE Trans. Inf. Forensics Secur. 2015, 10, 2016–2027. [Google Scholar]
  22. Qi, W.; Li, X.; Zhang, T.; Guo, Z. Optimal Reversible Data Hiding Scheme Based on Multiple Histograms Modification. IEEE Trans. Circuits Syst. Video Technol. 2020, 30, 2300–2312. [Google Scholar] [CrossRef]
  23. Wang, J.; Chen, X.; Ni, J.; Mao, N.; Shi, Y. Multiple Histograms-Based Reversible Data Hiding: Framework and Realization. IEEE Trans. Circuits Syst. Video Technol. 2020, 30, 2313–2328. [Google Scholar] [CrossRef]
  24. Ou, B.; Zhao, Y. High capacity reversible data hiding based on multiple histograms modification. IEEE Trans. Circuits Syst. Video Technol. 2020, 30, 2329–2342. [Google Scholar] [CrossRef]
  25. Zhang, T.; Li, X.; Qi, W.; Guo, Z. Location-based pvo and adaptive pairwise modification for efficient reversible data hiding. IEEE Trans. Inf. Forensics Secur. 2020, 15, 2306–2319. [Google Scholar] [CrossRef]
  26. Wang, X.; Wang, X.; Ma, B.; Li, Q.; Shi, Y.-Q. High Precision Error Prediction Algorithm Based on Ridge Regression Predictor for Reversible Data Hiding. IEEE Signal Process. Lett. 2021, 28, 1125–1129. [Google Scholar] [CrossRef]
  27. Weng, S.; Zhou, Y.; Zhang, T.; Xiao, M.; Zhao, Y. General Framework to Reversible Data Hiding for JPEG Images With Multiple Two-Dimensional Histograms. IEEE Trans. Multimed. 2023, 25, 5747–5762. [Google Scholar] [CrossRef]
  28. Mao, N.; He, H.; Chen, F.; Yuan, Y.; Qu, L. Reversible Data Hiding of JPEG Image Based on Adaptive Frequency Band Length. IEEE Trans. Circuits Syst. Video Technol. 2023, 33, 7212–7223. [Google Scholar] [CrossRef]
  29. Zhou, X.; Hou, K.; Zhuang, Y.; Yin, Z.; Han, W. General Pairwise Modification Framework for Reversible Data Hiding in JPEG Images. IEEE Trans. Circuits Syst. Video Technol. 2024, 34, 153–167. [Google Scholar] [CrossRef]
  30. Li, F.; Qi, Z.; Zhang, X.; Qin, C. Progressive Histogram Modification for JPEG Reversible Data Hiding. IEEE Trans. Circuits Syst. Video Technol. 2024, 34, 1241–1254. [Google Scholar] [CrossRef]
  31. Qiu, Y.; Qian, Z.; He, H.; Tian, H.; Zhang, X. Optimized lossless data hiding in JPEG bitstream and relay transfer based extension. IEEE Trans. Circuits Syst. Video Technol. 2021, 31, 1380–1394. [Google Scholar] [CrossRef]
  32. Du, Y.; Yin, Z.; Zhang, X. High capacity lossless data hiding in JPEG bitstream based on general VLC mapping. IEEE Trans. Depend. Secure Comput. 2022, 19, 1420–1433. [Google Scholar] [CrossRef]
  33. Maas, A.L.; Awni, Y.H.; Andrew, Y.N. Rectifier nonlinearities improve neural network acoustic models. In Proceedings of the 30th International Conference on Machine Learning (ICML-13), Atlanta, GA, USA, 16–21 June 2013. [Google Scholar]
  34. Lecun, Y.; Bottou, L.; Bengio, Y.; Haffner, P. Gradient-based learning applied to document recognition. Proc. IEEE 1998, 86, 2278–2324. [Google Scholar] [CrossRef]
  35. Kingma, D.; Ba, J. Adam: A method for stochastic optimization. In Proceedings of the 3rd International Conference on Learning Representations, San Diego, CA, USA, 7–9 May 2015. [Google Scholar]
  36. Bas, P.; Furon, T. Image Database of Bows-2. 2017. Available online: http://bows2.ec-lille.fr/ (accessed on 21 June 2019).
Figure 1. Illustration of splitting an original image into two sub-images. (a) Original image I. (b) “Circle” sub-image I1. (c) “Square” sub-image I2.
Figure 1. Illustration of splitting an original image into two sub-images. (a) Original image I. (b) “Circle” sub-image I1. (c) “Square” sub-image I2.
Entropy 27 00159 g001
Figure 2. The overall architecture of the proposed ICNNP.
Figure 2. The overall architecture of the proposed ICNNP.
Entropy 27 00159 g002
Figure 3. The flowchart of data embedding using the ICNNP-based RDH scheme.
Figure 3. The flowchart of data embedding using the ICNNP-based RDH scheme.
Entropy 27 00159 g003
Figure 4. Illustration of image rotation of marked “Circle” image. (a) M I 1 . (b) M I 1 .
Figure 4. Illustration of image rotation of marked “Circle” image. (a) M I 1 . (b) M I 1 .
Entropy 27 00159 g004
Figure 5. The flowchart of data extraction and image recovery using the ICNNP-based RDH scheme.
Figure 5. The flowchart of data extraction and image recovery using the ICNNP-based RDH scheme.
Entropy 27 00159 g005
Figure 6. Four cover images.
Figure 6. Four cover images.
Entropy 27 00159 g006
Figure 7. Performance comparison of CNNP in [15] and ICNNP for RDH on four test images.
Figure 7. Performance comparison of CNNP in [15] and ICNNP for RDH on four test images.
Entropy 27 00159 g007
Figure 8. Some experimental results.
Figure 8. Some experimental results.
Entropy 27 00159 g008
Table 1. Average PSNR (dB) values on 100 test images of ICNNP-based RDH method and CNNP-based RDH method [15].
Table 1. Average PSNR (dB) values on 100 test images of ICNNP-based RDH method and CNNP-based RDH method [15].
Embedding Capacity (Bits)Embedding Rate (bpp)CNNP [15]ICNNP
10,0000.03861.3162.37
20,0000.07658.0158.82
30,0000.11455.9856.58
40,0000.15354.4354.98
50,0000.19153.1053.66
60,0000.22951.9452.49
70,0000.26750.8651.40
80,0000.30549.8550.38
90,0000.34348.8849.39
100,0000.38147.9148.38
110,0000.42046.9547.36
120,0000.45846.0346.40
130,0000.49645.1345.46
140,0000.53444.2944.55
150,0000.57243.4743.67
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Qiu, Y.; Peng, W.; Lin, X. Improved CNN Prediction Based Reversible Data Hiding for Images. Entropy 2025, 27, 159. https://doi.org/10.3390/e27020159

AMA Style

Qiu Y, Peng W, Lin X. Improved CNN Prediction Based Reversible Data Hiding for Images. Entropy. 2025; 27(2):159. https://doi.org/10.3390/e27020159

Chicago/Turabian Style

Qiu, Yingqiang, Wanli Peng, and Xiaodan Lin. 2025. "Improved CNN Prediction Based Reversible Data Hiding for Images" Entropy 27, no. 2: 159. https://doi.org/10.3390/e27020159

APA Style

Qiu, Y., Peng, W., & Lin, X. (2025). Improved CNN Prediction Based Reversible Data Hiding for Images. Entropy, 27(2), 159. https://doi.org/10.3390/e27020159

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop