Deep Learning for Image Watermarking: A Comprehensive Review and Analysis of Techniques, Challenges, and Applications
Highlights
- Deep learning-based watermarking methods (CNN, GAN, Transformers, and diffusion models) significantly outperform traditional spatial- and frequency-domain techniques in terms of robustness, transparency, and adaptability to modern attack types.
- Emerging architectures such as Vision Transformers, Swin Transformers, and diffusion models introduce new capabilities, notably higher resistance to generative and latent-space attacks, as well as increased watermark capacity.
- The rapid evolution of neural network architectures accelerates the development of watermarking systems capable of protecting digital content against increasingly sophisticated threats, including AI-generated manipulations.
- Future watermarking deployments will require optimized, scalable, and computationally efficient deep learning architectures to support real-time applications in cybersecurity, multimedia distribution, IoT systems, and content authenticity verification.
Abstract
1. Introduction
- A review and taxonomy of classical watermarking methods for images and video frames.
- A comprehensive overview and taxonomy of deep learning-based image watermarking techniques, highlighting their advantages, limitations, and potential application areas.
- Detailed comparisons of various deep learning architectures (CNNs, GANs, Transformers, and diffusion models) used in watermarking, with particular emphasis on their performance, robustness, and computational complexity.
- A review and comparison of key datasets used for training watermarking algorithms.
- An analysis of future research directions and practical challenges in areas such as deepfake detection, cybersecurity, and applications in IoT systems, with a special focus on the integration of deep learning methods into watermarking solutions.
- A discussion of dataset availability, training strategies, and the role of transparency metrics based on neural networks, as well as specialized robustness metrics tailored to assess the impact of generative and adversarial attacks, providing practical guidelines useful for real-world implementations.
2. Fundamentals of Digital Watermarking
2.1. Watermarking Workflow
2.2. Watermarking Taxonomy
2.3. Watermarking Paradigms and Metrics
2.3.1. Transparency
- MSE (Mean Squared Error)—the average squared error between the pixel values of the original image and the watermarked image.
- f—the matrix data of the original image,
- g—the matrix data of the watermarked image,
- m—the number of pixel rows in the images and i represents the index of that row,
- n—the number of pixel columns in the image and j represents the index of that column.
- PSNR (Peak Signal-to-Noise Ratio)—measures the ratio of the original image signal to the noise introduced by the watermarking process.
- SSIM (Structural Similarity Index) [41]—measures the structural similarity between two images by analyzing contrast, brightness, and texture. The values range from 0 to 1, with values closer to 1 indicating higher similarity.
- µx, µy–mean luminance values for X and Y images,
- σx σy—standard deviation values for X and Y images,
- σxy—covariance value between X and Y images,
- C1 C2—stabilizing constants to prevent division by zero, where
- L—dynamic range of the pixel values (255 for 8-bit grayscale images),
- K1, K2—small constant, K1 ≪ 1 and K2 ≪ 1.
- MS-SSIM (Multiscale Structural Similarity Index) [42]—an extension of SSIM that considers multiple spatial scales, calculated through a multi-stage downsampling process.
- M—number of scale levels,
- αj—weights (usually αj = 1/M),
- —SSIM value at the j level.
- VIF (Visual Information Fidelity)—measures the amount of visual information transferred from the original image to the watermarked image using the HVS (Human Visual System) model.
- I0—the amount of information in the original channel,
- —the variance of signal at the k level,
- —the variance of noise at the k level.
- Iw—the amount of information in the modified channel,
- —the variance of error resulting from the difference between the original and the modified image.
- FSIM (Feature Similarity Index)—compares key visual features from the perspective of human perception.
- —phase congruence at a point i,
- —luminance similarity function,
- —phase congruence similarity function.
- DSIS (Double Stimulus Impairment Scale)—the method involves comparing two versions of the same material: the original reference version and the processed version. Users view both versions sequentially and then assess the degree of quality degradation in the processed version relative to the original. Ratings are collected on a quality scale from 1 to 5, where 5 indicates that the differences between the original and the modified content are imperceptible, and 1 signifies that the content quality has significantly deteriorated and is unacceptable.
- DSCQS (Double Stimulus Continuous Quality Scale)—users are presented with both the original and the modified versions of the content, but they are not explicitly informed which is the original or the modified version. Similar to DSIS, users evaluate the content quality; however, the lack of clarity regarding which material has been altered provides a more objective assessment from the perspective of human perception.
- Paired Comparison Test—users are shown two versions of the content: the original and the modified, without indicating which one has been altered. Participants evaluate which version they believe has higher quality or whether they can notice any differences.
- LPIPS (Learned Perceptual Image Patch Similarity) [46]—this method employs convolutional networks trained on large datasets to measure perceptual differences between images. Metric values close to 0 indicate smaller differences and higher transparency of the method.
- DISTS (Deep Image Structure and Texture Similarity) [47]—a metric that combines texture and structure analysis using deep features from neural networks, providing improved assessment of images with complex details.
- PieAPP (Perceptual Image-Error Assessment through Pairwise Preference) [48]—a trained model that predicts image quality based on user preferences by evaluating pairs of images.
2.3.2. Robustness
- BER (Bit Error Rate)—measures the percentage of bits that have been incorrectly extracted from the embedded watermark compared to the original watermark.
- wi—value of the i-th bit in the original watermark,
- —value of the i-th bit in the extracted watermark,
- N—total number of bits.
- NC (Normalized Correlation)—a metric that measures the similarity between the original and the extracted watermark. A value close to 1 indicates high resistance to attacks.
- ASR (Attack Success Rate)—it represents the percentage of successful attempts to weaken or remove the watermark as a result of adversarial attacks [49].
- RGA (Robustness Against Generative Attacks)—a metric evaluating the system’s resistance to attacks utilizing generative models. It analyzes the extent to which the watermark remains intact after being processed by these models [50].
- APT (Adversarial Perturbation Tolerance)—defines the minimum level of perturbations introduced by adversarial attacks necessary to successfully remove or distort the watermark. A higher APT value indicates greater system robustness [51].
2.3.3. Capacity
- Payload Capacity—measures the number of watermark bits relative to the given carrier.
- Nbit—number of bits embedded as watermark,
- Nhost—number of host units (pixels for images, seconds for video or audio).
- Embedding Capacity Efficiency—measures the efficiency with which the system utilizes the carrier’s space for embedding the watermark, taking into account the impact on quality and robustness.
2.4. Attacks on Watermarking Systems
3. Traditional Image Watermarking Methods
3.1. Spatial Domain
3.2. Frequency Domain
3.3. Hybrid Domains
3.4. Summary
4. Deep Learning-Based Watermarking
4.1. Deep Learning Architectures Used in Image Watermarking
- Encoder, which transforms the input data into a lower-dimensional representation, with the goal of reducing the data size while capturing the most important features of the input,
- Decoder, which reconstructs the data based on the representation by applying reverse transformations to those used in the encoder.
- Generator, based on provided features or random noise, learns to generate new data,
- Discriminator attempts to distinguish between real data and data generated by the generator.
4.2. Overview of Deep Learning-Based Image Watermarking Algorithms
4.3. Summary of Deep Learning-Based Image Watermarking Algorithms
5. Datasets for Image Watermarking
- High resolution—Modern watermarking methods should be tested not only on standard images of 128 × 128 or 256 × 256 pixels but also on high-resolution images such as HD (1080p) and 4K, which is essential for practical applications such as multimedia content protection [170];
- Content diversity—The dataset should include both real-world images (landscapes, faces, animals, objects, vehicles) and graphics or textures. This is particularly important for methods utilizing attention mechanisms, which rely on contextual relationships between image elements [129];
- Open access—Free and open access to data facilitates research replication and the comparison of different solutions’ effectiveness, forming the foundation for reliable evaluation of watermarking methods [171];
- High visual quality—Images should be artifact-free, clear, and detailed, allowing for precise evaluation of watermark transparency and its impact on the visual quality of the image after embedding and extraction [172];
- No lossy compression—Lossless formats (e.g., TIFF or PNG) are preferred to avoid artifacts resulting from lossy compression (e.g., JPEG), ensuring a reliable assessment of the watermarking method’s resistance to image degradation [173];
- Synthetic and real images—With the growing popularity of generative models such as DALL-E [174] and Midjourney, there is an increasing need to watermark content generated by artificial intelligence. Therefore, the dataset should include both real and synthetic images to ensure algorithm effectiveness in both contexts [175].
- Benchmark datasets—These are classical image sets widely used in image processing and deep learning research. They are characterized by standard resolutions, usually 256 × 256 pixels, and a high diversity of content, enabling versatile usage;
- High-resolution datasets—These include images with HD and 4K resolutions. Originally intended for training super-resolution algorithms, they are now successfully used to evaluate watermarking effectiveness in real-world applications where high visual quality is essential;
- Synthetic datasets—Comprising images generated by AI models, these datasets feature visual characteristics that differ from those of real-world images. As a result, models trained on such datasets may require adapted watermarking methods to perform effectively;
- Specialized datasets—These include images from specific fields, such as medicine, geoinformatics, security, or digital documents, where watermarking plays a key role in ensuring data integrity and authenticity. Such images are characterized by a high level of detail and specific visual features, often necessitating tailored watermarking approaches.
6. New Research Directions and Challenges in Image Watermarking
6.1. Key Challenges in Implementing Image Watermarking Systems
6.2. Research Directions for Methods and Algorithms
6.3. Application-Oriented Research in Watermarking
6.3.1. Watermarking in Identification of Deepfakes
6.3.2. Watermarking in Cybersecurity
6.3.3. Watermarking in Monitoring and IoT Systems
7. Conclusions
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Conflicts of Interest
Abbreviations
| CNN | Convolutional Neural Network |
| DCT | Discrete Cosine Transform |
| DFT | Discrete Fourier Transform |
| DWT | Discrete Wavelet Transform |
| GAN | Generative Adversarial Network |
| IoT | Internet of Things |
| PSNR | Peak Signal-to-Noise Ratio |
| SSIM | Structural Similarity Index |
| ViT | Vision Transformer |
| VIF | Visual Information Fidelity |
References
- Spaiser, V. Chapter 19: Digital Data and Methods. In Sociology, Social Policy and Education; Elgar Publishing: Cheltenham, UK, 2021; ISBN 978-1-78990-685-1. [Google Scholar]
- Seetharamu, S.; Cn, L.M.; Bhattacharya, A.; Bt, D.C. Digital Data Protection Laws: A Review. Int. J. Sci. Res. Sci. Eng. Technol. 2024, 11, 64–75. [Google Scholar] [CrossRef]
- Czetwertyński, S. Digital Piracy: The Issue of Knowledge of the Institution of Copyright Law. Ekon. Prawo Econ. Law 2023, 22, 69–102. [Google Scholar] [CrossRef]
- Townsened, C. The Consequences of Digital Piracy. Available online: https://www.uscybersecurity.net/digital-piracy/ (accessed on 7 October 2024).
- Sadiku, M.N.O.; Ashaolu, T.J.; Ajayi-Majebi, A.; Musa, S.M. Digital Piracy. IJSCIA 2021, 2, 797–800. [Google Scholar] [CrossRef]
- Javed, A.R.; Jalil, Z.; Zehra, W.; Gadekallu, T.R.; Suh, D.Y.; Piran, M.J. A Comprehensive Survey on Digital Video Forensics: Taxonomy, Challenges, and Future Directions. Eng. Appl. Artif. Intell. 2021, 106, 104456. [Google Scholar] [CrossRef]
- Walczyna, T.; Piotrowski, Z. Quick Overview of Face Swap Deep Fakes. Appl. Sci. 2023, 13, 6711. [Google Scholar] [CrossRef]
- Petitcolas, F.; Anderson, R.; Kuhn, M. Information Hiding—A Survey. Proc. IEEE 1999, 87, 1062–1078. [Google Scholar] [CrossRef]
- Cheddad, A.; Condell, J.; Curran, K.; Mc Kevitt, P. Digital Image Steganography: Survey and Analysis of Current Methods. Signal Process. 2010, 90, 727–752. [Google Scholar] [CrossRef]
- Jekateryńczuk, G.; Jankowski, D.; Veyland, R.; Piotrowski, Z. Detecting Malicious Devices in IPSEC Traffic with IPv4 Steganography. Appl. Sci. 2024, 14, 3934. [Google Scholar] [CrossRef]
- Rocha, A.; Goldenstein, S. Steganography and Steganalysis in Digital Multimedia: Hype or Hallelujah? Rev. Informática Teórica Apl. 2008, 15, 83–110. [Google Scholar] [CrossRef]
- Zhang, Y. Digital Watermarking Technology: A Review. In Proceedings of the 2009 ETP International Conference on Future Computer and Communication, Wuhan, China, 6–7 June 2009; pp. 250–252. [Google Scholar]
- Bailey, J. Watermarking vs. Fingerprinting: A War in Terminology. Plagiarism Today 2007. Available online: https://www.plagiarismtoday.com/2007/10/09/watermarking-vs-fingerprinting-a-war-in-terminology/ (accessed on 6 November 2025).
- Hosny, K.M.; Magdi, A.; ElKomy, O.; Hamza, H.M. Digital Image Watermarking Using Deep Learning: A Survey. Comput. Sci. Rev. 2024, 53, 100662. [Google Scholar] [CrossRef]
- Hu, K.; Wang, M.; Ma, X.; Chen, J.; Wang, X.; Wang, X. Learning-Based Image Steganography and Watermarking: A Survey. Expert Syst. Appl. 2024, 249, 123715. [Google Scholar] [CrossRef]
- Piotrowski, Z. Drift Correction Modulation Scheme for Digital Signal Processing. Math. Comput. Model. 2013, 57, 2660–2670. [Google Scholar] [CrossRef]
- Piotrowski, Z. Angle Phase Drift Correction Method Effectiveness. In Proceedings of the Signal Processing Algorithms, Architectures, Arrangements, and Applications SPA 2009, Poznan, Poland, 24–26 September 2009; pp. 82–86. [Google Scholar]
- Piotrowski, Z.; Lenarczyk, P. Blind Image Counterwatermarking—Hidden Data Filter. Multimed. Tools Appl. 2017, 76, 10119–10131. [Google Scholar] [CrossRef]
- Hua, G.; Huang, J.; Shi, Y.Q.; Goh, J.; Thing, V.L.L. Twenty Years of Digital Audio Watermarking—A Comprehensive Review. Signal Process. 2016, 128, 222–242. [Google Scholar] [CrossRef]
- Uddin, M.S.; Ohidujjaman; Hasan, M.; Shimamura, T. Audio Watermarking: A Comprehensive Review. Int. J. Adv. Comput. Sci. Appl. 2024, 15, 1410–1418. [Google Scholar] [CrossRef]
- Shelke, R.D.; Nemade, M.U. Audio Watermarking Techniques for Copyright Protection: A Review. In Proceedings of the 2016 International Conference on Global Trends in Signal Processing, Information Computing and Communication (ICGTSPICC), Jalgaon, India, 22–24 December 2016; pp. 634–640. [Google Scholar]
- Al-khafaji, A.; Alwan; Nur, N.; Sjarif, N.N.A. Digital Text Watermarking Techniques Classification and Open Research Challenges: A Review. Technol. Rep. Kansai Univ. 2020, 62, 18. [Google Scholar]
- Kaur, M.; Mahajan, K. An Existential Review on Text Watermarking Techniques. Int. J. Comput. Appl. 2015, 120, 29–32. [Google Scholar] [CrossRef]
- Kamaruddin, N.S.; Kamsin, A.; Por, L.Y.; Rahman, H. A Review of Text Watermarking: Theory, Methods, and Applications. IEEE Access 2018, 6, 8011–8028. [Google Scholar] [CrossRef]
- Begum, M.; Uddin, M.S. Digital Image Watermarking Techniques: A Review. Information 2020, 11, 110. [Google Scholar] [CrossRef]
- Wadhera, S.; Kamra, D.; Rajpal, A.; Jain, A.; Jain, V. A Comprehensive Review on Digital Image Watermarking. In Proceedings of the 5th International Conference on Computing Sciences (ICCS 2021), Beijing China, 4–6 December 2021. [Google Scholar]
- Sharma, S.; Zou, J.J.; Fang, G.; Shukla, P.; Cai, W. A Review of Image Watermarking for Identity Protection and Verification. Multimed. Tools Appl. 2024, 83, 31829–31891. [Google Scholar] [CrossRef]
- Joseph, I.; Mandala, J. Comprehensive Review on Video Watermarking Security Threats, Challenges, and Its Applications. ECS Trans. 2022, 107, 13833. Available online: https://iopscience.iop.org/article/10.1149/10701.13833ecst (accessed on 10 October 2024). [CrossRef]
- Paul, R.T. Review of Robust Video Watermarking Techniques. IJCA Spec. Issue Comput. Sci. 2011, 3, 90–95. [Google Scholar]
- Dey, A.; Bhattacharya, S.; Chaki, N. Software Watermarking: Progress and Challenges. INAE Lett. 2019, 4, 65–75. [Google Scholar] [CrossRef]
- Dalla Preda, M.; Pasqua, M. Software Watermarking: A Semantics-Based Approach. Electron. Notes Theor. Comput. Sci. 2017, 331, 71–85. [Google Scholar] [CrossRef]
- Li, Y. Database Watermarking: A Systematic View. In Handbook of Database Security: Applications and Trends; Gertz, M., Jajodia, S., Eds.; Springer: Boston, MA, USA, 2008; pp. 329–355. ISBN 978-0-387-48533-1. [Google Scholar]
- Alqassab, A.; Alanezi, M. Relational Database Watermarking Techniques: A Survey. J. Phys. Conf. Ser. 2021, 1818, 012185. [Google Scholar] [CrossRef]
- Boenisch, F. A Systematic Review on Model Watermarking for Neural Networks. Front. Big Data 2021, 4, 729663. [Google Scholar] [CrossRef] [PubMed]
- Li, Y.; Wang, H.; Barni, M. A Survey of Deep Neural Network Watermarking Techniques. Neurocomputing 2021, 461, 171–193. [Google Scholar] [CrossRef]
- Vasiljević, I.; Obradović, R.; Đurić, I.; Popkonstantinović, B.; Budak, I.; Kulić, L.; Milojević, Z.; Vasiljević, I.; Obradović, R.; Đurić, I.; et al. Copyright Protection of 3D Digitized Artistic Sculptures by Adding Unique Local Inconspicuous Errors by Sculptors. Appl. Sci. 2021, 11, 7481. [Google Scholar] [CrossRef]
- Huang, T.; Huang, J.; Pang, Y.; Yan, H. Smart Contract Watermarking Based on Code Obfuscation. Inf. Sci. 2023, 628, 439–448. [Google Scholar] [CrossRef]
- Li, L.; Jiang, B.; Wang, P.; Ren, K.; Yan, H.; Qiu, X. Watermarking LLMs with Weight Quantization. In Proceedings of the Findings of the Association for Computational Linguistics: EMNLP 2023, Singapore, 6–10 December 2023. [Google Scholar]
- Rathi, P.; Bhadauria, S.; Rathi, S. Watermarking of Deep Recurrent Neural Network Using Adversarial Examples to Protect Intellectual Property. Appl. Artif. Intell. 2022, 36, 2008613. [Google Scholar] [CrossRef]
- Lenarczyk, P.; Piotrowski, Z. Parallel Blind Digital Image Watermarking in Spatial and Frequency Domains. Telecommun. Syst. 2013, 54, 287–303. [Google Scholar] [CrossRef]
- Wang, Z.; Bovik, A.C.; Sheikh, H.R.; Simoncelli, E.P. Image Quality Assessment: From Error Visibility to Structural Similarity. IEEE Trans. Image Process. 2004, 13, 600–612. [Google Scholar] [CrossRef]
- Wang, Z.; Simoncelli, E.P.; Bovik, A.C. Multiscale Structural Similarity for Image Quality Assessment. In Proceedings of the Thrity-Seventh Asilomar Conference on Signals, Systems & Computers, Pacific Grove, CA, USA, 9–12 November 2003; pp. 1398–1402. [Google Scholar]
- Recommendation ITU-R BT.500-13; Methodology for the Subjective Assessment of the Quality of Television Pictures. ITU-R: Geneva, Switzerland, 2023.
- Recommendation ITU-T P.910 (10/2023); Subjective Video Quality Assessment Methods for Multimedia Applications. ITU-R: Geneva, Switzerland, 2023. Available online: https://www.itu.int/epublications/publication/itu-t-p-910-2023-10-subjective-video-quality-assessment-methods-for-multimedia-applications (accessed on 19 November 2025).
- Vallez, C.; Kucharavy, A.; Dolamic, L. Needle In A Haystack, Fast: Benchmarking Image Perceptual Similarity Metrics At Scale. arXiv 2022, arXiv:2206.00282. [Google Scholar] [CrossRef]
- Zhang, R.; Isola, P.; Efros, A.A.; Shechtman, E.; Wang, O. The Unreasonable Effectiveness of Deep Features as a Perceptual Metric. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 586–595. [Google Scholar]
- Ding, K.; Ma, K.; Wang, S.; Simoncelli, E.P. Image Quality Assessment: Unifying Structure and Texture Similarity. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 44, 2567–2581. [Google Scholar] [CrossRef] [PubMed]
- Prashnani, E.; Cai, H.; Mostofi, Y.; Sen, P. PieAPP: Perceptual Image-Error Assessment through Pairwise Preference. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018. [Google Scholar]
- Wu, J.; Zhou, M.; Zhu, C.; Liu, Y.; Harandi, M.; Li, L. Performance Evaluation of Adversarial Attacks: Discrepancies and Solutions. arXiv 2021, arXiv:2104.11103. [Google Scholar] [CrossRef]
- Buzhinsky, I.; Nerinovsky, A.; Tripakis, S. Metrics and Methods for Robustness Evaluation of Neural Networks with Generative Models. arXiv 2020, arXiv:2003.01993. [Google Scholar] [CrossRef]
- Rabhi, M.; Pietro, R.D. Adversarial Attacks Neutralization via Data Set Randomization. arXiv 2023, arXiv:2306.12161. [Google Scholar] [CrossRef]
- Fang, H.; Chen, K.; Qiu, Y.; Ma, Z.; Zhang, W.; Chang, E.-C. DERO: Diffusion-Model-Erasure Robust Watermarking. In Proceedings of the 32nd ACM International Conference on Multimedia, Melbourne, VIC, Australia, 28 October–1 November 2024; Association for Computing Machinery: New York, NY, USA, 2024; pp. 2973–2981. [Google Scholar]
- Guo, Y.; Li, R.; Hui, M.; Guo, H.; Zhang, C.; Cai, C.; Wan, L.; Wang, S. FreqMark: Invisible Image Watermarking via Frequency Based Optimization in Latent Space. arXiv 2024, arXiv:2410.20824. [Google Scholar] [CrossRef]
- Cox, I.J. Digital Watermarking and Steganography, 2nd ed.; The Morgan Kaufmann Series in Multimedia Information and Systems; Morgan Kaufmann Publishers: Amsterdam, The Netherlands; Boston, MA, USA, 2008; ISBN 978-0-12-372585-1. [Google Scholar]
- Tuner, L.F. Digital Data Security System. Patent IPN WO 89/08915, 21 September 1989. [Google Scholar]
- van Schyndel, R.G.; Tirkel, A.Z.; Osborne, C.F. A Digital Watermark. In Proceedings of the 1st International Conference on Image Processing, Austin, TX, USA, 13–16 November 1994; Volume 2, pp. 86–90. [Google Scholar]
- Bamatraf, A.; Ibrahim, R.; Salleh, M. Digital Watermarking Algorithm Using LSB. In Proceedings of the 2010 International Conference on Computer Applications and Industrial Electronics, Kuala Lumpur, Malaysia, 5–8 December 2010; p. 159. [Google Scholar]
- Bamatraf, A.; Ibrahim, R.; Salleh, M.N.M. A New Digital Watermarking Algorithm Using Combination of Least Significant Bit (LSB) and Inverse Bit. arXiv 2011, arXiv:1111.6727. [Google Scholar] [CrossRef]
- Arya, R.K.; Saharan, R. Algorithm to Enhance the Robustness and Imperceptibility of LSB. In Proceedings of the 2015 Second International Conference on Advances in Computing and Communication Engineering, Dehradun, India, 1–2 May 2015; pp. 583–587. [Google Scholar]
- Muyco, S.D.; Hernandez, A.A. Least Significant Bit Hash Algorithm for Digital Image Watermarking Authentication. In Proceedings of the 2019 5th International Conference on Computing and Artificial Intelligence; Association for Computing Machinery: New York, NY, USA, 2019; pp. 150–154. [Google Scholar]
- Faheem, Z.B.; Ishaq, A.; Rustam, F.; de la Torre Díez, I.; Gavilanes, D.; Vergara, M.M.; Ashraf, I. Image Watermarking Using Least Significant Bit and Canny Edge Detection. Sensors 2023, 23, 1210. [Google Scholar] [CrossRef]
- Yasin, H.M.; Sallow, A.B.; Mahmood, R.Z. High-Speed FPGA-Based Video Watermarking Using LSB Technique in the Spatial Domain. Int. J. Intell. Syst. Appl. Eng. 2024, 12, 644–653. [Google Scholar]
- Kadhim, I.J.; Premaratne, P.; Alwan, Z.A. Invisible Video Watermarking Based on Frame Selection and LSB Embedding. Int. J. Intell. Eng. Syst. 2024, 17, 41–53. [Google Scholar] [CrossRef]
- Gottimukkala, A.R.; Kumar, N.; Dash, J.K.; Swain, G. Image Watermarking Based on Remainder Value Differencing and Extended Hamming Code. J. Electron. Imaging 2023, 33, 011003. [Google Scholar] [CrossRef]
- Venugopala, P.S.; Sarojadevi, H.; Chiplunkar, N.N.; Bhat, V. Video Watermarking by Adjusting the Pixel Values and Using Scene Change Detection. In Proceedings of the 2014 Fifth International Conference on Signal and Image Processing, Bangalore, India, 8–10 January 2014; pp. 259–264. [Google Scholar]
- Yeo, I.-K.; Kim, H.J. Generalized Patchwork Algorithm for Image Watermarking. Multimed. Syst. 2003, 9, 261–265. [Google Scholar] [CrossRef]
- Purna Kumari, B.; Subramanyam Rallabandi, V.P. Modified Patchwork-Based Watermarking Scheme for Satellite Imagery. Signal Process. 2008, 88, 891–904. [Google Scholar] [CrossRef]
- Kang, H.; Yamaguchi, K.; Kurkoski, B.; Yamaguchi, K. Psychoacoustically-Adapted Patchwork Algorithm for Watermarking. In Proceedings of the Third International Conference on Intelligent Information Hiding and Multimedia Signal Processing (IIH-MSP 2007), Kaohsiung, Taiwan, 26–28 November 2007; pp. 267–270. [Google Scholar]
- Ali, M. Robust Image Watermarking in Spatial Domain Utilizing Features Equivalent to SVD Transform. Appl. Sci. 2023, 13, 6105. [Google Scholar] [CrossRef]
- Wang, T. Digital Image Watermarking Using Dual-Scrambling and Singular Value Decomposition. In Proceedings of the 2017 IEEE International Conference on Computational Science and Engineering (CSE) and IEEE International Conference on Embedded and Ubiquitous Computing (EUC), Guangzhou, China, 21–24 July 2017; Volume 1, pp. 724–727. [Google Scholar]
- Wang, H.; Hao, J.; Cui, F. Colour Image Watermarking Algorithm Based on the Arnold Transform. In Proceedings of the 2010 International Conference on Communications and Mobile Computing, Shenzhen, China, 12–14 April 2010; IEEE Computer Society: Washington, DC, USA, 2010; pp. 66–69. [Google Scholar]
- Li, X.; Wang, X.; Yang, W.; Wang, X. A Robust Video Watermarking Scheme to Scalable Recompression and Transcoding. In Proceedings of the 2016 6th International Conference on Electronics Information and Emergency Communication (ICEIEC), Beijing, China, 17–19 June 2016; pp. 257–260. [Google Scholar]
- Li, X.; Wang, X.; Chen, A.; Xiao, L. A Simplified and Robust DCT-Based Watermarking Algorithm. In Proceedings of the 2017 2nd International Conference on Multimedia and Image Processing (ICMIP), Wuhan, China, 17–19 March 2017; pp. 167–171. [Google Scholar]
- Moosazadeh, M.; Ekbatanifard, G. An Improved Robust Image Watermarking Method Using DCT and YCoCg-R Color Space. Optik 2017, 140, 975–988. [Google Scholar] [CrossRef]
- Badran, E.F.; Ghobashy, A.; El-Shennawy, K. DCT-Based Digital Image Watermarking Via Image Segmentation Techniques. In Proceedings of the 2006 ITI 4th International Conference on Information & Communications Technology, Cairo, Egypt, 10–12 December 2006; p. 1. [Google Scholar]
- Alomoush, W.; Khashan, O.A.; Alrosan, A.; Attar, H.H.; Almomani, A.; Alhosban, F.; Makhadmeh, S.N. Digital Image Watermarking Using Discrete Cosine Transformation Based Linear Modulation. J. Cloud Comput. 2023, 12, 96. [Google Scholar] [CrossRef]
- Li, H.; Guo, X. Embedding and Extracting Digital Watermark Based on DCT Algorithm. J. Comput. Commun. 2018, 6, 287–298. [Google Scholar] [CrossRef]
- Xu, Z.J.; Wang, Z.Z.; Lu, Q. Research on Image Watermarking Algorithm Based on DCT. Procedia Environ. Sci. 2011, 10, 1129–1135. [Google Scholar] [CrossRef]
- Senthilkumaran, N.; Abinaya, S. Digital Image Watermarking Using Dft Algorithm. Adv. Comput. Int. J. 2016, 7, 9–17. [Google Scholar] [CrossRef]
- Cedillo-Hernandez, M.; Cedillo-Hernandez, A.; Garcia-Ugalde, F.J. Improving DFT-Based Image Watermarking Using Particle Swarm Optimization Algorithm. Mathematics 2021, 9, 1795. [Google Scholar] [CrossRef]
- Jimson, N.; Hemachandran, K. DFT Based Digital Image Watermarking: A Survey. Int. J. Adv. Res. Comput. Sci. 2018, 9, 540–544. [Google Scholar] [CrossRef][Green Version]
- Pun, C. A Novel DFT-Based Digital Watermarking System for Images. In Proceedings of the 2006 8th international Conference on Signal Processing, Guilin, China, 16–20 November 2006; Volume 2. [Google Scholar]
- Malonia, M.; Agarwal, S.K. Digital Image Watermarking Using Discrete Wavelet Transform and Arithmetic Progression Technique. In Proceedings of the 2016 IEEE Students’ Conference on Electrical, Electronics and Computer Science (SCEECS), Bhopal, India, 5–6 March 2016; pp. 1–6. [Google Scholar]
- Begum, M.; Shorif, S.B.; Uddin, M.S.; Ferdush, J.; Jan, T.; Barros, A.; Whaiduzzaman, M. Image Watermarking Using Discrete Wavelet Transform and Singular Value Decomposition for Enhanced Imperceptibility and Robustness. Algorithms 2024, 17, 32. [Google Scholar] [CrossRef]
- Giri, K.J.; Quadri, S.M.K.; Bashir, R.; Bhat, J.I. DWT Based Color Image Watermarking: A Review. Multimed. Tools Appl. 2020, 79, 32881–32895. [Google Scholar] [CrossRef]
- Ramos, A.M.; Artiles, J.A.P.; Chaves, D.P.B.; Pimentel, C. A Fragile Image Watermarking Scheme in DWT Domain Using Chaotic Sequences and Error-Correcting Codes. Entropy 2023, 25, 508. [Google Scholar] [CrossRef]
- Al-Haj, A. Combined DWT-DCT Digital Image Watermarking. J. Comput. Sci. 2007, 3, 740–746. [Google Scholar] [CrossRef]
- Fazli, S.; Moeini, M. A Robust Image Watermarking Method Based on DWT, DCT, and SVD Using a New Technique for Correction of Main Geometric Attacks. Optik 2016, 127, 964–972. [Google Scholar] [CrossRef]
- Hamidi, M.; El Haziti, M.; Cherifi, H.; El Hassouni, M. A Hybrid Robust Image Watermarking Method Based on DWT-DCT and SIFT for Copyright Protection. J. Imaging 2021, 7, 218. [Google Scholar] [CrossRef]
- Amirgholipour Kasmani, S.; Naghsh-Nilchi, A. Robust Digital Image Watermarking Based on Joint DWT-DCT. JDCTA 2009, 3, 42–54. [Google Scholar] [CrossRef][Green Version]
- Radhi, H.Y.; Yousif, S.F.; Mohamed, W.Q.; Mohammed, A.H. Efficient Image Watermark Scheme Based on DCT-DFT and Singular Value Decomposition Algorithm with Ikeda/Arnold Maps. J. Wuhan Univ. Technol. Mater. Sci. Ed. 2023, 47, 885–898. [Google Scholar]
- Li, L.; Zhang, H.-J.; Fan, H.-Y.; Lu, Z.-M. A DFT and IWT-DCT Based Image Watermarking Scheme for Industry. IEICE Trans. Inf. Syst. 2023, 106, 1916–1921. [Google Scholar] [CrossRef]
- Varghese, J.; Bin Hussain, O.; Subash, S.; Razak, A. An Effective Digital Image Watermarking Scheme Incorporating DCT, DFT and SVD Transformations. PeerJ Comput. Sci. 2023, 9, e1427. [Google Scholar] [CrossRef]
- Ansari, R.; Devanalamath, M.M.; Manikantan, K.; Ramachandran, S. Robust Digital Image Watermarking Algorithm in DWT-DFT-SVD Domain for Color Images. In Proceedings of the 2012 International Conference on Communication, Information & Computing Technology (ICCICT), Mumbai, India, 19–20 October 2012; IEEE Conference Publication; IEEE Xplore. Available online: https://ieeexplore.ieee.org/document/6398160 (accessed on 15 October 2024).
- Chen, A.; Wang, X. An Image Watermarking Scheme Based on DWT and DFT. In Proceedings of the 2017 2nd International Conference on Multimedia and Image Processing (ICMIP), Wuhan, China, 17–19 March 2017; pp. 177–180. [Google Scholar]
- Fullea, E.; Mart”nez, J.M. Robust Digital Image Watermarking Using DWT, DFT and Quality Based Average. In Proceedings of the Ninth ACM International Conference on Multimedia, Ottawa, ON, Canada, 30 September–5 October 2001; pp. 489–491. [Google Scholar]
- Varghese, J.; Hussain, O.B.; Razak, T.A.; Subash, S. A Hybrid Digital Image Watermarking Scheme Incorporating DWT, DFT, DCT and SVD Transformations. J. Eng. Res. 2022, 10, 113–130. [Google Scholar] [CrossRef]
- Li, J.; Sui, A. A Digital Video Watermarking Algorithm Based on DCT Domain. In Proceedings of the 2012 Fifth International Joint Conference on Computational Sciences and Optimization, Harbin, China, 23–26 June 2012; pp. 557–560. [Google Scholar]
- Salih, M.; Ahmed, E.; Abdulkareem Al-Abaji, M. Digital Video Watermarking Methods Using DCT and SVD. J. Educ. Sci. 2019, 29, 266–278. [Google Scholar] [CrossRef]
- Al-Gindy, A.; Omar, A.A.-C.; Mashal, O.; Shaker, Y.; Alhogaraty, E.; Moussa, S. A New Watermarking Scheme for Digital Videos Using DCT. Open Comput. Sci. 2022, 12, 248–259. [Google Scholar] [CrossRef]
- Yang, C.-H.; Huang, H.-Y.; Hsu, W.-H. An Adaptive Video Watermarking Technique Based on DCT Domain. In Proceedings of the 2008 8th IEEE International Conference on Computer and Information Technology, Sydney, NSW, Australia, 8–11 July 2008; pp. 589–594. [Google Scholar]
- Hu, J.C.; Huang, H.Y.; Liang, A.H. Video Watermarking Algorithm Based on DCT and SVD. Appl. Mech. Mater. 2013, 321–324, 2688–2692. [Google Scholar] [CrossRef]
- Rajab, L.; Al-Khatib, T.; Al-Haj, A. A Blind DWT-SCHUR Based Digital Video Watermarking Technique. J. Softw. Eng. Appl. 2015, 8, 224–233. Available online: https://www.scirp.org/journal/paperinformation?paperid=55955 (accessed on 15 October 2024). [CrossRef]
- Sen, C.; Kashyap, T. Digital Video Watermarking Using DWT for Data Security. IJARCCE 2015, 307–309. [Google Scholar] [CrossRef]
- Chan, P.-W.; Lyu, M.R. A DWT-Based Digital Video Watermarking Scheme with Error Correcting Code. In Proceedings of the Information and Communications Security; Qing, S., Gollmann, D., Zhou, J., Eds.; Springer: Huhehaote, China, 2003; pp. 202–213. [Google Scholar]
- Patil, S.A. Digital Video Watermarking Using Dwt and Pca. IOSR J. Eng. 2013, 3, 45–49. [Google Scholar] [CrossRef]
- Liu, Y.; Zhao, J. A New Video Watermarking Algorithm Based on 1D DFT and Radon Transform. Signal Process. 2010, 90, 626–639. [Google Scholar] [CrossRef]
- Sun, X.-C.; Lu, Z.-M.; Wang, Z.; Liu, Y.-L. A Geometrically Robust Multi-Bit Video Watermarking Algorithm Based on 2-D DFT. Multimed. Tools Appl. 2021, 80, 13491–13511. [Google Scholar] [CrossRef]
- Yang, X.; Zhang, Z.; Jiao, Y.; Li, Z. A Robust Video Watermarking Algorithm Based on Two-Dimensional Discrete Fourier Transform. Electronics 2023, 12, 3271. [Google Scholar] [CrossRef]
- Sang, J.; Liu, Q.; Song, C.-L. Robust Video Watermarking Using a Hybrid DCT-DWT Approach. J. Electron. Sci. Technol. 2020, 18, 100052. [Google Scholar] [CrossRef]
- Mawande, S.; Dakhore, H. Video Watermarking Using DWT-DCT-SVD Algorithms. In Proceedings of the 2017 International Conference on Computing Methodologies and Communication (ICCMC), Erode, India, 18–19 July 2017; pp. 1161–1164. [Google Scholar]
- Agarwal, A.; Bhadana, R.; Chavan, S. A Robust Video Watermarking Scheme Using DWT and DCT. Int. J. Comput. Sci. Inf. Technol. IJCSIT 2011, 2, 1711–1716. [Google Scholar]
- Palaiyappan, C.; Raja Jeya Sekhar, T. A Block Based Novel Digital Video Watermarking Scheme Using Dct. IOSR J. Electron. Commun. Eng. 2013, 5, 34–44. [Google Scholar] [CrossRef]
- Zuo, T.; Duan, Y.; Du, Q.; Tao, X. Semantic Security: A Digital Watermark Method for Image Semantic Preservation. In Proceedings of the ICASSP 2024—2024 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Seoul, Republic of Korea, 14–19 April 2024; pp. 4645–4649. [Google Scholar]
- Goodfellow, I.; Bengio, Y.; Courville, A. Deep Learning, 196th ed.; MIT Press: Cambridge, MA, USA, 2016. [Google Scholar]
- LeCun, Y.; Bengio, Y.; Hinton, G. Deep Learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef] [PubMed]
- Dean, J.; Corrado, G.; Monga, R.; Chen, K.; Devin, M.; Le, Q.V.; Mao, M.; Ranzato, M.; Senior, A.; Tucker, P.; et al. Large Scale Distributed Deep Networks. In Proceedings of the Advances in Neural Information Processing Systems, Red Hook, NY, USA, 3–6 December 2012; Curran Associates, Inc.: Red Hook, NY, USA, 2012; Volume 25. [Google Scholar]
- Krizhevsky, A.; Sutskever, I.; Hinton, G.E. ImageNet Classification with Deep Convolutional Neural Networks. In Proceedings of the Advances in Neural Information Processing Systems, Red Hook, NY, USA, 3–6 December 2012; Curran Associates, Inc.: Red Hook, NY, USA, 2012; Volume 25. [Google Scholar]
- Hinton, G.E.; Salakhutdinov, R.R. Reducing the Dimensionality of Data with Neural Networks. Science 2006, 313, 504–507. [Google Scholar] [CrossRef]
- Baldi, P. Autoencoders, Unsupervised Learning, and Deep Architectures. In Proceedings of the ICML Workshop on Unsupervised and Transfer Learning, Bellevue, WA, USA, 2 July 2011; JMLR Workshop and Conference Proceedings. pp. 37–49. [Google Scholar]
- Ayinde, B.O.; Zurada, J.M. Deep Learning of Constrained Autoencoders for Enhanced Understanding of Data. IEEE Trans. Neural Netw. Learn. Syst. 2018, 29, 3969–3979. [Google Scholar] [CrossRef] [PubMed]
- Hosseini-Asl, E.; Zurada, J.M.; Nasraoui, O. Deep Learning of Part-Based Representation of Data Using Sparse Autoencoders with Nonnegativity Constraints. IEEE Trans. Neural Netw. Learn. Syst. 2016, 27, 2486–2498. [Google Scholar] [CrossRef]
- Goodfellow, I.J.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative Adversarial Networks. arXiv 2014, arXiv:1406.2661. [Google Scholar] [CrossRef]
- Radford, A.; Metz, L.; Chintala, S. Unsupervised Representation Learning with Deep Convolutional Generative Adversarial Networks. arXiv 2015, arXiv:1511.06434. [Google Scholar]
- Ho, J.; Jain, A.; Abbeel, P. Denoising Diffusion Probabilistic Models. In Proceedings of the Advances in Neural Information Processing Systems; Curran Associates, Inc.: Red Hook, NY, USA, 2020; Volume 33, pp. 6840–6851. [Google Scholar]
- Nichol, A.Q.; Dhariwal, P. Improved Denoising Diffusion Probabilistic Models. In Proceedings of the 38th International Conference on Machine Learning, Virtual, 18–24 July 2021; pp. 8162–8171. [Google Scholar]
- Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention Is All You Need. In Proceedings of the Advances in Neural Information Processing Systems; Curran Associates, Inc.: Red Hook, NY, USA, 2017; Volume 30. [Google Scholar]
- Hu, J.; Shen, L.; Sun, G. Squeeze-and-Excitation Networks. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 7132–7141. [Google Scholar]
- Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An Image Is Worth 16 × 16 Words: Transformers for Image Recognition at Scale. arXiv 2020, arXiv:2010.11929. [Google Scholar]
- Ben Jabra, S.; Ben Farah, M. Deep Learning-Based Watermarking Techniques Challenges: A Review of Current and Future Trends. Circuits Syst. Signal Process. 2024, 43, 4339–4368. [Google Scholar] [CrossRef]
- Wang, Z.; Byrnes, O.; Wang, H.; Sun, R.; Ma, C.; Chen, H.; Wu, Q.; Xue, M. Data Hiding with Deep Learning: A Survey Unifying Digital Watermarking and Steganography. IEEE Trans. Comput. Soc. Syst. 2023, 10, 2985–2999. [Google Scholar] [CrossRef]
- Zhong, X.; Das, A.; Alrasheedi, F.; Tanvir, A. A Brief, In-Depth Survey of Deep Learning-Based Image Watermarking. Appl. Sci. 2023, 13, 11852. [Google Scholar] [CrossRef]
- Kandi, H.; Mishra, D.; Gorthi, S.R.K.S. Exploring the Learning Capabilities of Convolutional Neural Networks for Robust Image Watermarking. Comput. Secur. 2017, 65, 247–268. [Google Scholar] [CrossRef]
- Mun, S.-M.; Nam, S.-H.; Jang, H.-U.; Kim, D.; Lee, H.-K. A Robust Blind Watermarking Using Convolutional Neural Network. Neurocomputing 2019, 337, 191–202. [Google Scholar] [CrossRef]
- Das, A.; Zhong, X. A Deep Learning-Based Audio-in-Image Watermarking Scheme. In Proceedings of the 2021 International Conference on Visual Communications and Image Processing (VCIP), Munich, Germany, 5–8 December 2021; pp. 1–5. [Google Scholar]
- Zhong, X.; Huang, P.-C.; Mastorakis, S.; Shih, F.Y. An Automated and Robust Image Watermarking Scheme Based on Deep Neural Networks. arXiv 2020, arXiv:2007.02460. [Google Scholar] [CrossRef]
- Tavakoli, A.; Honjani, Z.; Sajedi, H. Convolutional Neural Network-Based Image Watermarking Using Discrete Wavelet Transform. Int. J. Inf. Technol. 2023, 15, 2021–2029. [Google Scholar] [CrossRef]
- Ouyang, C.; Wei, Z. Deep Neural Network-Based Image Watermarking in Wavelet Transform Domain. In Proceedings of the 2023 International Conference on Artificial Intelligence and Automation Control (AIAC), Xiamen, China, 17–19 November 2023; pp. 164–167. [Google Scholar]
- Lu, J.; Ni, J.; Su, W.; Xie, H. Wavelet-Based CNN for Robust and High-Capacity Image Watermarking. In Proceedings of the 2022 IEEE International Conference on Multimedia and Expo (ICME), Taipei, Taiwan, 18–22 July 2022; pp. 1–6. [Google Scholar]
- Ronneberger, O.; Fischer, P.; Brox, T. U-Net: Convolutional Networks for Biomedical Image Segmentation. In Proceedings of the Medical Image Computing and Computer-Assisted Intervention—MICCAI 2015, Munich, Germany, 5–9 October 2015. [Google Scholar]
- Tancik, M.; Mildenhall, B.; Ng, R. StegaStamp: Invisible Hyperlinks in Physical Photographs. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 2114–2123. [Google Scholar]
- Singh, H.K.; Singh, A.K. Digital Image Watermarking Using Deep Learning. Multimed. Tools Appl. 2024, 83, 2979–2994. [Google Scholar] [CrossRef]
- Dasgupta, A.; Zhong, X. Robust Image Watermarking Based on Cross-Attention and Invariant Domain Learning. In Proceedings of the 2023 International Conference on Computational Science and Computational Intelligence (CSCI), Las Vegas, NV, USA, 13–15 December 2023. [Google Scholar]
- Schroff, F.; Kalenichenko, D.; Philbin, J. FaceNet: A Unified Embedding for Face Recognition and Clustering. In Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015; pp. 815–823. [Google Scholar]
- Zhu, J.; Kaplan, R.; Johnson, J.; Fei-Fei, L. HiDDeN: Hiding Data With Deep Networks. arXiv 2018, arXiv:1807.09937. [Google Scholar]
- Wen, B.; Aydore, S. ROMark: A Robust Watermarking System Using Adversarial Training. arXiv 2019, arXiv:1910.01221. [Google Scholar] [CrossRef]
- Hamamoto, I.; Kawamura, M. Neural Watermarking Method Including an Attack Simulator against Rotation and Compression Attacks. IEICE Trans. Inf. Syst. 2020, 103, 33–41. [Google Scholar] [CrossRef]
- Zhang, L.; Li, W.; Ye, H. A Blind Watermarking System Based on Deep Learning Model. In Proceedings of the 2021 IEEE 20th International Conference on Trust, Security and Privacy in Computing and Communications (TrustCom), Shenyang, China, 20–22 October 2021; IEEE: Piscataway, NJ, USA, 2021; pp. 1208–1213. [Google Scholar]
- Zhang, H.; Wang, H.; Cao, Y.; Shen, C.; Li, Y. Robust Data Hiding Using Inverse Gradient Attention. arXiv 2022, arXiv:2011.10850. [Google Scholar] [CrossRef]
- Hao, K.; Feng, G.; Zhang, X. Robust Image Watermarking Based on Generative Adversarial Network. China Commun. 2020, 17, 131–140. [Google Scholar] [CrossRef]
- Huang, J.; Luo, T.; Li, L.; Yang, G.; Xu, H.; Chang, C.-C. ARWGAN: Attention-Guided Robust Image Watermarking Model Based on GAN. IEEE Trans. Instrum. Meas. 2023, 72, 5018417. [Google Scholar] [CrossRef]
- Shedole, S.M.; Santhi, V. Hybrid Deep Learning Based Digital Image Watermarking Using GAN-LSTM and Adaptive Gannet Optimization Techniques. Multimed. Tools Appl. 2024, 84, 20661–20691. [Google Scholar] [CrossRef]
- Karki, B.; Tsai, C.-H.; Huang, P.-C.; Zhong, X. Deep Learning-Based Text-in-Image Watermarking. In Proceedings of the 2024 IEEE 7th International Conference on Multimedia Information Processing and Retrieval (MIPR), San Jose, CA, USA, 7–9 August 2024. [Google Scholar]
- Aberna, P.; Agilandeeswari, L.; Aashich, B. Vision Transformer-Based Watermark Generation for Authentication and Tamper Detection Using Schur Decomposition and Hybrid Transforms. Int. J. Comput. Inf. Syst. Ind. Manag. Appl. 2023, 15, 107–121. [Google Scholar]
- Liu, Z.; Li, Z.; Zheng, L.; Li, D. Two-Stage Robust Lossless DWI Watermarking Based on Transformer Networks in the Wavelet Domain. Appl. Sci. 2023, 13, 6886. [Google Scholar] [CrossRef]
- Luo, T.; Wu, J.; He, Z.; Xu, H.; Jiang, G.; Chang, C.-C. WFormer: A Transformer-Based Soft Fusion Model for Robust Image Watermarking. IEEE Trans. Emerg. Top. Comput. Intell. 2024, 8, 4179–4196. [Google Scholar] [CrossRef]
- Wang, B.; Song, Z.; Wu, Y. Robust Blind Watermarking Framework for Hybrid Networks Combining CNN and Transformer. In Proceedings of the 15th Asian Conference on Machine Learning, Istanbul, Turkey, 27 February 2024; pp. 1417–1432. [Google Scholar]
- Chen, W.; Li, Y. RoWSFormer: A Robust Watermarking Framework with Swin Transformer for Enhanced Geometric Attack Resilience. arXiv 2024, arXiv:2409.14829. [Google Scholar] [CrossRef]
- Liu, Z.; Lin, Y.; Cao, Y.; Hu, H.; Wei, Y.; Zhang, Z.; Lin, S.; Guo, B. Swin Transformer: Hierarchical Vision Transformer Using Shifted Windows. arXiv 2021, arXiv:2103.14030. [Google Scholar] [CrossRef]
- Cui, Y.; Ren, J.; Xu, H.; He, P.; Liu, H.; Sun, L.; Xing, Y.; Tang, J. DiffusionShield: A Watermark for Copyright Protection against Generative Diffusion Models. arXiv 2024, arXiv:2306.04642. [Google Scholar] [CrossRef]
- Wen, Y.; Kirchenbauer, J.; Geiping, J.; Goldstein, T. Tree-Ring Watermarks: Fingerprints for Diffusion Images That Are Invisible and Robust. arXiv 2023, arXiv:2305.20030. [Google Scholar] [CrossRef]
- Sha, Z.; Li, Z.; Yu, N.; Zhang, Y. DE-FAKE: Detection and Attribution of Fake Images Generated by Text-to-Image Generation Models. arXiv 2023, arXiv:2210.06998. [Google Scholar]
- Zhang, L.; Liu, X.; Martin, A.V.; Bearfield, C.X.; Brun, Y.; Guan, H. Attack-Resilient Image Watermarking Using Stable Diffusion. arXiv 2024, arXiv:2401.04247. [Google Scholar]
- Rombach, R.; Blattmann, A.; Lorenz, D.; Esser, P.; Ommer, B. High-Resolution Image Synthesis with Latent Diffusion Models. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022. [Google Scholar]
- Tan, Y.; Peng, Y.; Fang, H.; Chen, B.; Xia, S.-T. WaterDiff: Perceptual Image Watermarks Via Diffusion Model. In Proceedings of the ICASSP 2024—2024 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Seul, Republic of Korea, 14–19 April 2024; pp. 3250–3254. [Google Scholar]
- Hu, R.; Zhang, J.; Li, Y.; Li, J.; Guo, Q.; Qiu, H.; Zhang, T. SuperMark: Robust and Training-Free Image Watermarking via Diffusion-Based Super-Resolution. arXiv 2024, arXiv:2412.10049. [Google Scholar]
- Bui, T.; Agarwal, S.; Collomosse, J. TrustMark: Universal Watermarking for Arbitrary Resolution Images. arXiv 2023, arXiv:2311.18297. [Google Scholar] [CrossRef]
- Wang, Y.; Zhu, X.; Ye, G.; Zhang, S.; Wei, X. Achieving Resolution-Agnostic DNN-Based Image Watermarking: A Novel Perspective of Implicit Neural Representation. In Proceedings of the 32nd ACM International Conference on Multimedia, New York, NY, USA, 28 October–1 November 2024; pp. 10354–10362. [Google Scholar]
- Kutter, M.; Petitcolas, F.A.P. Fair Benchmark for Image Watermarking Systems. In Proceedings of the Security and Watermarking of Multimedia Contents; SPIE: Bellingham, WA, USA, 1999; Volume 3657, pp. 226–239. [Google Scholar]
- Duszejko, P.; Piotrowski, Z. 4KSecure: A Universal Method for Active Manipulation Detection in Images of Any Resolution. Appl. Sci. 2025, 15, 4469. [Google Scholar] [CrossRef]
- Numajiri, H.; Hayashi, T. Analysis on Open Data as a Foundation for Data-Driven Research. Scientometrics 2024, 129, 6315–6332. [Google Scholar] [CrossRef]
- Hadhoud, M.M.; Abd El-Samie, F.; El-Khamy, S.E. New Trends in High Resolution Image Processing. In Proceedings of the Fourth Workshop on Photonics and Its Application, Giza, Egypt, 4 May 2004; pp. 2–23. [Google Scholar]
- Satone, K.N.; Deshmukh, A.S.; Ulhe, P.B. A Review of Image Compression Techniques. In Proceedings of the 2017 International conference of Electronics, Communication and Aerospace Technology (ICECA), Coimbatore, India, 20–22 April 2017; Volume 1, pp. 97–101. [Google Scholar]
- Ramesh, A.; Pavlov, M.; Goh, G.; Gray, S.; Voss, C.; Radford, A.; Chen, M.; Sutskever, I. Zero-Shot Text-to-Image Generation. arXiv 2021, arXiv:2102.12092. [Google Scholar]
- Dathathri, S.; See, A.; Ghaisas, S.; Huang, P.-S.; McAdam, R.; Welbl, J.; Bachani, V.; Kaskasoli, A.; Stanforth, R.; Matejovicova, T.; et al. Scalable Watermarking for Identifying Large Language Model Outputs. Nature 2024, 634, 818–823. [Google Scholar] [CrossRef]
- Deng, J.; Dong, W.; Socher, R.; Li, L.-J.; Li, K.; Fei-Fei, L. ImageNet: A Large-Scale Hierarchical Image Database. In Proceedings of the 2009 IEEE Conference on Computer Vision and Pattern Recognition, Miami, FL, USA, 20–25 June 2009; pp. 248–255. [Google Scholar]
- Lin, T.-Y.; Maire, M.; Belongie, S.; Hays, J.; Perona, P.; Ramanan, D.; Dollár, P.; Zitnick, C.L. Microsoft COCO: Common Objects in Context. In Proceedings of the Computer Vision—ECCV 2014; Fleet, D., Pajdla, T., Schiele, B., Tuytelaars, T., Eds.; Springer International Publishing: Cham, Switzerland, 2014; pp. 740–755. [Google Scholar]
- Krizhevsky, A. Learning Multiple Layers of Features from Tiny Images; University of Toronto: Toronto, ON, Canada, 2009. [Google Scholar]
- Everingham, M.; Van Gool, L.; Williams, C.K.I.; Winn, J.; Zisserman, A. The Pascal Visual Object Classes (VOC) Challenge. Int. J. Comput. Vis. 2010, 88, 303–338. [Google Scholar] [CrossRef]
- Bas, P.; Filler, T.; Pevný, T. “Break Our Steganographic System”: The Ins and Outs of Organizing BOSS. In Proceedings of the 13th International Conference on Information Hiding, Prague, Czech Republic, 18–20 May 2011; Springer: Berlin/Heidelberg, Germany, 2011; pp. 59–70. [Google Scholar]
- Agustsson, E.; Timofte, R. NTIRE 2017 Challenge on Single Image Super-Resolution: Dataset and Study. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Honolulu, HI, USA, 21–26 July 2017; pp. 1122–1131. [Google Scholar]
- Liu, J.; Liu, D.; Yang, W.; Xia, S.; Zhang, X.; Dai, Y. A Comprehensive Benchmark for Single Image Compression Artifact Reduction. IEEE Trans. Image Process. 2020, 29, 7845–7860. [Google Scholar] [CrossRef]
- Zhang, K.; Li, D.; Luo, W.; Ren, W.; Stenger, B.; Liu, W.; Li, H.; Yang, M.-H. Benchmarking Ultra-High-Definition Image Super-Resolution. In Proceedings of the 2021 IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, QC, Canada, 10–17 October 2021; IEEE: Piscataway, NJ, USA; pp. 14749–14758. [Google Scholar]
- Schuhmann, C.; Beaumont, R.; Vencu, R.; Gordon, C.; Wightman, R.; Cherti, M.; Coombes, T.; Katta, A.; Mullis, C.; Wortsman, M.; et al. LAION-5B: An Open Large-Scale Dataset for Training next Generation Image-Text Models. arXiv 2022, arXiv:2210.08402. [Google Scholar]
- Bird, J.J.; Lotfi, A. CIFAKE: Image Classification and Explainable Identification of AI-Generated Synthetic Images. IEEE Access 2024, 12, 15642–15650. [Google Scholar] [CrossRef]
- Rahman, M.A.; Paul, B.; Sarker, N.H.; Hakim, Z.I.A.; Fattah, S.A. Artifact: A Large-Scale Dataset with Artificial and Factual Images for Generalizable and Robust Synthetic Image Detection. In Proceedings of the 2023 IEEE International Conference on Image Processing (ICIP), Kuala Lumpur, Malaysia, 8–11 October 2023; pp. 2200–2204. [Google Scholar]
- Boychev, D.; Cholakov, R. ImagiNet: A Multi-Content Benchmark for Synthetic Image Detection. arXiv 2024, arXiv:2407.20020. [Google Scholar]
- Wang, X.; Peng, Y.; Lu, L.; Lu, Z.; Bagheri, M.; Summers, R.M. ChestX-Ray8: Hospital-Scale Chest X-Ray Database and Benchmarks on Weakly-Supervised Classification and Localization of Common Thorax Diseases. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 3462–3471. [Google Scholar]
- Helber, P.; Bischke, B.; Dengel, A.; Borth, D. EuroSAT: A Novel Dataset and Deep Learning Benchmark for Land Use and Land Cover Classification. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2019, 12, 2217–2226. [Google Scholar] [CrossRef]
- Knoche, M.; Hörmann, S.; Rigoll, G. Cross-Quality LFW: A Database for Analyzing Cross-Resolution Image Face Recognition in Unconstrained Environments. In Proceedings of the 2021 16th IEEE International Conference on Automatic Face and Gesture Recognition (FG 2021), Jodhpur, India, 15–18 December 2021; pp. 1–5. [Google Scholar]
- Sood, A.; Mishra, D.; Surya, V.; Singh, H.; Sundaresan, R.; Pal, D.; Dharmaraju, R.; Satish, R.; Mishra, S.; Chavan, N.A.; et al. Challenges and Recommendations for Enhancing Digital Data Protection in Indian Medical Research and Healthcare Sector. Npj Digit. Med. 2025, 8, 48. [Google Scholar] [CrossRef] [PubMed]
- Barni, M.; Bartolini, F. Watermarking Systems Engineering: Enabling Digital Assets Security and Other Applications (Signal Processing and Communications); CRC: Boca Raton, FL, USA, 2004; ISBN 978-0-8247-4806-7. [Google Scholar]
- Pibre, L.; Jérôme, P.; Ienco, D.; Chaumont, M. Deep Learning Is a Good Steganalysis Tool When Embedding Key Is Reused for Different Images, Even If There Is a Cover Source-Mismatch. arXiv 2015, arXiv:1511.04855. [Google Scholar] [CrossRef]
- Yang, L.; Zhang, Z.; Song, Y.; Hong, S.; Xu, R.; Zhao, Y.; Zhang, W.; Cui, B.; Yang, M.-H. Diffusion Models: A Comprehensive Survey of Methods and Applications. ACM Comput. Surv. 2023, 56, 1–39. [Google Scholar] [CrossRef]
- Fuest, M.; Ma, P.; Gui, M.; Schusterbauer, J.; Hu, V.T.; Ommer, B. Diffusion Models and Representation Learning: A Survey. arXiv 2024, arXiv:2407.00783. [Google Scholar] [CrossRef]
- Krichen, M.; Abdalzaher, M.S. Performance Enhancement of Artificial Intelligence: A Survey. J. Netw. Comput. Appl. 2024, 232, 104034. [Google Scholar] [CrossRef]
- Zhou, Y.; Chen, S.; Wang, Y.; Huan, W. Review of Research on Lightweight Convolutional Neural Networks. In Proceedings of the 2020 IEEE 5th Information Technology and Mechatronics Engineering Conference (ITOEC), Chongqing, China, 12–14 June 2020; pp. 1713–1720. [Google Scholar]
- Howard, A.G.; Zhu, M.; Chen, B.; Kalenichenko, D.; Wang, W.; Weyand, T.; Andreetto, M.; Adam, H. MobileNets: Efficient Convolutional Neural Networks for Mobile Vision Applications. arXiv 2017, arXiv:1704.04861. [Google Scholar] [CrossRef]
- Iandola, F.N.; Han, S.; Moskewicz, M.W.; Ashraf, K.; Dally, W.J.; Keutzer, K. SqueezeNet: AlexNet-Level Accuracy with 50× Fewer Parameters and <0.5 MB Model Size. arXiv 2016, arXiv:1602.07360. [Google Scholar]
- Cheng, H.; Zhang, M.; Shi, J.Q. A Survey on Deep Neural Network Pruning-Taxonomy, Comparison, Analysis, and Recommendations. arXiv 2024, arXiv:2308.06767. [Google Scholar] [CrossRef] [PubMed]
- Ma, X.; Lin, S.; Ye, S.; He, Z.; Zhang, L.; Yuan, G.; Tan, S.H.; Li, Z.; Fan, D.; Qian, X.; et al. Non-Structured DNN Weight Pruning—Is It Beneficial in Any Platform? IEEE Trans. Neural Netw. Learn. Syst. 2021, 33, 4930–4944. [Google Scholar] [CrossRef]
- Nagel, M.; Fournarakis, M.; Amjad, R.A.; Bondarenko, Y.; van Baalen, M.; Blankevoort, T. A White Paper on Neural Network Quantization. arXiv 2021, arXiv:2106.08295. [Google Scholar] [CrossRef]
- Chen, Z.; Tan, X.; Wang, K.; Pan, S.; Mandic, D.; He, L.; Zhao, S. InferGrad: Improving Diffusion Models for Vocoder by Considering Inference in Training. arXiv 2022, arXiv:2202.03751. [Google Scholar] [CrossRef]
- Shih, A.; Belkhale, S.; Ermon, S.; Sadigh, D.; Anari, N. Parallel Sampling of Diffusion Models. arXiv 2023, arXiv:2305.16317. [Google Scholar] [CrossRef]
- Mahon, S.; Varrette, S.; Plugaru, V.; Pinel, F.; Bouvry, P. Performance Analysis of Distributed and Scalable Deep Learning. In Proceedings of the 2020 20th IEEE/ACM International Symposium on Cluster, Cloud and Internet Computing (CCGRID), Melbourne, VIC, Australia, 11–14 May 2020; pp. 760–766. [Google Scholar]
- Pham, M.; Li, H.; Yuan, Y.; Mou, C.; Ramachandran, K.; Xu, Z.; Tu, Y. Dynamic Memory Management in Massively Parallel Systems: A Case on GPUs. In Proceedings of the 36th ACM International Conference on Supercomputing; Association for Computing Machinery: New York, NY, USA, 2022; pp. 1–13. [Google Scholar]
- Sakthidevi, I.; Rajkumar, G.V.; Sunitha, R.; Sangeetha, A.; Krishnan, R.S.; Sundararajan, S. Machine Learning Orchestration in Cloud Environments: Automating the Training and Deployment of Distributed Machine Learning AI Model. In Proceedings of the 2023 7th International Conference on I-SMAC (IoT in Social, Mobile, Analytics and Cloud) (I-SMAC), Kirtipur, Nepal, 11–13 October 2023; pp. 376–384. [Google Scholar]
- Jangda, A.; Pinckney, D.; Brun, Y.; Guha, A. Formal Foundations of Serverless Computing. Proc. ACM Program. Lang. 2019, 3, 149. [Google Scholar] [CrossRef]
- Liang, S.; Jin, S.; Chen, Y. A Review of Edge Computing Technology and Its Applications in Power Systems. Energies 2024, 17, 3230. [Google Scholar] [CrossRef]
- Andriulo, F.C.; Fiore, M.; Mongiello, M.; Traversa, E.; Zizzo, V. Edge Computing and Cloud Computing for Internet of Things: A Review. Informatics 2024, 11, 71. [Google Scholar] [CrossRef]
- Zeyu, H.; Geming, X.; Zhaohang, W.; Sen, Y. Survey on Edge Computing Security. In Proceedings of the 2020 International Conference on Big Data, Artificial Intelligence and Internet of Things Engineering (ICBAIE), Virtual Event, 28–30 June 2020; pp. 96–105. [Google Scholar]
- Yan, Z.; Yi, Y.Z.; JiLin, Z.; NaiLiang, Z.; YongJian, R.; Jian, W.; Jun, Y. Federated Learning Model Training Method Based on Data Features Perception Aggregation. In Proceedings of the 2021 IEEE 94th Vehicular Technology Conference (VTC2021-Fall), Norman, OK, USA, 27–30 September 2021; pp. 1–7. [Google Scholar]
- Sáinz-Pardo Díaz, J.; López García, Á. Study of the Performance and Scalability of Federated Learning for Medical Imaging with Intermittent Clients. Neurocomputing 2023, 518, 142–154. [Google Scholar] [CrossRef]
- de Goede, M.; Cox, B.; Decouchant, J. Training Diffusion Models with Federated Learning. arXiv 2024, arXiv:2406.12575. [Google Scholar] [CrossRef]
- Qi, Y.; Feng, Y.; Wang, X.; Li, H.; Tian, J. Leveraging Federated Learning and Edge Computing for Recommendation Systems within Cloud Computing Networks. arXiv 2024, arXiv:2403.03165. [Google Scholar] [CrossRef]
- Oubadriss, A.; Laassiri, J.; Makrani, A.E. An Overview Comparison between Convolutional Neural Networks and Vision Transformers. In Proceedings of the 7th International Conference on Networking, Intelligent Systems and Security, Ado Ekiti, Nigeria, 26–28 November 2024; Association for Computing Machinery: New York, NY, USA, 2024; pp. 1–9. [Google Scholar]
- Chen, M.; Mei, S.; Fan, J.; Wang, M. Opportunities and Challenges of Diffusion Models for Generative AI. Natl. Sci. Rev. 2024, 11, nwae348. [Google Scholar] [CrossRef] [PubMed]
- Jamil, S.; Jalil Piran, M.; Kwon, O.-J. A Comprehensive Survey of Transformers for Computer Vision. Drones 2023, 7, 287. [Google Scholar] [CrossRef]
- von Krause, M.; Radev, S.T.; Voss, A.; Quintus, M.; Egloff, B.; Wrzus, C. Stability and Change in Diffusion Model Parameters over Two Years. J. Intell. 2021, 9, 26. [Google Scholar] [CrossRef]
- Ravi, A.; Chaturvedi, V.; Shafique, M. ViT4Mal: Lightweight Vision Transformer for Malware Detection on Edge Devices. ACM Trans. Embed. Comput. Syst. 2023, 22, 117. [Google Scholar] [CrossRef]
- Nag, S.; Liberty, L.; Sivakumar, A.; Yadwadkar, N.J.; John, L.K. Lightweight Vision Transformers for Low Energy Edge Inference. In Proceedings of the Machine Learning for Computer Architecture and Systems 2024, Buenos Aires, Argentina, 30 June 2024. [Google Scholar]
- Setyawan, N.; Sun, C.-C.; Hsu, M.-H.; Kuo, W.-K.; Hsieh, J.-W. MicroViT: A Vision Transformer with Low Complexity Self Attention for Edge Device. arXiv 2025, arXiv:2502.05800. [Google Scholar]
- Tang, H.; Chen, Y.; Wang, T.; Zhou, Y.; Zhao, L.; Gao, Q.; Du, M.; Tan, T.; Zhang, X.; Tong, T. HTC-Net: A Hybrid CNN-Transformer Framework for Medical Image Segmentation. Biomed. Signal Process. Control 2024, 88, 105605. [Google Scholar] [CrossRef]
- Liu, X.; Liu, J.; Bai, Y.; Gu, J.; Chen, T.; Jia, X.; Cao, X. Watermark Vaccine: Adversarial Attacks to Prevent Watermark Removal. arXiv 2022, arXiv:2207.08178. [Google Scholar] [CrossRef]
- Bistroń, M.; Piotrowski, Z. Efficient Video Watermarking Algorithm Based on Convolutional Neural Networks with Entropy-Based Information Mapper. Entropy 2023, 25, 284. [Google Scholar] [CrossRef] [PubMed]
- Xu, T.; Mi, P.; Wang, R.; Chen, Y. Towards Faster Training of Diffusion Models: An Inspiration of A Consistency Phenomenon. arXiv 2024. [Google Scholar] [CrossRef]
- Li, C.; Zhang, C.; Xu, W.; Xie, J.; Feng, W.; Peng, B.; Xing, W. LatentSync: Audio Conditioned Latent Diffusion Models for Lip Sync. arXiv 2024, arXiv:2412.09262. [Google Scholar] [CrossRef]
- Fei, Z.; Fan, M.; Yu, C.; Li, D.; Zhang, Y.; Huang, J. Dimba: Transformer-Mamba Diffusion Models. arXiv 2024, arXiv:2406.01159. [Google Scholar] [CrossRef]
- Liu, W.; Zhang, S.Q. HQ-DiT: Efficient Diffusion Transformer with FP4 Hybrid Quantization. arXiv 2024, arXiv:2405.19751. [Google Scholar]
- De Falco, F.; Ceschini, A.; Sebastianelli, A.; Le Saux, B.; Panella, M. Quantum Hybrid Diffusion Models for Image Synthesis. KI–Künstl. Intell. 2024, 38, 311–326. [Google Scholar] [CrossRef]
- Uehira, K.; Suzuki, K.; Ikeda, H. Applications of Optoelectronic Watermarking Technique to New Business and Industry Systems Utilizing Flat-Panel Displays and Smart Devices. In Proceedings of the 2014 IEEE Industry Application Society Annual Meeting, Vancouver, BC, Canada, 5–9 October 2014; pp. 1–9. [Google Scholar]
- Juefei-Xu, F.; Wang, R.; Huang, Y.; Guo, Q.; Ma, L.; Liu, Y. Countering Malicious DeepFakes: Survey, Battleground, and Horizon. Int. J. Comput. Vis. 2022, 130, 1678–1734. [Google Scholar] [CrossRef]
- Hameleers, M.; van der Meer, T.G.L.A.; Dobber, T. Distorting the Truth versus Blatant Lies: The Effects of Different Degrees of Deception in Domestic and Foreign Political Deepfakes. Comput. Hum. Behav. 2024, 152, 108096. [Google Scholar] [CrossRef]
- Zhao, X.; Gunn, S.; Christ, M.; Fairoze, J.; Fabrega, A.; Carlini, N.; Garg, S.; Hong, S.; Nasr, M.; Tramer, F.; et al. SoK: Watermarking for AI-Generated Content. arXiv 2024, arXiv:2411.18479. [Google Scholar]
- SynthID. Available online: https://deepmind.google/models/synthid/ (accessed on 20 November 2025).
- Agbaje, M.; Awodele, O.; Ogbonna, C. Applications of Digital Watermarking to Cyber Security (Cyber Watermarking). In Proceedings of the Informing Science & IT Education Conference (InSITE), Tampa, FL, USA, 29 June–5 July 2015; pp. 1–11. [Google Scholar]
- Dinesh Arokia Raj, A.; Jha, R.R.; Yadav, M.; Sam, D.; Jayanthi, K. Role of Blockchain and Watermarking Toward Cybersecurity. In Multimedia Watermarking: Latest Developments and Trends; Kumar Sahu, A., Ed.; Springer Nature: Singapore, 2024; pp. 103–123. ISBN 978-981-99-9803-6. [Google Scholar]
- Blake, S. Embedded Blockchains: A Synthesis of Blockchains, Spread Spectrum Watermarking, Perceptual Hashing & Digital Signatures. arXiv 2020, arXiv:2009.00951. [Google Scholar]
- Singh, B.; Sharma, M.K. Efficient Watermarking Technique for Protection and Authentication of Document Images. Multimed. Tools Appl. 2022, 81, 22985–23005. [Google Scholar] [CrossRef]
- Kapre, B.S.; Rajurkar, A.M.; Guru, D.S. The Blind Robust Video Watermarking Scheme in Video Surveillance Context. Multimed. Tools Appl. 2024, 83, 38999–39025. [Google Scholar] [CrossRef]
- Wazirali, R.; Ahmad, R.; Al-Amayreh, A.; Al-Madi, M.; Khalifeh, A. Secure Watermarking Schemes and Their Approaches in the IoT Technology: An Overview. Electronics 2021, 10, 1744. [Google Scholar] [CrossRef]











| Group of Attacks | Type of Attacks | Description | Examples of Operations | Effect on the Watermark |
|---|---|---|---|---|
| Untargeted attacks | - | Attacks resulting from routine processing of the media, with no intention of removing the watermark, but which may affect its integrity. | Lossy compression (JPEG, MPEG), scaling, filtering. | Partial loss or distortion of the watermark. |
| Targeted attacks | General | Intentional manipulation of the media to remove, distort, or weaken the watermark. | Rotate, scale, change resolution. | Total or partial loss of the watermark. |
| Targeted attacks | Statistical attacks | Modifications using statistical analysis of the media to identify and remove the watermark. | Histogram attack, frequency distribution analysis, autocorrelation attack. | Removal of the watermark without significant changes in the perception of the media. |
| Targeted attacks | Sensitivity attacks | Minimal modifications to the media that do not affect the visual quality, but destroy the watermark. | Bit depth reduction, delicate pixel changes. | Distortion or complete loss of the watermark without visible changes in the media. |
| Targeted attacks | Destructive compression | Aggressive compression to remove the watermark by extremely reducing the media data. | High-loss JPEG compression. | Significant loss of data, complete destruction of the watermark, degradation of media quality. |
| Targeted attacks | Geometry attacks | Manipulations in the spatial structure of the media that distort the position of the watermark. | Rotation, translation, change of proportions. | Disturbance of the watermark position, loss of synchronization. |
| Deep learning-based attacks | Generative attacks | Use of generative models to regenerate media content and remove embedded watermark. | Image inpainting, deepfake generation, AI-based restoration. | Complete removal of the watermark without perceptual changes. |
| Deep learning-based attacks | Adversarial attacks | Modifications generated by neural networks to fool detection systems and weaken watermark extraction. | Adversarial noise, gradient-based attacks (FGSM, PGD). | Degradation or undetectability of the watermark. |
| Deep learning-based attacks | Neural network removal | Use of DL models trained to detect and remove watermarks. | CNN-based watermark removal, encoder–decoder architectures. | High probability of the watermark elimination with minimal distortion. |
| Deep learning-based attacks | Content replacement | Media content is regenerated using deep learning models to overwrite or bypass the watermark layer. | GAN-based texture replacement, style transfer techniques. | Loss or severe weakening of the watermark. |
| Deep learning-based attacks | Latent-space attacks | Attacks that exploit generative models (diffusion or VAE-based) to regenerate content in the latent space, effectively removing embedded watermarks. | Stable Diffusion regeneration, DERO [52], VAE sampling attacks [53]. | Total removal of the watermark, especially in latent-domain watermarking schemes. |
| Year | Reference | Architecture and Technology | Watermark Capacity | Host Image Resolution | Watermark Robustness |
|---|---|---|---|---|---|
| 2017 | [133] | CNN as autoencoder | 64 × 64 pixels | 128 × 128 pixels | Noise, cropping, JPEG compression, rotation |
| 2017 | [134] | CNN with residual blocks | 4096 bits | 512 × 512 pixels | Affine transform, cropping, JPEG compression, filtering, rotation, rescaling |
| 2021 | [135] | CNN and LSTM for audio mapping | 8192 audio samples | 128 × 128 pixels | Not implemented |
| 2021 | [136] | CNN and fully connected Invariance Layer | 32 × 32 pixels/1024 bits | 128 × 128 pixels | Noise, cropping, blur, JPEG compression |
| 2023 | [137] | CNN with DWT as preprocessing | 256 bits | 256 × 256 pixels | Noise, dropout, JPEG compression |
| 2023 | [138] | CNN with DWT as preprocessing | 32 × 32 pixels/1024 bits | 512 × 512 pixels | Noise, sharpening, smoothing, dropout, JPEG compression |
| 2022 | [139] | CNN autoencoder with DWT and IDWT blocks | 50–700 bits | 400 × 400 pixels | Perspective warp, motion, blur, noise, color manipulation, JPEG compression |
| 2020 | [141] | CNN autoencoder | 100 bits | 400 × 400 pixels | Perspective warp, camera misalignment, blur, color distortion, noise, JPEG compression |
| 2023 | [142] | CNN autoencoder and CNN denoising autoencoder | 32 × 32 pixels/1024 bits | 128 × 128 pixels | Noise, rotation, JPEG compression |
| 2023 | [143] | CNN with MHA in Invariant Domain | 8 × 8 pixels/64 bits | 128 × 128 pixels | Horizontal flip, blur, solarization, brightness adjustment, contrast variation, hue and saturation modulation |
| 2018 | [145] | GAN | 30 bits | 128 × 128 pixels (training) 512 × 512 pixels (testing) | JPEG compression, blur, cropping, dropout |
| 2019 | [146] | GAN with min-max optimization | 30 bits | 128 × 128 pixels | Cropping, cropout, dropout, blur, JPEG compression and combinations |
| 2020 | [147] | GAN | 64 bits | 512 × 512 pixels | Rotation, JPEG compression, noise, cropping, blur, brightness adjustment |
| 2021 | [148] | GAN | 64 bits | 256 × 256 pixels | JPEG compression, rotation, noise, blur cropping, brightness and contrast adjustment, color inversion |
| 2020 | [149] | GAN and IGA | 256 bits | 256 × 256 pixels | Cropping, dropout, JPEG compression, resizing |
| 2020 | [150] | GAN and attention | 30 bits | 64 × 64 pixels | Cropping, cropout, blur, flip, JPEG compression |
| 2023 | [151] | attention module, GAN, feature fusion | 30 bits | 128 × 128 pixels | Cropping, dropout, blur, JPEG compression, resizing |
| 2024 | [152] | GAN-LSTM, Adaptive Gannet Optimization | 256 × 256 to 1024 × 1024 pixels | 256 × 256 to 1024 × 1024 pixels | Noise, median filtering, blur, JPEG compression, cropping, rotation, scaling |
| 2024 | [153] | Transformer and ViT | 16-word segments | 224 × 224 pixels | JPEG compression, noise, rotation, cropping |
| 2023 | [154] | ViT | 128 bits | 256 × 256 pixels | Noise, median filtering, rotation, scaling |
| 2023 | [155] | Transformer with DWT preprocessing | binary image 24 × 24 pixels | 96 × 96 pixels | Median and gaussian filtering, noise, SPN, JPEG compression, rotation, cropping, scaling |
| 2023 | [156] | Transformer, GAN | 36 to 100 bits | 128 × 128 pixels | Noise, cropout, dropout, JPEG compression, affine transformation |
| 2023 | [157] | Swin Transformer, CNN, MA-FFM, Identity module | 64 bits | 128 × 128 pixels | Cropping, noise, dropout, gaussian and median filtering, JPEG compression |
| 2024 | [158] | Swin Transformer with DCT attention block | 64 bits | 128 × 128 pixels | Cropout, dropout, rotation, scaling, affine transform |
| 2024 | [163] | Stable Diffusion | 32 bits | 64 × 64 × 4 (latent) | JPEG compression, rotation, noise, blur, generative attacks |
| 2024 | [165] | Diffusion Probabilistic Model | 16384 bits | 128 × 128 pixels | JPEG compression, regeneration attacks |
| 2024 | [166] | Diffusion Probabilistic Model | 32 bits | 512 × 512 pixels | JPEG compression, blur, noise, cropping, brightness adjustment, adaptive attacks |
| Name | Category | Number of Images | Resolution [Pixels] | Type | Limitations/ Notes | Reference |
|---|---|---|---|---|---|---|
| ImageNet | Benchmark | 14 million | ~224 × 224 to 256 × 256 (resized) | Animals, vehicles, plants, tools | Large scale, requires preprocessing, good for robustness tests. | [176] |
| COCO | Benchmark | 330,000 | 640 × 480 | People, vehicles, food, animals | Moderate resolution, suitable for general-training. | [177] |
| CIFAR-10/CIFAR-100 | Benchmark | 60,000/100,000 | 32 × 32 | Animals, vehicles | Very low resolution, not suitable for perceptual metrics, useful for capacity tests. | [178] |
| Pascal VOC | Benchmark | 21,000 | 500 × 375 | Animals, vehicles, people | Limited scale and resolution, used in simple robustness evaluations. | [179] |
| BOSSbase | Benchmark | 10,000 | 512 × 512 | Grayscale natural images | Designed for steganalysis, great for statistical robustness tests. | [180] |
| DIV2K | High resolution | 1000 | ~2K (e.g., 2040 × 1080) | Landscape, buildings, architecture, | High quality, ideal for transparency tests. | [181] |
| Flickr2K | High resolution | 2650 | ~2K (e.g., 2040 × 1350) | Natural photos: portraits, landscapes, | Unprocessed, variable quality, useful for perceptual metrics. | [181] |
| LIU4K | High resolution | 2100 | 4K (3840 × 2160) | Different background and objects | High resolution, good for visual quality and real-world simulations. | [182] |
| UHD4K | High resolution | 5000+ | 4K (3840 × 2160) | Satellite images, films, urban scenes | Very high resolution, good for high-end use cases. | [183] |
| UHD8K | High resolution | 2966 | 8K (7680 × 4320) | Satellite images, films, urban scenes | Extremely high resolution, useful for stress testing. | [183] |
| LAION-5B | Synthetic | 5.85 billion | from 256 × 256 to 4K | Images paired with text prompts (mixture of real and AI-generated) | Unprocessed and noisy, not ideal for reproducible benchmarking. | [184] |
| CIFAKE | Synthetic | 120,000 | 32 × 32 | Real images from CIFAR 10 and synthetic images | Low resolution, designed for deepfake detection benchmarks. | [185] |
| ArtiFact | Synthetic | 1.5 million | from 256 × 256 to 1024 × 1024 | People, animals, vehicles, artworks | Moderate resolution, good for testing synthetic distortions. | [186] |
| ImagiNet | Synthetic | 200,000 | from 256 × 256 to 2K | Photos, painting | Well-balanced synthetic content, useful for hybrid real/synthetic training. | [187] |
| NIH Chest X-ray | Specialized (medical) | 112,000 | 1024 × 1024 | Chest X-rays | Suitable for medical robustness/embedding studies. | [188] |
| EuroSAT | Specialized (satellite images) | 27,000 | 64 × 64 | Satellite images: forests, urban areas, fields | Low resolution, useful for satellite-specific tests. | [189] |
| LFW (Labeled Faces in the Wild) | Specialized (faces) | 13,000 | 250 × 250 | Facial photos in natural conditions | Standard for face datasets, useful for privacy, detection, and watermarking on identity data. | [190] |
| Feature | CNN | ViT | Swin Transformer |
|---|---|---|---|
| Feature processing method | Local (by convolutional filters) | Global (by self-attention) | Local and global (by shifted windows) |
| Resistance to traditional attacks | Medium | High | High |
| Resistance to generative attacks | Medium | High | Very high |
| Computational complexity | Low to medium | High | Medium (optimized) |
| Ability to capture context | Limited | High | High (with local optimization) |
| Scalability to high resolution | Limited | Limited (without optimization) | High |
| Potential in watermarking | Well verified but limited | High (based on previous research) | Very high |
| Feature | GAN | Diffusion Model |
|---|---|---|
| Training stability | Low (frequent convergence problems) | High |
| Quality of generated images | High | Very high |
| Generation time | Relatively short | Longer (if no optimization methods were used) |
| Resistance to attacks | Medium | High |
| Computational complexity | Medium | High |
| Possibility of embedding in latent space | Limited | Yes |
| Potential in watermarking | Well verified but limited | High (based on previous research) |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2026 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license.
Share and Cite
Bistroń, M.; Żurada, J.M.; Piotrowski, Z. Deep Learning for Image Watermarking: A Comprehensive Review and Analysis of Techniques, Challenges, and Applications. Sensors 2026, 26, 444. https://doi.org/10.3390/s26020444
Bistroń M, Żurada JM, Piotrowski Z. Deep Learning for Image Watermarking: A Comprehensive Review and Analysis of Techniques, Challenges, and Applications. Sensors. 2026; 26(2):444. https://doi.org/10.3390/s26020444
Chicago/Turabian StyleBistroń, Marta, Jacek M. Żurada, and Zbigniew Piotrowski. 2026. "Deep Learning for Image Watermarking: A Comprehensive Review and Analysis of Techniques, Challenges, and Applications" Sensors 26, no. 2: 444. https://doi.org/10.3390/s26020444
APA StyleBistroń, M., Żurada, J. M., & Piotrowski, Z. (2026). Deep Learning for Image Watermarking: A Comprehensive Review and Analysis of Techniques, Challenges, and Applications. Sensors, 26(2), 444. https://doi.org/10.3390/s26020444

