Next Article in Journal
Unpacking Prediction: Contextualized and Interpretable Academic Risk Modeling with XAI for Small Cohorts
Previous Article in Journal
A Novel Multi-Rate Simulation Method of Power System Based on State-Space Branch Cutting and Variable Slope Interpolation Interface Algorithm
Previous Article in Special Issue
A Real-Time Improved YOLOv10 Model for Small and Multi-Scale Ground Target Detection in UAV LiDAR Range Images of Complex Scenes
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Active Defense for Deepfakes Using Watermark-Guided Original Face Recovery

College of Cyber Security, Jinan University, Guangzhou 510632, China
*
Author to whom correspondence should be addressed.
Electronics 2026, 15(3), 625; https://doi.org/10.3390/electronics15030625
Submission received: 14 December 2025 / Revised: 21 January 2026 / Accepted: 27 January 2026 / Published: 2 February 2026
(This article belongs to the Special Issue Image Processing for Intelligent Electronics in Multimedia Systems)

Abstract

At present, active defense strategies based on digital watermarking mainly rely on post-event watermark extraction, which verifies the occurrence of deepfake events by measuring the degree of watermark degradation, or on adversarial watermarks to interfere with image generation. To overcome these limitations, we propose a unified watermarking framework that can restore the original content of images tampered with by deepfakes. This scheme integrates three core components: an encoder for watermark pre-embedding, a decoder for robust watermark extraction, and a face restorer for watermark-guided image restoration. Numerous experiments have shown that this method has achieved good results in terms of extraction accuracy and recovery performance, thereby verifying the effectiveness of this approach.

1. Introduction

With the continuous advancement of deep learning, an increasing number of researchers have become engaged in this field, driving breakthroughs in related technologies across multiple domains, such as remote sensing image scene classification [1], deepfake detection [2,3,4], vulnerability detection [5], etc. Deepfakes are a technology capable of tampering with human faces. After such tampering, facial images may exhibit obvious differences and changes [6,7,8,9,10]. The societal impact of deepfakes is profound, and their potential risks are multifaceted. For instance, malicious actors may manipulate facial images of political figures to disrupt public order or engage in other harmful activities [11,12]. Consequently, developing effective defenses against deepfakes is imperative.
Defense against deepfakes mainly consists of proactive and passive approaches. Passive defense [13,14] focuses on capturing the subtle traces left by facial tampering to achieve security goals. Yu et al. [15] proposed a method that provides auxiliary detection for invisible face forgery through cross-domain comparison. However, passive approaches—due to the lack of a universal model—may not robustly cover images generated by known or unknown deep forgery tools. Lu et al. [16] designed a spatiotemporal model that regards spatial and temporal cues as a whole and synchronously captures forgery traces across frames and regions. Nie et al. [17] proposed a diffusion-based inconsistent pattern learning mechanism, which is specifically designed to uncover subtle contradictions at the temporal level.
However, these passive detection technologies cannot promptly prevent deepfakes from occurring and often can only be verified after the fact. Therefore, proactive detection technologies can play a more effective role in this regard. Proactive deepfake defense is generally achieved by adding perturbations [18,19] to the image or embedding watermarks [20,21].
Watermarks can be further refined into three mainstream types, namely adversarial watermarks, fragile watermarks, and robust watermarks. All three types involve pre-embedded watermarks, which are then used to detect and defend against deepfakes. Among them, adversarial watermarking focuses on interfering with the generation of deepfakes through watermarking, thereby resisting deepfakes at their source. Lv et al. [20] proposed an adversarial watermarking technique, embedding imperceptible perturbations in images, which cause the images to blur and degrade semantic fidelity during the processing of deepfake models, thereby simultaneously enhancing the anti-counterfeiting capabilities of both visual and automatic detection. Fragile watermarking is achieved by deliberately allowing the watermark to be destroyed by a deepfake for post-event comparison and detection. Neekhara et al. [22] designed a watermarking scheme that remains robust to ordinary interference but is vulnerable to deepfakes. Robust watermarks maintain a high extraction rate under any circumstance, making them suitable for post-event tracking. Wu et al. [23] constructed a watermarking scheme that can effectively resist interference in different deepfake scenarios, thereby maintaining the watermark’s robustness.
While the above-mentioned schemes offer some resistance against deepfakes, they generally lack the capability to restore manipulated images. Recent research, however, has begun to address this gap. For example, Yu et al. [24] proposed DFREC, a deepfake identity recovery method that simultaneously reconstructs source and target faces using identity segmentation and dual-path reconstruction, achieving high fidelity and traceability. Similarly, Ai et al. [25] introduced DeepReversion, an end-to-end UNet-based network that learns an inverse mapping from fake to original faces, enabling high-fidelity facial inversion and identity recovery. Shao et al. [26] further advanced the field with SeqFakeFormer, which performs sequential prediction of facial manipulations, allowing both the detection of tampered sequences and high-fidelity restoration of original faces.
It is important to clarify that “restoration” in this context refers primarily to the semantic recovery of facial attributes and overall visual content, rather than pixel-level reconstruction of high-frequency details. Moreover, these approaches typically operate without access to true attribute information, which limits their recovery precision. Critically, they also lack the ability to perform copyright verification. Therefore, a unified framework capable of simultaneous copyright verification, semantic content recovery, and forensic traceability remains urgently needed.
In this paper, we propose a unified watermarking framework that integrates the aforementioned functionalities. Our approach employs an encoder−decoder architecture to imperceptibly embed and reliably extract a robust watermark. A dedicated face restorer is further developed to recover images altered by deepfake manipulations. Crucially, the embedded watermark is designed to maintain high robustness even after deepfake processing. The main contributions of this work are as follows:
1.
A facial image recovery framework leveraging auxiliary watermark information has been developed. The framework restores deepfake images by extracting and utilizing guidance information from watermarks present in the images.
2.
A two-stage training procedure was implemented. First, an encoding and decoding architecture was fine-tuned to enhance the robustness of the watermark against both deepfake manipulations and subsequent image recovery operations. Subsequently, a face restorer network was fine-tuned, enabling it to recover deepfake-corrupted images under the guidance of the embedded watermarks.
3.
A dual watermarking framework is proposed. The first-layer watermark is used to provide the labels required for face recovery and can maintain robustness after various sources of noise interference. The second-layer watermark is used to hide additional information by extending the bit length. This integrated approach enables the watermarking system to support both face recovery and forensic traceability.

2. Related Work

2.1. Deepfakes Based on Attribute Editing

In recent years, the continuous advancement of deep learning techniques has driven the rapid evolution of deepfake technologies. Among them, deepfakes based on facial attribute editing [27,28] have gradually attracted widespread attention. This paradigm typically takes a host image and a target attribute vector as inputs, and outputs a forged image exhibiting noticeable alterations in hair color, facial characteristics, and other identity-related features, thereby achieving identity spoofing. StarGAN [6] is a representative method within this category; it attains high-fidelity generation via adversarial training between a generator and a discriminator. Similar frameworks include AttGAN [8], CafeGAN [7], PA-GAN [29], etc., all of which annotate specific facial attributes and leverage these annotations to guide the synthesis of manipulated images. Inspired by these approaches, our solution employs watermarks as supervisory labels to steer the recovery of deepfake images back to the original host image. As a supplement, here are three representative loss functions in attribute editing:
L c l s = E I , V t a r [ l o g ( D i s ( V t a r | G ( I , V t a r ) ) ) ] + E I , V o r i [ l o g ( D i s ( V o r i | I ) ) ]
L r e c = E I , V t a r , V o r i [ I G ( G ( I , V t a r ) , V o r i ) ]
L a d v = E I [ l o g ( D i s ( I ) ) ] + E I , V t a r [ l o g ( 1 D i s ( G ( I , V t a r ) ) ]
Among them, I is the host image, V t a r is the target domain label, V o r i is the source domain label, G is the deepfake generator, and D i s is the deepfake discriminator.

2.2. Proactive Defense Against Deepfakes Based on Digital Watermarking

With the abuse of deepfake technology, its defense countermeasures have become an urgent research topic that must be addressed. Compared with passive detection, active defense mechanisms have received extensive attention due to their advantages of proactive and real-time performance. Against this backdrop, the active defense paradigm based on digital watermarking, as a proactive solution, has become an important research direction in this field. Wu et al. [23] embedded watermarks in images in advance, achieving real-time detection of deepfakes through fragile watermarks, and then achieving post-event tracking of deepfakes by training robust watermarks. Zhang et al. [21] implemented an architectural solution that can first counter deepfakes and still maintain the robustness of the watermark when the countermeasure fails. This achieves dual protection based on the watermark, ensuring that the watermark can be successfully extracted after a deepfake occurs. Nadimpalli et al. [30] proposed an active deepfake detection technology based on GAN visible watermarks. By adding reconstruction regularization to the loss function, the generated fake images come with detectable watermarks. This further enables active detection of deepfakes based on watermarks. Wang et al. [31] designed a robust scheme. By using chaotic encryption and jointly training the encoder−decoder, they embedded identity semantics as watermarks into images, achieving high-precision active detection and source tracing of deepfake face-changing. Furthermore, they proposed a “LampMark” keypoint robust perception watermark. Through end-to-end embedding in real-person images, deepfakes can be actively detected by restoring consistency between the watermark and the test image [32].
Although the current defense schemes can achieve effective deepfake defense, their limitation lies in the fact that once the image is tampered with, it loses its value because it cannot be recovered. Therefore, this research aims to propose a reliable watermarking mechanism. This mechanism should not only utilize the watermark information to guide the restoration of image content, but also ensure the robustness of the watermark throughout the entire process, thereby ensuring the security of the entire procedure.

3. Security Objective

In the envisioned scenario, given that facial images contain highly sensitive information such as identity, physiological features, and behavioral patterns, attackers can generate realistic deepfake content based on them, thereby triggering risks such as identity forgery, reputation damage, and social engineering attacks. Therefore, it is urgently necessary to build a proactive defense mechanism for facial images to weaken or block the information sources of potential deepfake attacks at the data level.
Therefore, we propose a digital watermarking-centered propagation process protection paradigm: embedding a robust watermark that is imperceptible and carries attribute information in the protected image before transmission, and providing the user with a decoder in conjunction. Users can utilize this decoder to extract the embedded watermark, parse its specific bit segments to obtain image attributes, and thereby achieve restoration of the original content. The overall pipeline is illustrated in Figure 1. Mathematically, we have the following:
W enc ( I h , w ) I w
G ( I w ) I f
W dec ( I f ) w
A ( w ) A e x t
Gen ( I f , A e x t ) I r e s
Among them, W e n c is the watermark encoder, W d e c is the watermark decoder, G e n is the face restorer, I h is the host image, I w is the encoded image after embedding the watermark, w is the embedded watermark, w is the extracted watermark, A is a function corresponding to Equation (12), A e x t is the attribute label recovered from the watermark, I f is the image tampered with by a deepfake, G is the generator used for deepfake generation, and I r e s is the recovered image.

4. Proposed Method

This section elaborates on, as thoroughly as possible, the overall idea, the technical route, and implementation details of the proposed solution.

4.1. Framework Overview

We employ a robust watermarking technique to realize the envisioned protection mechanism. Specifically, we adopt the end-to-end embedding–extraction pipeline proposed in [23] as our backbone framework, given its ability to maintain high watermark fidelity even under facial attribute manipulations. The overall architecture, depicted in Figure 2, comprises three key components: an encoder, a decoder, and a face restorer. The encoder embeds a watermark sequence containing facial attribute labels and copyright information into the original facial image. The decoder then extracts this embedded information from a potentially distorted or manipulated image. Finally, the face restorer reconstructs the original facial appearance using the extracted data.

4.2. Network Architecture

4.2.1. Encoder and Decoder

The encoder works in coordination along two parallel paths to achieve watermark embedding. Specifically, the first path extracts the bottleneck feature representation of the host image with the help of the UNet structure; the second path diffuses the watermark bits to different resolution levels through multi-level diffusion and upsampling modules. In the upsampling path of the UNet, the features extracted by the contraction layer in the first path and the extended watermark features in the second path are fused into the corresponding expanded features. In addition, the SE-block is deeply integrated with the ResBlock in the upsampling stage to efficiently complete the watermark embedding task.
The decoder is composed of a UNet structure, a 1 × 1 convolutional layer, and a downsampling operation in sequence. This decoder has been specially trained to enhance the robustness of the watermark extraction process, ensuring accurate restoration of watermark information even under complex conditions.

4.2.2. Face Restorer

The input to this face restorer is the concatenation tensor of a three-channel image and a conditional vector. First, the initial feature mapping is completed through a 7 × 7 convolution, and then high-dimensional bottleneck features are extracted by downsampling. Multiple residual blocks are stacked to enhance the feature representation. Subsequently, the feature dimensions are restored through transposed convolution upsampling. Finally, the image is output by a 7 × 7 convolution and an activation function. The conditional vector participates in the feature transformation throughout the process after spatial replication.
Please note that while training the face restorer, in order to improve the quality of the images, we also incorporated a discriminator for adversarial joint training. It is a PatchGAN [33].

4.3. Dual Watermarking Mechanism

In this section, we will elaborate on the proposed dual watermarking mechanism.

4.3.1. Dual Watermarking Construction

Dual watermarking w = w 1 , w 2 , . . . , w n is composed of the symbol bit A e x t { 0 , 1 } and the extended bit A c o p y { 0 , 1 } : the first layer A e x t is used for face restoration labels, and the second layer A c o p y is used to hide additional information by extending the bits.
The first layer construction. For the first layer, we set the watermark according to the following formula:
w 1 st i = + 0.1 , if A a t t i = 1 0.1 , if A a t t i = 0
Among them, A a t t i is the i-th facial attribute label, and w 1 st i is the i-th first-layer watermark.
The second layer construction. We dynamically allocate the amplitude based on A c o p y as follows:
Δ A o f f i = + 0.025 , if A c o p y i = 1 0.025 , if A c o p y i = 0
Among them, Δ A o f f i is the offset of the i-th bit, and A c o p y i is the i-th bit of A c o p y .
Then, substitute the watermark offset and the first-layer watermark into the following formula to obtain the final watermark:
w i = sign ( w 1 st i ) ( | w 1 st i | + Δ A o f f )
Here, w 1 st i is the first layer watermark at the i-th bit, and w i is the embedded watermark at the i-th bit. It should be noted that—in order to improve the accuracy of subsequent extraction of the second-layer watermark—we will repeat the embedding of each bit five times during the embedding process, that is, a total of six watermark bits will be embedded.

4.3.2. Dual Watermarking Extraction

The first layer extraction. Specifically, after the user extracts the watermark, since the watermark is inevitably subject to slight interference during the embedding and extraction process, in order to improve the restoration accuracy of the image, the user binarizes the received watermark signal: each soft bit estimation undergoes a hard decision with a threshold of 0.5, thereby obtaining a noise-free attribute vector label.
A e x t i = 1 , if w i 0 0 , otherwise
Among them, w i is the i-th bit for extracting the watermark, and A e x t i is the i-th bit for restoring the facial attributes. Here, we adopt the indices i [ 0 , 4 ] to denote the extracted facial-attribute label as A e x t .
Subsequently, for the tampered image, the user sends it along with the attribute label A e x t into the face restorer. The face restorer then outputs the restored image based on this, making it visually approach the original image as closely as possible to achieve the security goal.
The second layer extraction. The extraction of the second-layer watermark first calculates the value of each bit of A c o e x according to the following formula:
A c o e x i = 1 , if | w i |   > 0.1 0 , otherwise
Among them, A c o e x i is the i-th bit of the second-layer watermark for extraction and determination.
Then, since we mentioned in the previous section that each digit is actually repeated five times in the embedding, a majority vote is used to determine the final result, as shown below:
A r e s u l t ( g ) = 1 , if j = 0 4 A v o t e ( 5 g + j ) 3 , g = 0 , 1 , , 5 . 0 , otherwise
Among them, A v o t e represents every five elements, and A r e s u l t is the voting result, that is, the element with more than half of the votes selected from every five is taken as the final result.

4.4. Loss Function and Training Procedure

In this section, we will describe the training measures and loss functions. Besides extracting the embedded watermark bits, the authenticity of facial images altered by deepfake manipulations must also be restored. We draw on the training architecture in [6]. The generator, jointly optimized with an adversarial reconstruction loss and an attribute-classification loss, is capable of reconstructing the original facial appearance when conditioned on the true attribute vector recovered via the watermark. We also leverage its reconstruction loss to aid in image restoration.
Our training is divided into two stages. First, we fine-tune the encoding and decoding parts to ensure that the watermark maintains high robustness in specific environments. Then, we fine-tune the face restorer network to recover the original image and improve the restoration effect in terms of image quality.

4.4.1. Watermark Decoding Stage

We perform fine-tuning based on the framework proposed in [23]. During training, we keep the encoder and discriminator frozen, and introduce a secondary noise simulation step—each encoded image is passed through the same noise perturbation twice consecutively. This strategy enhances the robustness of our fine-tuned model against various deepfake manipulations. The corresponding loss function is defined as follows:
I n o i s e = N ( N ( I w ) )
L d e = w W dec ( I n o i s e )
Here, N represents various types of noise, including identity transformation, common signal processing (JpegTest with Q = 50 , Resize with p = 50 % , GaussianBlur with k = 3 , σ = 2 , MedianBlur with k = 3 , Brightness with f = 0.5 , Contrast with f = 0.5 , Saturation with f = 0.5 , Hue with f = 0.1 , and GaussianNoise with σ = 0.1 ), adversarial attack (SPSA [34]), and generative attacks (StarGAN, AttGAN, and CafeGAN). Among these, SPSA is a gradient-free adversarial attack strategy configured with the following parameters: steps = 20, learning rate = 0.2/255, and standard deviation = 0.005 . To ensure that the embedded watermark information resists a series of such perturbations, we apply the noise twice to the image, as shown in Equation (15), aiming to achieve enhanced robustness.
Therefore, the total loss at this stage is as follows:
L S 1 = L d e
The pseudocode for the training of watermark decoding is detailed in Algorithm 1.
Algorithm 1: Training of watermark decoding
     Input: 
Trained W enc ( ) , host images I h , watermark w , and epoch number i n .
     Output: 
Trained W dec ( ) .
       1:
for  i = 0  to  i n  do
       2:
    Randomly select a host image I h .
       3:
    Generate the watermarked image I w using w .
       4:
    Generate a noised image I n o i s e using N .
       5:
    Compute L d e using w and I n o i s e .
       6:
    Optimize W dec ( ) by minimizing L d e .
       7:
end for

4.4.2. Image Recovery Stage

Based on the successful optimization of the watermark extraction task, this paper further focuses on improving image restoration performance. Specifically, we take the Network [6] as the backbone and migrate it to the image restoration scenario through fine-tuning strategies. During the training procedure, we observed that the reconstruction loss introduced by [6] has a significant benefit for maintaining the consistency of details with the structure. Based on this, we explicitly introduce this reconstruction term as an auxiliary supervision signal in the original objective function to strengthen the optimization constraints of the image restoration branch, thereby systematically improving the visual quality of the recovered image.
Inspired by [24], we further introduce the following loss function to assist in restoring images to be closer to their original state during training. This loss term relies on a VGG19 model pre-trained on ImageNet [35] and operates by quantifying the feature-level differences between the original face image and the restored face image:
L p c e = n = 1 5 b n ( I h ) b n ( I rec ) 2
Here, b n denotes the feature map from the n-th layer of VGG; the first five layers used are relu1_2, relu2_2, relu3_4, relu4_4, and relu5_4 as in [35]. I rec represents the restored image.
Therefore, the complete loss function is defined as follows:
L S 2 = λ 1 L c l s + λ 2 L r e c + λ 3 L a d v + λ 4 L p c e
where L c l s is the classification loss from Equation  (1), L r e c is the reconstruction loss from Equation (2), L a d v is the adversarial loss from Equation (3), and L p c e is the newly introduced auxiliary loss. The coefficients λ 1 , λ 2 , λ 3 , and λ 4 are hyperparameters tuned to ensure high-quality image recovery.
The training pseudocode for this stage is shown in Algorithm 2.
Algorithm 2: Training of image restoration
    Input: 
Host images I h , watermark w , target domain
         
vector V t a r , source domain vector V o r i , and epoch number i n .
    Output: 
Trained Gen ( ) .
      1:
for  i = 0  to  i n  do
      2:
    Randomly select a host image I h .
      3:
    Generate a watermarked image I w using the host image I h and watermark w .
      4:
    Generate a deepfake image I f using the target domain vector V t a r and the watermarked image I w .
      5:
    Generate a recovered image I r e s using V o r i and the deepfake image I f .
      6:
    Compute L S 2 using V o r i , V t a r , and I h .
      7:
    Optimize Gen ( ) by minimizing L S 2 .
      8:
end for

5. Experimental Results

5.1. Experimental Setup

5.1.1. Datasets

We conducted our experiments using three publicly available face image datasets: CelebA-HQ [36] (30,000 images), FFHQ [37] (70,000 images), and HumanFace [38] (1000 images). All images, originally provided at 1024 × 1024 resolution, were uniformly resized to 128 × 128 pixels to meet computational requirements.

5.1.2. Baseline Methods

Our method was compared against several state-of-the-art watermarking schemes [23,39,40,41]. For fair comparison, we used the official code and pre-trained models released by the respective authors, while maintaining consistent experimental conditions, including image size, watermark length, and noise layer configurations. To evaluate robustness against deepfake attacks, we employed three attribute-manipulation models: StarGAN, CafeGAN, and AttGAN [6,7,8]. Our experiments specifically focused on deepfake-induced tampering involving hair color alteration and gender modification.

5.1.3. Implementation Details

The watermark length for all the compared schemes was fixed at 30 bits, consistent with the design in [23], to ensure fair comparison. For deepfake generation, we employed officially released pre-trained models for all methods except CafeGAN. Since no pre-trained CafeGAN model was available, we retrained it using the original paper’s publicly released code and configurations.
Our experiments utilized a total of 30,000 images. For watermark extraction training, 24,195 images were used for training and 3003 for validation. A separate set of 200 images was randomly selected for testing. The model was fine-tuned for three epochs with a batch size of 16, using the Adam optimizer ( l r = 0.0002 , β = 0.5 ).
For the final evaluation, the dataset was split into 28,000 training images and 2000 test images. To rigorously assess robustness, the test images were processed through three different deepfake algorithms to simulate challenging real-world conditions. A final test set of 2000 images was then constructed from these processed outputs. The image restoration model was fine-tuned for 1000 iterations with a batch size of 32, using the Adam optimizer ( α = 1 × 10 5 , β 1 = 0.5 , β 2 = 0.999 ). The learning rate decayed every 1000 iterations. The loss function comprised a weighted combination of multiple terms, with coefficients set as λ 1 = 1 , λ 2 = 10 , λ 3 = 1 , and λ 4 = 0.5 .
All implementations were developed in PyTorch 2025.2 and executed on an NVIDIA RTX 3090 GPU.

5.1.4. Evaluation Metrics

A comprehensive set of metrics was employed to evaluate the experimental results. Watermark robustness was assessed using the average bit error rate (BER). For ablation studies and image quality assessment, we utilized the peak signal-to-noise ratio (PSNR), structural similarity index measure (SSIM), and learned perceptual image patch similarity (LPIPS).

5.2. Watermark Extraction Accuracy Evaluation

We first evaluate the accuracy of watermark extraction in the absence of distortions. Our framework employs a dual watermarking scheme. The first layer serves as a robust watermark, while the second layer functions as an extended robust watermark. Initially, we individually evaluated the robustness of each watermark layer across different datasets under no-interference conditions, as summarized in Table 1.
As shown in the results, both the first-layer and second-layer watermarks maintain strong robustness. This demonstrates that in the absence of attacks, our scheme can reliably verify facial attributes and image copyright.

5.3. Recovery Evaluation

This section presents the restoration results of images subjected to deepfake interference. The receiver extracts the embedded watermark bits and submits them together with the corrupted images to a face restorer model. The model then returns the recovered images.
We investigated two typical tampering scenarios, namely, color alteration and gender modification, to validate the effectiveness of the proposed method. Specifically, we altered the hair color in a facial image from black to blonde and reversed the gender of the subject. Subsequently, the tampered face image, along with its embedded watermark tags, was jointly fed into the constructed generator for recovery. Representative results are illustrated in Figure 3, Figure 4 and Figure 5. These results demonstrate that both the hair color and the gender of the subject were successfully restored, confirming the efficacy of the proposed approach in handling the aforementioned tampering scenarios.

6. Discussion

6.1. Robustness Analysis

This section begins with an evaluation of the robustness of each watermark layer across different datasets under various distortion conditions. The experimental results are summarized in Table 2. These results indicate that the BER scores for the first-layer watermark in our scheme remain close to zero across all tested datasets and distortions, demonstrating its high robustness. Concurrently, the BER for the second-layer watermark remains stable, but slightly higher than that of the first layer. This confirms that the second-layer watermark is suitable for serving as an extension for concealing additional information.
Subsequently, we evaluate the first-layer watermark under different deepfake reconstructions across multiple datasets. Its BER values under various distortions are listed in Table 3, Table 4, Table 5, Table 6, Table 7 and Table 8. The experimental results demonstrate that the proposed scheme exhibits exceptional robustness: in most scenarios, its BER remains below 10 % , significantly outperforming the comparative schemes.
From the perspectives of dataset adaptability and scene adaptability, the proposed scheme maintains a low watermark extraction BER on three mainstream facial datasets across various experimental scenarios involving specific deepfake perturbation-recovery followed by common noise attacks. In StarGAN-related perturbation scenarios—whether after hair color alteration, gender modification, or subsequent common noise attacks—the proposed scheme sustains a low BER on each dataset (primarily between 0.4% and 7%). In corresponding scenarios for AttGAN and CafeGAN involving hair color or gender perturbation-recovery followed by common noise, the BER performance of our scheme remains stable, generally within the range of 1.7% to 14.8%. Even when compared against the best-performing benchmark, our scheme maintains a clear advantage in most settings (with an average BER reduction ranging from approximately 0.02% to 2.3%), fully demonstrating its strong dataset generalization capability and scene adaptability. Remarkably, even after undergoing specific deepfake perturbation and recovery processes, the scheme can effectively withstand subsequent common noise attacks, unaffected by differences in facial data distribution or specific perturbation types.
For scenarios involving specific deepfake perturbations followed by strong interference, such as Gaussian noise or other common noises, the proposed solution still maintains excellent watermark extraction performance. Across all experimental settings, even after images undergo color or transformation perturbations by StarGAN, AttGAN, or CafeGAN followed by restoration and subsequent strong interference (e.g., Gaussian noise, SPSA), the BER of our solution remains significantly lower than that of all benchmark schemes. In contrast, the BER of the comparative schemes typically increases substantially in such complex scenarios, sometimes exceeding 50%. This further verifies that the proposed solution can effectively recover the watermark after specific deepfake perturbations and restoration, exhibiting outstanding robustness.
In summary, the proposed scheme exhibits strong robustness, maintaining stable performance even against subsequent common noise attacks following facial attribute manipulation. This confirms its capability to provide reliable traceability for facial image copyright protection.

6.2. Recovered Images Quality Analysis

This section compares our proposed approach with the image restoration solution GFP-GAN [42]. We have separately evaluated the image restoration quality after being processed by different deepfakes under each dataset, as shown in Table 9.
The results show that although the recovery performance for some specific distortion types is slightly better or worse than that of GFP-GAN, overall, the method proposed in this paper demonstrates superior image recovery performance over GFP-GAN in the vast majority of deepfake distortion types and across all tested datasets. This fully validates that the proposed method has stronger recovery capabilities and generalization in deepfake image restoration.

6.3. Complexity Analysis

As our networks are largely built upon convolutional layers, we provide a complexity analysis of the adopted model in this section. The computational complexity can be derived from the following formula [43]:
O ( N e t ) = i = 1 d l i 1 × t i 2 × g i × p i 2
Here, d represents the depth, and l i 1 , t i 2 , g i , and p i 2 are the number of input channels of the i-th layer, the size of the filters, the number of filters, and the size of the output feature map, respectively.
Thus, the complexity of the encoder in our method is 2.84 × 10 9 , the complexity of the decoder is 1.88 × 10 9 , the complexity of the face restorer is 5.66 × 10 9 , and the complexity of the discriminator is 3.44 × 10 10 .

6.4. Ablation Study

All experimental images and model configurations in this section remain consistent with those described in Section 5.1. For testing, 200 images of size 128 × 128 were selected from the dataset.

6.4.1. Parameter Sensitivity

When embedding the watermark into the image using the encoder, we also introduce a strength factor to control the embedding intensity. This design balances the visibility and robustness of the watermark, ensuring it can be reliably extracted in various application scenarios without significantly degrading the visual quality of the host image, as written as follows:
I w = I h + γ · ( I w I h )
where γ is the strength factor.
To systematically evaluate its impact, we performed a sensitivity analysis on the strength factor parameter. This parameter governs the watermark embedding intensity, thereby directly influencing both the robustness of the watermark and the visual fidelity of the host image. We conducted comparative experiments with three distinct values of the strength factor: 0.5, 1, and 2, as detailed in Table 10 and Table 11.
The experimental results indicate that when the strength factor is set to 0.5, the image quality improves—as evidenced by higher PSNR and SSIM values—at the cost of a reduced watermark extraction rate, which suggests compromised watermark robustness under this setting. In contrast, increasing the strength factor to 2 significantly improves the extraction rate but degrades image quality, indicating that an excessively high embedding intensity harms visual fidelity. Setting the strength factor to 1 achieves an optimal balance, ensuring both reliable watermark extraction and satisfactory visual preservation.

6.4.2. Effectiveness of Watermark on Image Recovery

This section examines the necessity of using watermarks to guide image recovery. In practice, relying on image labels for restoration faces significant limitations. Under the worst-case scenario, users may possess no prior knowledge of the label’s structure, making correct recovery impossible. Even when users are aware of basic formatting constraints, such as the number of digits and the valid value range, randomly generated labels still rarely match the true content. Providing an incorrect label causes the face recovery model to fail, as demonstrated in Figure 6. Therefore, guiding the recovery process through extractable watermarks presents a reliable and necessary alternative. All erroneous labels in this section are assigned a value of 1.

6.4.3. Training Procedure

We first clarify the necessity of retraining the decoder. The proposed framework builds upon SepMark [23] but introduces a decoder retraining stage. To validate the importance of this stage, we quantitatively compare the performance of our watermarking scheme with that of SepMark. As shown in the first two rows of the experimental results in Table 3, Table 4, Table 5, Table 6, Table 7 and Table 8, our scheme achieves a lower bit error rate (BER) than SepMark in the StarGAN restoration scenario. Moreover, in both AttGAN and CafeGAN restoration scenarios, the proposed scheme consistently demonstrates stronger robustness against various distortions introduced by deepfake manipulations. These results confirm that the retraining stage is essential, as it significantly enhances watermark robustness, ensuring reliable watermark preservation after image restoration operations.
Then, we assess the necessity of the specialized training for the face restorer. As shown in Table 12 (the three values in parentheses denote PSNR, SSIM, and LPIPS, respectively), the images restored by our fine-tuned face restorer achieve significantly better visual quality than those generated by the benchmark method [6]. This performance gain confirms the effectiveness and necessity of the dedicated optimization employed in our training.

7. Conclusions

In this paper, we propose an active defense scheme based on digital watermarking to recover images compromised by deepfake attacks. The scheme employs a trained encoder−decoder model that effectively embeds watermark information into images and enables stable extraction even after various types of interference, including deepfake manipulations. Furthermore, an image recovery model is introduced, which utilizes the extracted watermark information to effectively recover tampered images.
Experimental results demonstrate the high robustness of the proposed watermarking mechanism, confirming the overall feasibility and effectiveness of the defense scheme. In future work, we will focus on enhancing the quality of the recovered images through approaches such as optimizing the loss function and refining the model architecture, thereby improving the scheme’s practicality and visual performance.

Author Contributions

Conceptualization, B.F.; methodology, Y.G. and B.F.; software, Y.G.; validation, Y.G.; formal analysis, Y.G., Z.L., Y.L. and B.F.; investigation, Y.G., Z.L., Y.L. and B.F.; resources, B.F.; data curation, Y.G., Z.L. and B.F.; writing—original draft preparation, Y.G. and B.F.; writing—review and editing, Y.G. and B.F.; visualization, Y.G. and B.F.; supervision, B.F.; project administration, B.F.; funding acquisition, B.F. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China (Grant No. 62472199, 62261160653, 62441237), the Natural Science Foundation of Guangdong Province, China (Grant No. 2025A1515011601), and the Opening Project of the MoE Key Laboratory of Information Technology (Sun Yat-sen University) (Grant No. 2024ZD001).

Data Availability Statement

The original data presented in this study are all included in the article. If you have any questions, please consult the author for details.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Huo, Y.; Gang, S.; Guan, C. FCIHMRT: Feature cross-layer interaction hybrid method based on Res2Net and transformer for remote sensing scene classification. Electronics 2023, 12, 4362. [Google Scholar] [CrossRef]
  2. Hashmi, A.; Shahzad, S.A.; Lin, C.W.; Tsao, Y.; Wang, H.M. AVTENet: A human-cognition-inspired audio-visual transformer-based ensemble network for video deepfake detection. IEEE Trans. Cogn. Dev. Syst. 2025, in press. [Google Scholar]
  3. Katı, B.E.; Küçüksille, E.U.; Sarıman, G. Enhancing deepfake detection through quantum transfer learning and class-attention vision transformer architecture. Appl. Sci. 2025, 15, 525. [Google Scholar] [CrossRef]
  4. Deressa, D.W.; Mareen, H.; Lambert, P.; Atnafu, S.; Akhtar, Z.; Van Wallendael, G. GenConViT: Deepfake video detection using generative convolutional vision transformer. Appl. Sci. 2025, 15, 6622. [Google Scholar] [CrossRef]
  5. Zhang, X.; Zhang, F.; Zhao, B.; Zhou, B.; Xiao, B. VulD-Transformer: Source code vulnerability detection via transformer. In Proceedings of the 14th Asia-Pacific Symposium on Internetware; Association for Computing Machinery: New York, NY, USA, 2023; pp. 185–193. [Google Scholar]
  6. Choi, Y.; Choi, M.; Kim, M.; Ha, J.-W.; Kim, S.; Choo, J. StarGAN: Unified generative adversarial networks for multi-domain image-to-image translation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–22 June 2018; pp. 8789–8797. [Google Scholar]
  7. Kwak, J.; Han, D.-K.; Ko, H. CAFE-GAN: Arbitrary face attribute editing with complementary attention feature. In European Conference on Computer Vision; Springer International Publishing: Cham, Switzerland, 2020; pp. 524–540. [Google Scholar]
  8. He, Z.; Zuo, W.; Kan, M.; Shan, S.; Chen, X. AttGAN: Facial attribute editing by only changing what you want. IEEE Trans. Image Process. 2019, 28, 5464–5478. [Google Scholar] [CrossRef] [PubMed]
  9. Pumarola, A.; Agudo, A.; Martinez, A.M.; Sanfeliu, A.; Moreno-Noguer, F. GANimation: Anatomically-aware facial animation from a single image. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 818–833. [Google Scholar]
  10. Chen, R.; Chen, X.; Ni, B.; Ge, Y.; Zhang, J. SimSwap: An efficient framework for high fidelity face swapping. In Proceedings of the 28th ACM International Conference on Multimedia (ACM MM); Association for Computing Machinery: New York, NY, USA, 2020; pp. 2003–2011. [Google Scholar]
  11. Huang, Q.; Zhang, J.; Zhou, W.; Liu, J.; Lu, J. Initiative defense against facial manipulation. Proc. AAAI Conf. Artif. Intell. 2021, 35, 1619–1627. [Google Scholar] [CrossRef]
  12. Wang, X.; Huang, J.; Ma, S.; Liu, Y.; Wang, S. Deepfake disrupter: The detector of deepfake is my friend. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18–24 June 2022; pp. 14920–14929. [Google Scholar]
  13. Shi, Z.; Liu, W.; Chen, H. Face reconstruction-based generalized deepfake detection model with residual outlook attention. ACM Trans. Multimed. Comput. Commun. Appl. 2025, 21, 1–19. [Google Scholar] [CrossRef]
  14. Cao, J.; Ma, C.; Yao, T.; Chen, X.; Ding, S. End-to-end reconstruction-classification learning for face forgery detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18–24 June 2022; pp. 4113–4122. [Google Scholar]
  15. Yu, Y.; Ni, R.; Yang, S.; Zhao, Y.; Wang, L. Narrowing domain gaps with bridging samples for generalized face forgery detection. IEEE Trans. Multimed. 2023, 26, 3405–3417. [Google Scholar] [CrossRef]
  16. Lu, W.; Liu, L.; Zhang, B.; Liu, J.; Tao, D. Detection of deepfake videos using long-distance attention. IEEE Trans. Neural Netw. Learn. Syst. 2023, 35, 9366–9379. [Google Scholar] [CrossRef] [PubMed]
  17. Nie, F.; Ni, J.; Zhang, J.; Wang, X.; Liu, H. DIP: Diffusion learning of inconsistency pattern for general deepfake detection. IEEE Trans. Multimed. 2024, 26, 1–14. [Google Scholar] [CrossRef]
  18. Yang, C.; Ding, L.; Chen, Y.; Zhang, H.; Xiang, T. Defending against GAN-based deepfake attacks via transformation-aware adversarial faces. In 2021 International Joint Conference on Neural Networks (IJCNN); IEEE: Piscataway, NJ, USA, 2021; pp. 1–8. [Google Scholar]
  19. Qu, Z.; Xi, Z.; Lu, W.; Liu, L.; Tao, D. DF-RAP: A robust adversarial perturbation for defending against deepfakes in real-world social network scenarios. IEEE Trans. Inf. Forensics Secur. 2024, 19, 3943–3957. [Google Scholar] [CrossRef]
  20. Lv, L. Smart watermark to defend against deepfake image manipulation. In 2021 IEEE 6th International Conference on Computer and Communication Systems (ICCCS); IEEE: Piscataway, NJ, USA, 2021; pp. 380–384. [Google Scholar]
  21. Zhang, Y.; Ye, D.; Xie, C.; Zhang, Y.; Wang, Y. Dual defense: Adversarial, traceable, and invisible robust watermarking against face swapping. IEEE Trans. Inf. Forensics Secur. 2024, 19, 4628–4641. [Google Scholar] [CrossRef]
  22. Neekhara, P.; Hussain, S.; Zhang, X.; Jia, J.; Huang, J. FaceSigns: Semi-fragile watermarks for media authentication. ACM Trans. Multimed. Comput. Commun. Appl. 2024, 20, 1–21. [Google Scholar] [CrossRef]
  23. Wu, X.; Liao, X.; Ou, B. Sepmark: Deep separable watermarking for unified source tracing and deepfake detection. In Proceedings of the 31st ACM International Conference on Multimedia; Association for Computing Machinery: New York, NY, USA, 2023; pp. 1190–1201. [Google Scholar]
  24. Yu, P.; Gao, H.; Fei, J.; Zhang, Y.; Liu, J.; Wang, S. DFREC: DeepFake Identity Recovery Based on Identity-aware Masked Autoencoder. arXiv 2024, arXiv:2412.07260. [Google Scholar]
  25. Ai, J.; Wang, Z.; Huang, B.; Liu, Y.; Zhang, X. DeepReversion: Reversely inferring the original face from the deepfake face. In 2023 International Joint Conference on Neural Networks (IJCNN); IEEE: Piscataway, NJ, USA, 2023; pp. 1–7. [Google Scholar]
  26. Shao, R.; Wu, T.; Liu, Z. Detecting and recovering sequential deepfake manipulation. In European Conference on Computer Vision (ECCV); Springer Nature Switzerland: Cham, Switzerland, 2022; pp. 712–728. [Google Scholar]
  27. Wu, P.W.; Lin, Y.J.; Chang, C.H.; Lee, H.Y. RelGAN: Multi-domain image-to-image translation via relative attributes. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 16–20 June 2019; pp. 5914–5922. [Google Scholar]
  28. Li, X.; Zhang, S.; Hu, J.; Yang, J.; Ni, Z. Image-to-image translation via hierarchical style disentanglement. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Virtual, 19–25 June 2021; pp. 8639–8648. [Google Scholar]
  29. He, Z.; Kan, M.; Zhang, J.; Shan, S.; Chen, X. PA-GAN: Progressive attention generative adversarial network for facial attribute editing. arXiv 2020, arXiv:2007.05892. [Google Scholar] [CrossRef]
  30. Nadimpalli, A.V.; Rattani, A. Proactive deepfake detection using GAN-based visible watermarking. ACM Trans. Multimed. Comput. Commun. Appl. 2024, 20, 1–27. [Google Scholar] [CrossRef]
  31. Wang, T.; Huang, M.; Cheng, H.; Ma, B.; Wang, Y. Robust identity perceptual watermark against deepfake face swapping. arXiv 2023, arXiv:2311.01357. [Google Scholar]
  32. Wang, T.; Huang, M.; Cheng, H.; Zhang, X.; Shen, Z. Lampmark: Proactive deepfake detection via training-free landmark perceptual watermarks. In Proceedings of the 32nd ACM International Conference on Multimedia; Association for Computing Machinery: New York, NY, USA, 2024; pp. 10515–10524. [Google Scholar]
  33. Isola, P.; Zhu, J.Y.; Zhou, T.; Efros, A.A. Image-to-image translation with conditional adversarial networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 1125–1134. [Google Scholar]
  34. Uesato, J.; O’donoghue, B.; Kohli, P.; Oord, A. Adversarial risk and the dangers of evaluating against weak attacks. In Proceedings of the 35th International Conference on Machine Learning (ICML); PMLR: Cambridge, MA, USA, 2018; pp. 5025–5034. [Google Scholar]
  35. Deng, J.; Dong, W.; Socher, R.; Li, L.-J.; Li, K.; Fei-Fei, L. ImageNet: A large-scale hierarchical image database. In 2009 IEEE Conference on Computer Vision and Pattern Recognition; IEEE: Piscataway, NJ, USA, 2009; pp. 248–255. [Google Scholar]
  36. Karras, T.; Aila, T.; Laine, S.; Lehtinen, J. Progressive growing of GANs for improved quality, stability, and variation. arXiv 2017, arXiv:1710.10196. [Google Scholar]
  37. Karras, T.; Laine, S.; Aila, T. A style-based generator architecture for generative adversarial networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 16–20 June 2019; pp. 4401–4410. [Google Scholar]
  38. JoinDatawithme. Humanface of variOus Age Groups. Available online: https://github.com/JoinDatawithme/Humanface_of_various_age_groups (accessed on 13 January 2026).
  39. Jia, Z.; Fang, H.; Zhang, W. MBRS: Enhancing robustness of DNN-based watermarking by mini-batch of real and simulated JPEG compression. In Proceedings of the 29th ACM International Conference on Multimedia; Association for Computing Machinery: New York, NY, USA, 2021; pp. 41–49. [Google Scholar]
  40. Ma, R.; Guo, M.; Hou, Y.; Zhang, C.; Liu, Y. Towards blind watermarking: Combining invertible and non-invertible mechanisms. In Proceedings of the 30th ACM International Conference on Multimedia; Association for Computing Machinery: New York, NY, USA, 2022; pp. 1532–1542. [Google Scholar]
  41. Fang, H.; Jia, Z.; Ma, Z.; Wang, X.; Zhang, W. PIMoG: An effective screen-shooting noise-layer simulation for deep-learning-based watermarking network. In Proceedings of the 30th ACM International Conference on Multimedia; Association for Computing Machinery: New York, NY, USA, 2022; pp. 2267–2275. [Google Scholar]
  42. Wang, X.; Li, Y.; Zhang, H.; Shan, Y. Towards real-world blind face restoration with generative facial prior. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Virtual, 19–25 June 2021; pp. 9168–9178. [Google Scholar]
  43. Shah, B.; Bhavsar, H. Time complexity in deep learning models. Procedia Comput. Sci. 2022, 215, 202–210. [Google Scholar] [CrossRef]
Figure 1. The scenario we envisioned. We can embed watermarks in the images to be disseminated in advance. Then, after the user receives the images, they can recover the images based on the extracted watermarks to resist deepfake attacks.
Figure 1. The scenario we envisioned. We can embed watermarks in the images to be disseminated in advance. Then, after the user receives the images, they can recover the images based on the extracted watermarks to resist deepfake attacks.
Electronics 15 00625 g001
Figure 2. The framework of the proposed scheme. In the embedding phase, a watermark containing the original attributes is embedded into the host image using Encoder . In the extraction and recovery phase, the watermark is extracted using Decoder , and the original face is recovered using Face   restorer .
Figure 2. The framework of the proposed scheme. In the embedding phase, a watermark containing the original attributes is embedded into the host image using Encoder . In the extraction and recovery phase, the watermark is extracted using Decoder , and the original face is recovered using Face   restorer .
Electronics 15 00625 g002
Figure 3. The image recovery effect of CelebAHQ. The first row is the host image, the second row is the watermark image, the third row is the deepfake image, and the fourth row is the recovered image. (af) are StarGAN (blond), AttGAN (blond), CafeGAN (blond), StarGAN (male), AttGAN (male), and CafeGAN (male).
Figure 3. The image recovery effect of CelebAHQ. The first row is the host image, the second row is the watermark image, the third row is the deepfake image, and the fourth row is the recovered image. (af) are StarGAN (blond), AttGAN (blond), CafeGAN (blond), StarGAN (male), AttGAN (male), and CafeGAN (male).
Electronics 15 00625 g003
Figure 4. The image recovery effect of FFHQ. The first row is the host image, the second row is the watermark image, the third row is the deepfake image, and the fourth row is the recovered image. (af) are StarGAN (blond), AttGAN (blond), CafeGAN (blond), StarGAN (male), AttGAN (male), and CafeGAN (male).
Figure 4. The image recovery effect of FFHQ. The first row is the host image, the second row is the watermark image, the third row is the deepfake image, and the fourth row is the recovered image. (af) are StarGAN (blond), AttGAN (blond), CafeGAN (blond), StarGAN (male), AttGAN (male), and CafeGAN (male).
Electronics 15 00625 g004
Figure 5. The image recovery effect of HumanFace. The first row is the host image, the second row is the watermark image, the third row is the deepfake image, and the fourth row is the recovered image. (af) are StarGAN (blond), AttGAN (blond), CafeGAN (blond), StarGAN (male), AttGAN (male), and CafeGAN (male).
Figure 5. The image recovery effect of HumanFace. The first row is the host image, the second row is the watermark image, the third row is the deepfake image, and the fourth row is the recovered image. (af) are StarGAN (blond), AttGAN (blond), CafeGAN (blond), StarGAN (male), AttGAN (male), and CafeGAN (male).
Electronics 15 00625 g005
Figure 6. Cases where the restoration failed due to incorrect labels. The first row is the host image, the second row is the watermark image, the third row is the deepfake image, and the fourth row is the recovered image. (af) are CelebAHQ (blond), CelebAHQ (male), FFHQ (blond), FFHQ (male), HumanFace (blond), HumanFace (male).
Figure 6. Cases where the restoration failed due to incorrect labels. The first row is the host image, the second row is the watermark image, the third row is the deepfake image, and the fourth row is the recovered image. (af) are CelebAHQ (blond), CelebAHQ (male), FFHQ (blond), FFHQ (male), HumanFace (blond), HumanFace (male).
Electronics 15 00625 g006
Table 1. Watermark extraction accuracy (BER) under distortion-free conditions.
Table 1. Watermark extraction accuracy (BER) under distortion-free conditions.
DatasetWatermark Extraction Accuracy (BER)
First Layer WatermarkSecond Layer Watermark
HumanFace 0.0000 % 9.3333 %
FFHQ 0.0000 % 10.2500 %
CelebAHQ 0.0000 % 3.4167 %
Average 0.0000 % 7.6667 %
Table 2. Watermark extraction accuracy (BER) under various distortions.
Table 2. Watermark extraction accuracy (BER) under various distortions.
Distortion TypeThe First Layer WatermarkThe Second Layer Watermark
HumanFaceFFHQCelebAHQHumanFaceFFHQCelebAHQ
Identity0.0000%0.0000%0.0000%9.3333%10.2500%3.4167%
Jpeg0.1333%0.2333%0.2167%11.5833%15.9167%14.9167%
Resize0.0000%0.0167%0.0000%8.5833%9.7500%6.5000%
GaussianBlur0.0000%0.0000%0.0000%31.6667%28.0833%32.1667%
MedianBlur0.0000%0.0000%0.0000%5.9167%4.0000%4.1667%
Brightness0.0000%0.0000%0.0000%11.3333%13.3333%7.4167%
Contrast0.0000%0.0000%0.0000%11.5000%14.9167%9.7500%
Saturation0.0000%0.0000%0.0000%10.5000%12.8333%5.7500%
Hue0.0000%0.0000%0.0000%13.1667%12.0000%8.2500%
GaussianNoise0.5000%0.6000%0.8167%11.1667%17.0833%17.5833%
SPSA0.4333%1.1833%1.4000%22.0000%24.2500%26.0000%
StarGAN (blond)0.2333%0.4333%0.1667%12.9167%12.9167%12.1667%
StarGAN (male)0.1333%0.2000%0.1500%12.5000%14.3333%14.6667%
AttGAN (blond)2.5667%2.2667%1.1000%22.5000%22.5000%20.4167%
AttGAN (male)1.1833%0.7333%0.5833%22.7500%20.7500%16.3333%
CafeGAN (blond)3.7500%3.3333%3.3667%26.0833%26.0833%23.6667%
CafeGAN (male)0.5333%0.6500%0.5333%17.2500%17.5000%17.8333%
Average0.5569%0.5676%0.4902%15.3382%16.2647%14.1765%
Table 3. Watermark extraction accuracy (BER) on the images processed by StarGAN (blond).
Table 3. Watermark extraction accuracy (BER) on the images processed by StarGAN (blond).
Scheme (Dataset)IdentityJpegTestResizeGaussianBlurMedianBlurBrightnessContrastSaturationHueGaussianNoiseSPSAAverage
HumanFace
Ours (HumanFace)0.5667%2.6500%1.4667%1.0000%1.0333%0.7333%1.2167%0.6000%0.6833%4.8500%13.6833%2.5893%
SepMark (HumanFace)0.8333%3.1333%1.5833%1.0667%1.0833%1.0667%1.3000%0.9000%0.9000%5.6667%12.0333%2.6879%
CIN (HumanFace)34.8500%42.2267%32.8167%53.5000%38.5000%35.6667%36.5167%35.4000%35.6500%45.8000%40.5167%39.2221%
MBRS (HumanFace)28.8333%42.7500%30.0500%37.9833%36.8667%30.6667%30.3667%29.8000%31.0333%46.7333%42.3167%35.2182%
PIMOG (HumanFace)23.9667%42.3000%22.5667%25.1000%24.8000%28.7500%27.4167%25.2500%26.1333%41.9667%44.1000%30.2136%
FFHQ
Ours (FFHQ)0.5000%2.9333%1.1500%0.8333%0.7667%0.6000%0.9500%0.5333%0.4667%4.5833%17.3333%2.7864%
SepMark (FFHQ)0.5833%3.0667%1.0167%0.8333%0.7000%0.7000%0.9000%0.5667%0.6000%6.1500%15.7667%2.8076%
CIN (FFHQ)33.6000%37.6167%47.8667%46.7667%40.8167%34.5667%35.6667%33.8500%34.1667%46.9667%41.9500%39.4394%
MBRS (FFHQ)25.6333%40.9333%27.0167%36.0333%34.6500%27.5500%27.9500%26.3500%28.1667%47.0000%42.2833%33.0515%
PIMOG (FFHQ)21.1500%40.9000%20.6167%22.3167%22.0833%25.9333%25.2000%22.4333%23.0333%41.1833%43.5667%28.0379%
CelebAHQ
Ours (CelebAHQ)0.4833%3.1333%1.2000%0.9333%0.7500%0.7167%1.4333%0.4500%0.7333%6.1833%17.7000%3.0651%
SepMark (CelebAHQ)0.5000%3.3500%1.0000%0.8667%0.8000%0.6167%1.1000%0.4833%0.4833%6.1833%16.9667%2.9409%
CIN (CelebAHQ)33.2333%40.5833%30.3500%52.9833%38.1500%34.5167%34.4833%33.0000%34.5000%33.5333%35.0000%36.3939%
MBRS (CelebAHQ)50.1000%43.3333%29.4333%37.3333%34.8333%29.5000%30.2999%25.6500%27.9000%33.4167%34.3667%34.1970%
PIMOG (CelebAHQ)22.1667%39.9500%21.0667%23.2667%22.7500%26.9333%25.7833%23.5000%23.5333%35.7833%33.8167%27.1409%
Note: The bold values indicate that this solution is the most effective.
Table 4. Watermark extraction accuracy (BER) on the images processed by StarGAN (male).
Table 4. Watermark extraction accuracy (BER) on the images processed by StarGAN (male).
Scheme (Dataset)IdentityJpegTestResizeGaussianBlurMedianBlurBrightnessContrastSaturationHueGaussianNoiseSPSAAverage
HumanFace
Ours (HumanFace)0.4333%2.8833%1.3167%0.8667%0.9667%0.6667%0.9667%0.4333%0.5500%5.0500%12.7333%2.4424%
SepMark (HumanFace)0.4833%2.8000%1.1500%1.1167%1.2000%0.6500%1.0500%0.5333%0.6500%5.4667%12.0167%2.4652%
CIN (HumanFace)35.7500%41.3333%33.5000%53.4833%39.0000%36.0500%37.0500%35.6333%36.4000%45.8833%41.0000%39.5530%
MBRS (HumanFace)29.7500%43.1000%32.4167%38.6333%37.1500%32.0500%30.9833%30.4667%32.0000%47.0167%42.1000%35.9697%
PIMOG (HumanFace)22.9000%41.1833%21.7833%24.3167%23.5833%28.1833%26.2833%24.4167%24.4333%42.0167%44.8833%29.4530%
FFHQ
Ours (FFHQ)0.4167%2.9333%1.0667%0.8167%0.7333%0.4833%0.8167%0.3833%0.4333%5.0667%15.7000%2.6227%
SepMark (FFHQ)0.5000%2.9500%1.2500%0.8333%0.7167%0.6333%0.8000%0.4667%0.5333%5.9667%14.0333%2.6076%
CIN (FFHQ)33.6167%37.6167%47.8667%46.7667%40.8167%34.5500%35.6667%33.8500%34.1500%49.9667%41.9667%39.7121%
MBRS (FFHQ)26.5167%42.7833%28.2667%37.5500%35.1000%29.1000%29.0000%27.4333%28.3300%47.5000%42.6500%34.0209%
PIMOG (FFHQ)21.5833%42.4167%20.4000%23.2667%22.2833%26.3167%24.4000%22.5667%23.4167%41.3667%42.9500%28.2697%
CelebAHQ
Ours (CelebAHQ)0.5333%3.4000%1.2167%0.8333%0.7667%0.7167%1.0500%0.4667%0.5167%6.2500%17.2667%3.0015%
SepMark (CelebAHQ)0.4833%3.4000%1.0833%0.9333%0.9000%0.6667%0.8167%0.5000%0.5167%6.6833%15.8500%2.8939%
CIN (CelebAHQ)41.6833%48.2000%33.1000%55.1667%44.8167%43.0667%42.1500%41.4000%42.1500%44.6500%43.2667%43.6046%
MBRS (CelebAHQ)27.1833%42.7000%29.3333%38.3833%35.7833%29.0833%29.7833%27.3833%29.3667%46.7000%42.2000%34.3545%
PIMOG (CelebAHQ)21.2167%40.6167%20.4167%22.6833%22.2333%25.5833%25.4333%22.5167%23.2667%41.6667%43.6667%28.1182%
Note: The bold values indicate that this solution is the most effective.
Table 5. Watermark extraction accuracy (BER) on the images processed by AttGAN (blond).
Table 5. Watermark extraction accuracy (BER) on the images processed by AttGAN (blond).
Scheme (Dataset)IdentityJpegTestResizeGaussianBlurMedianBlurBrightnessContrastSaturationHueGaussianNoiseSPSAAverage
HumanFace
Ours (HumanFace)4.7833%9.4667%7.0000%5.3667%6.4667%5.5500%5.4333%4.8333%4.7833%11.0833%19.4500%7.6561%
SepMark (HumanFace)6.1667%10.6333%8.2000%7.4333%7.6667%7.0667%6.7667%6.2333%6.3500%13.6500%19.8167%9.0894%
CIN (HumanFace)35.6833%39.3667%48.1333%49.0000%40.5833%37.1167%36.6000%35.7500%35.8500%47.4167%41.9333%40.6758%
MBRS (HumanFace)34.6833%36.7500%38.7000%34.3333%35.0833%36.2833%34.3667%34.7167%35.5000%50.2167%40.2500%37.3530%
PIMOG (HumanFace)30.6000%42.8333%30.2333%30.5500%31.0667%33.0333%31.7333%31.2500%31.4833%43.5167%41.9500%34.3864%
FFHQ
Ours (FFHQ)3.9833%8.5667%6.9000%4.6000%6.0667%4.4667%4.2000%4.0000%4.3333%10.6000%23.7667%7.4076%
SepMark (FFHQ)5.4333%9.0167%7.5333%6.4333%6.7667%5.9000%6.1500%5.3500%5.6333%11.8833%24.3333%8.5848%
CIN (FFHQ)33.6167%37.6167%47.8667%46.7667%40.8167%34.5667%35.6667%33.8500%34.1667%46.9667%41.9667%39.4425%
MBRS (FFHQ)31.6333%35.3833%34.5000%31.2833%32.6000%31.9500%31.5667%31.3167%33.2167%50.1667%40.3167%34.9030%
PIMOG (FFHQ)28.5333%40.9333%28.0500%28.3000%28.7333%30.5667%29.7000%29.1667%23.7833%41.9667%42.8500%32.0530%
CelebAHQ
Ours (CelebAHQ)2.8667%7.9000%5.1500%3.5500%4.3000%3.4667%3.7167%2.9667%2.8167%10.6833%23.4500%6.4424%
SepMark (CelebAHQ)4.0167%8.5333%5.6500%4.9833%5.4500%4.2667%4.9500%4.0000%4.1167%11.6500%24.7833%7.4909%
CIN (CelebAHQ)33.0167%38.0500%48.4667%47.7833%39.1833%33.1167%34.9500%32.7667%33.2000%36.0167%44.6833%38.2939%
MBRS (CelebAHQ)27.3500%34.1167%30.5333%26.5167%28.1167%29.3667%27.5500%27.3500%29.0500%37.5333%42.4167%30.9000%
PIMOG (CelebAHQ)24.2667%39.5333%23.2000%23.8500%23.9333%27.1500%26.2000%24.3667%25.1667%31.3500%39.6500%28.0606%
Note: The bold values indicate that this solution is the most effective.
Table 6. Watermark extraction accuracy (BER) on the images processed by AttGAN (male).
Table 6. Watermark extraction accuracy (BER) on the images processed by AttGAN (male).
Scheme (Dataset)IdentityJpegTestResizeGaussianBlurMedianBlurBrightnessContrastSaturationHueGaussianNoiseSPSAAverage
HumanFace
Ours (HumanFace)2.1000%5.1833%3.6500%2.8000%3.4833%2.5833%3.3333%2.1333%2.2500%7.8167%19.7167%5.0045%
SepMark (HumanFace)2.8167%5.8000%4.5667%3.5667%3.6500%2.9833%3.8000%2.8333%3.0833%9.4500%19.9333%5.6803%
CIN (HumanFace)33.6167%37.6167%47.8667%46.7667%40.7833%34.5500%35.6833%33.8167%34.1167%46.9667%41.9333%39.4288%
MBRS (HumanFace)30.3333%36.6500%34.9667%29.7833%30.9667%32.5667%32.0833%32.1833%32.7500%49.7833%39.3833%34.6773%
PIMOG (HumanFace)26.7833%41.4000%25.9500%27.0000%26.9833%29.3000%28.2000%27.6500%28.4000%41.9500%42.7000%31.4833%
FFHQ
Ours (FFHQ)1.9667%4.3333%3.4667%2.2167%2.7167%2.2833%2.4000%1.9000%2.0000%7.0500%22.0500%4.7621%
SepMark (FFHQ)2.2167%4.7000%3.5667%2.6333%2.9833%2.3333%3.1000%2.2833%2.3833%7.9833%22.1000%5.1167%
CIN (FFHQ)33.6167%37.6333%47.8667%46.7667%40.8167%34.5500%35.6667%33.8500%34.1667%46.9667%41.9667%39.4424%
MBRS (FFHQ)26.8000%33.2167%31.2833%25.9500%28.0000%27.3833%28.4333%26.8167%28.4833%50.1333%40.5833%31.5530%
PIMOG (FFHQ)22.9667%40.0167%22.9500%23.4000%23.0500%26.0333%24.7167%23.6500%26.0333%39.9500%42.1833%28.6318%
CelebAHQ
Ours (CelebAHQ)1.7500%5.4500%3.1333%2.1000%2.4333%1.8167%2.4333%1.7167%1.8833%8.3667%21.9500%4.8212%
SepMark (CelebAHQ)2.0833%5.8167%3.4333%3.1000%3.0833%2.5500%3.0333%2.1333%2.3833%9.7167%21.8667%5.3818%
CIN (CelebAHQ)31.1667%38.4333%47.8167%47.4833%39.0333%32.9500%32.9500%32.9500%31.7500%46.7500%43.0167%38.5727%
MBRS (CelebAHQ)24.0667%33.1000%29.5000%23.5333%24.5667%25.8500%25.8333%24.4833%25.4833%49.0000%38.4667%29.4439%
PIMOG (CelebAHQ)22.1667%40.0333%21.9500%22.8667%22.7000%24.6000%24.9167%22.7333%23.8167%40.4667%38.4667%27.7015%
Note: The bold values indicate that this solution is the most effective.
Table 7. Watermark extraction accuracy (BER) on the images processed by CafeGAN (blond).
Table 7. Watermark extraction accuracy (BER) on the images processed by CafeGAN (blond).
Scheme (Dataset)IdentityJpegTestResizeGaussianBlurMedianBlurBrightnessContrastSaturationHueGaussianNoiseSPSAAverage
HumanFace
Ours (HumanFace)7.1667%11.7500%9.3833%6.0667%9.1833%7.7667%7.5333%7.1167%7.0333%13.8833%23.1000%9.9985%
SepMark (HumanFace)9.3667%12.9333%10.7167%10.9833%11.1833%10.2333%10.1000%9.4000%9.7833%15.5833%25.0167%12.3000%
CIN (HumanFace)42.1167%44.2333%48.1667%48.3333%45.0000%42.9667%42.5833%42.3500%42.4167%48.4833%45.1333%44.7076%
MBRS (HumanFace)37.7000%41.5833%38.0833%38.8167%39.3000%38.4500%38.2667%37.6833%39.4000%38.8167%43.6333%39.2485%
PIMOG (HumanFace)29.6500%42.5333%27.7000%30.5167%29.4667%32.0000%30.3500%29.5667%31.1500%43.8000%43.9167%33.6955%
FFHQ
Ours (FFHQ)7.0833%11.1500%9.2167%6.1667%9.1500%7.7333%7.7667%6.9667%6.9000%13.4000%26.4500%10.1803%
SepMark (FFHQ)9.4167%12.1000%10.9667%10.4667%11.2833%9.7000%9.7000%9.5333%9.5500%16.1167%27.9000%12.4303%
CIN (FFHQ)33.6000%37.6167%47.8667%46.7667%40.8167%34.5500%35.6667%33.8500%34.1667%46.9667%41.9667%39.4394%
MBRS (FFHQ)37.5833%43.3833%38.4833%38.0167%39.3000%38.0333%37.6000%37.4833%39.1833%47.6333%42.8500%40.1966%
PIMOG (FFHQ)26.4833%42.1333%25.2833%27.0167%26.6667%29.5833%27.4333%27.1833%27.5000%41.9500%42.7833%30.1233%
CelebAHQ
Ours (CelebAHQ)6.6000%11.8000%9.5333%5.9167%9.2333%7.2000%7.6500%6.6667%6.5000%14.8000%27.6333%10.3212%
SepMark (CelebAHQ)9.0667%13.1167%10.9833%10.6000%10.7167%9.8670%9.9167%9.3000%9.3167%17.0333%28.2000%12.5561%
CIN (CelebAHQ)41.2000%44.6667%47.5833%47.7333%45.0500%41.9333%41.8000%41.2333%41.6000%42.7833%50.8167%44.2182%
MBRS (CelebAHQ)38.5333%43.5167%37.6000%40.0333%40.0333%39.6833%39.4667%38.6333%39.4167%40.7500%46.6667%40.3939%
PIMOG (CelebAHQ)26.4167%43.0167%25.3667%27.0167%26.8667%29.3833%27.8167%27.1167%28.1667%42.2000%45.8333%31.7455%
Note: The bold values indicate that this solution is the most effective.
Table 8. Watermark extraction accuracy (BER) on the images processed by CafeGAN (male).
Table 8. Watermark extraction accuracy (BER) on the images processed by CafeGAN (male).
Scheme (Dataset)IdentityJpegTestResizeGaussianBlurMedianBlurBrightnessContrastSaturationHueGaussianNoiseSPSAAverage
HumanFace
Ours (HumanFace)1.7667%4.1667%3.3500%1.5833%3.0000%2.0000%2.8500%1.7667%1.8833%6.5000%19.8333%4.4273%
SepMark (HumanFace)2.5333%4.6333%3.5833%3.3667%3.3167%2.8500%6.6667%2.5167%2.5667%8.3833%20.5833%5.5455%
CIN (HumanFace)38.0667%40.9833%47.2500%47.5333%43.3167%38.5167%38.7500%37.9667%38.6333%47.3000%43.3500%41.9697%
MBRS (HumanFace)31.4667%39.5500%33.4833%32.2167%33.4500%33.3167%32.9000%32.3000%34.4167%48.0500%40.6667%35.6197%
PIMOG (HumanFace)17.9333%37.8333%16.0833%18.9500%18.5333%21.7500%20.1167%18.4333%20.2000%38.0000%40.7000%24.4121%
FFHQ
Ours (FFHQ)1.4500%4.1000%3.4333%1.6500%2.8833%1.7500%2.4833%1.6000%1.5833%6.9833%24.2500%4.7424%
SepMark (FFHQ)2.4167%4.5167%3.7333%3.1833%3.1500%2.3000%3.3300%2.2000%2.3500%8.5333%24.4000%5.4648%
CIN (FFHQ)33.6167%37.6167%47.8667%46.7667%40.8167%34.5500%35.6667%33.8500%34.1667%46.7667%41.9500%39.4212%
MBRS (FFHQ)31.1667%40.8000%32.7167%31.5167%32.7500%32.3667%32.3500%31.3333%33.9333%46.7167%42.2500%35.2636%
PIMOG (FFHQ)16.3667%36.8333%15.1000%17.1833%26.6167%18.9833%18.9000%17.6333%18.8000%37.2500%46.0167%24.5167%
CelebAHQ
Ours (CelebAHQ)1.8333%5.6333%3.6333%1.8000%3.0833%2.0333%3.3000%1.7500%2.0000%8.9000%24.8000%5.3424%
SepMark (CelebAHQ)2.5833%5.8167%3.8833%3.7000%3.2333%2.8167%3.5833%2.5333%2.6500%10.1667%25.6500%6.0561%
CIN (CelebAHQ)37.7333%42.0667%47.8667%47.2000%42.9333%38.5833%38.6000%37.5000%38.1000%47.2000%44.9167%42.0636%
MBRS (CelebAHQ)31.1667%40.5500%33.0333%31.5833%31.9167%32.3333%32.3500%31.4167%33.8333%48.6500%40.1000%35.1758%
PIMOG (CelebAHQ)16.9000%38.3667%16.3167%17.8000%17.5500%20.0833%20.1167%17.7833%19.0667%37.1000%40.3167%23.7636%
Note: The bold values indicate that this solution is the most effective.
Table 9. Image quality comparison of recovered images obtained by GFP-GAN and ours.
Table 9. Image quality comparison of recovered images obtained by GFP-GAN and ours.
Distortion TypeHumanFaceFFHQCelebAHQ
GFP-GANOursGFP-GANOursGFP-GANOurs
StarGAN(blond)(16.1569, 0.7429, 0.1083)(27.0001, 0.9112, 0.0466)(15.6818, 0.6878, 0.1388)(27.4159, 0.9022, 0.0489)(15.6297, 0.6841, 0.1461)(28.4617, 0.9159, 0.0380)
StarGAN(male)(22.8294, 0.8543, 0.0531)(27.5393, 0.9141, 0.0444)(23.0230, 0.8336, 0.0693)(27.8128, 0.9056, 0.0464)(23.1393, 0.8332, 0.0722)(28.7715, 0.9161, 0.0359)
AttGAN(blond)(13.6015, 0.6421, 0.1624)(18.3305, 0.7180, 0.1157)(14.7626, 0.6159, 0.1758)(18.7024, 0.6931, 0.1238)(15.1420, 0.6535, 0.1535)(21.8117, 0.7805, 0.0773)
AttGAN(male)(20.2485, 0.7911, 0.0995)(19.4030, 0.7639, 0.1063)(22.4829, 0.8191, 0.0865)(21.2123, 0.7871, 0.0912)(24.7060, 0.8533, 0.0658)(23.6924, 0.6932, 0.0678)
CafeGAN(blond)(15.7986, 0.6755, 0.1407)(19.8994, 0.7397, 0.0987)(13.4619, 0.5979, 0.1833)(19.4606, 0.7028, 0.1121)(15.2037, 0.6258, 0.1724)(20.1800, 0.5367, 0.1116)
CafeGAN(male)(25.9666, 0.8940, 0.0432)(23.9651, 0.8566, 0.0554)(26.1756, 0.8815, 0.0613)(24.4656, 0.8469, 0.0579)(26.7685, 0.8896, 0.0606)(25.5523, 0.7342, 0.0538)
Average(19.0988, 0.7667, 0.1012)(22.6896, 0.8173, 0.0779)(19.2646, 0.7393, 0.1192)(23.1783, 0.8063, 0.0801)(20.0982, 0.7566, 0.1118)(24.7449, 0.7628, 0.0641)
Note: The values in parentheses are the PSNR, SSIM, and LPIPS of the recovered image, respectively, and the bold values indicate that this solution is the most effective.
Table 10. The watermark extraction accuracy (BER) of each strength factor for each dataset under different noise conditions.
Table 10. The watermark extraction accuracy (BER) of each strength factor for each dataset under different noise conditions.
DistortionDataset
HumanFaceFFHQCelebAHQ
γ = 1 γ = 0.5 γ = 2 γ = 1 γ = 0.5 γ = 2 γ = 1 γ = 0.5 γ = 2
Identity0.0000%0.1333%0.0000%0.0000%0.2167%0.0000%0.0000%0.2833%0.0000%
JPEG Test0.1333%5.7667%0.0000%0.2333%6.4000%0.0000%0.2167%6.7333%0.0000%
Resize0.0000%1.5000%0.0000%0.0167%1.4333%0.0000%0.0000%1.2000%0.0000%
GaussianBlur0.0000%2.3167%0.0000%0.0000%2.2667%0.0000%0.0000%1.8333%0.0000%
MedianBlur0.0000%0.5167%0.0000%0.0000%0.5167%0.0000%0.0000%0.3333%0.0000%
Brightness0.0000%0.1667%0.0000%0.0000%0.3167%0.0000%0.0000%0.3833%0.0000%
Contrast0.0000%0.3333%0.0000%0.0000%0.4500%0.0000%0.0000%0.3833%0.0000%
Saturation0.0000%0.2333%0.0000%0.0000%0.2667%0.0000%0.0000%0.2500%0.0000%
Hue0.0000%0.2000%0.0000%0.0000%0.3333%0.0000%0.0000%0.2500%0.0000%
GaussianNoise0.5000%9.3167%0.0000%0.6000%9.9000%0.0000%0.8167%11.5000%0.0000%
SPSA0.4333%4.3333%0.0333%1.1833%6.4500%0.2500%1.4000%6.8667%0.2333%
StarGAN (blond)0.2333%6.8500%0.0167%0.4333%7.1167%0.0000%0.1667%6.1000%0.0333%
StarGAN (male)0.1333%6.3500%0.0000%0.2000%5.8000%0.0000%0.1500%5.2000%0.0000%
AttGAN (blond)2.5667%14.9500%0.1000%2.2667%14.7000%0.1000%1.1000%12.9833%0.0333%
AttGAN (male)1.1833%9.8667%0.0167%0.7333%8.7000%0.0000%0.5833%8.4667%0.0000%
CafeGAN (blond)3.7500%16.8000%0.1333%3.3333%16.3333%0.2000%3.3667%16.6500%0.1667%
CafeGAN (male)0.5333%8.3667%0.0000%0.6500%8.1000%0.0000%0.5333%7.9000%0.0000%
Average0.5569%5.1794%0.0176%0.5676%5.2529%0.0324%0.4902%5.1549%0.0274%
Table 11. Evaluation of the visual quality of watermarked images with different γ values.
Table 11. Evaluation of the visual quality of watermarked images with different γ values.
Dataset γ = 1 γ = 0.5 γ = 2
HumanFace(37.6321, 0.9559, 0.0025)(43.5612, 0.9856, 0.0006)(31.6845, 0.8750, 0.0135)
FFHQ(38.0710, 0.9567, 0.0030)(43.9921, 0.9862, 0.0008)(32.1337, 0.8756, 0.0159)
CelebAHQ(39.0441, 0.9619, 0.0025)(44.9401, 0.9880, 0.0007)(33.1328, 0.8882, 0.0134)
Average(38.2491, 0.9582, 0.0027)(44.1645, 0.9866, 0.0007)(32.3170, 0.8796, 0.0143)
Table 12. Image quality of recovered images obtained by StarGAN and ours.
Table 12. Image quality of recovered images obtained by StarGAN and ours.
Distortion TypeHumanFaceFFHQCelebAHQ
StarGANOursStarGANOursStarGANOurs
StarGAN (blond)(26.9548, 0.7846, 0.0658)(27.0001, 0.9112, 0.0466)(27.5291, 0.7932, 0.0688)(27.4159, 0.9022, 0.0489)(28.6113, 0.9239, 0.0442)(28.4617, 0.9159, 0.0380)
StarGAN (male)(27.1126, 0.7751, 0.0663)(27.5393, 0.9141, 0.0444)(27.4235, 0.7816, 0.0708)(27.8128, 0.9056, 0.0464)(28.4101, 0.8203, 0.0538)(28.7715, 0.9161, 0.0359)
AttGAN (blond)(17.9340, 0.4623, 0.1418)(18.3305, 0.7180, 0.1157)(18.4189, 0.4930, 0.1504)(18.7024, 0.6931, 0.1238)(21.0817, 0.7606, 0.0897)(21.8117, 0.7805, 0.0773)
AttGAN (male)(19.1163, 0.5203, 0.1139)(19.4030, 0.7639, 0.1063)(20.5325, 0.5883, 0.1091)(21.2123, 0.7871, 0.0912)(22.6571, 0.6727, 0.0803)(23.6924, 0.6932, 0.0678)
CafeGAN (blond)(19.2086, 0.4813, 0.1242)(19.8994, 0.7397, 0.0987)(18.9439, 0.4789, 0.1385)(19.4606, 0.7028, 0.1121)(19.6383, 0.7115, 0.1096)(20.1800, 0.5367, 0.1116)
CafeGAN (male)(22.8836, 0.6549, 0.0736)(23.9651, 0.8566, 0.0554)(22.9488, 0.6608, 0.0818)(24.4656, 0.8469, 0.0579)(23.9011, 0.7080, 0.0681)(25.5523, 0.7342, 0.0538)
Average(22.2617, 0.6131, 0.0976)(22.6896, 0.8173, 0.0779)(22.6328, 0.6326, 0.1032)(23.1783, 0.8063, 0.0801)(24.0499, 0.7662, 0.0743)(24.7449, 0.7628, 0.0641)
Note: The values in parentheses are the PSNR, SSIM, and LPIPS of the recovered image, respectively, and the bold values indicate that this solution is the most effective.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Guo, Y.; Liu, Z.; Li, Y.; Feng, B. Active Defense for Deepfakes Using Watermark-Guided Original Face Recovery. Electronics 2026, 15, 625. https://doi.org/10.3390/electronics15030625

AMA Style

Guo Y, Liu Z, Li Y, Feng B. Active Defense for Deepfakes Using Watermark-Guided Original Face Recovery. Electronics. 2026; 15(3):625. https://doi.org/10.3390/electronics15030625

Chicago/Turabian Style

Guo, Yizhi, Ziqiao Liu, Yantao Li, and Bingwen Feng. 2026. "Active Defense for Deepfakes Using Watermark-Guided Original Face Recovery" Electronics 15, no. 3: 625. https://doi.org/10.3390/electronics15030625

APA Style

Guo, Y., Liu, Z., Li, Y., & Feng, B. (2026). Active Defense for Deepfakes Using Watermark-Guided Original Face Recovery. Electronics, 15(3), 625. https://doi.org/10.3390/electronics15030625

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop