Next Article in Journal
Reconstructing Image Composition: Computation of Leading Lines
Next Article in Special Issue
Reliable Out-of-Distribution Recognition of Synthetic Images
Previous Article in Journal
Exocentric and Egocentric Views for Biomedical Data Analytics in Virtual Environments—A Usability Study
Previous Article in Special Issue
A Robust Approach to Multimodal Deepfake Detection
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Harmonizing Image Forgery Detection & Localization: Fusion of Complementary Approaches

Internet Technology and Data Science Lab (IDLab), Ghent University—imec, 9052 Ghent, Belgium
*
Author to whom correspondence should be addressed.
J. Imaging 2024, 10(1), 4; https://doi.org/10.3390/jimaging10010004
Submission received: 27 October 2023 / Revised: 29 November 2023 / Accepted: 19 December 2023 / Published: 25 December 2023
(This article belongs to the Special Issue Robust Deep Learning Techniques for Multimedia Forensics and Security)

Abstract

:
Image manipulation is easier than ever, often facilitated using accessible AI-based tools. This poses significant risks when used to disseminate disinformation, false evidence, or fraud, which highlights the need for image forgery detection and localization methods to combat this issue. While some recent detection methods demonstrate good performance, there is still a significant gap to be closed to consistently and accurately detect image manipulations in the wild. This paper aims to enhance forgery detection and localization by combining existing detection methods that complement each other. First, we analyze these methods’ complementarity, with an objective measurement of complementariness, and calculation of a target performance value using a theoretical oracle fusion. Then, we propose a novel fusion method that combines the existing methods’ outputs. The proposed fusion method is trained using a Generative Adversarial Network architecture. Our experiments demonstrate improved detection and localization performance on a variety of datasets. Although our fusion method is hindered by a lack of generalization, this is a common problem in supervised learning, and hence a motivation for future work. In conclusion, this work deepens our understanding of forgery detection methods’ complementariness and how to harmonize them. As such, we contribute to better protection against image manipulations and the battle against disinformation.

1. Introduction

Manipulating images is easier than ever using tools such as Adobe Photoshop or recently accessible AI-based tools. This has raised significant concerns about the spread of deceptive content, which poses risks when used for misinformation, as false evidence, or for other fraudulent purposes. The potential misuse of manipulated images underscores the pressing need for robust image forgery detection methods. These methods serve as a crucial line of defense in safeguarding the integrity of visual information and mitigating the social ramifications of image-based deception.
While substantial progress has been made in the development of image forgery detection techniques [1,2], a considerable gap remains in their capacity to effectively detect forgeries in real-world scenarios. That is, manipulated images found in the wild are often subjected to alterations that were not seen in a detection method’s training set. For example, methods targeting deepfake (face) forgery detection only [3,4,5] may not be able to detect other types of fakes, such as splicing or copy-move attacks. In other words, generalization is a major challenge of image forgery detection methods. Additionally, the detection performance is hindered by post-manipulation processing, such as decreasing the quality due to compression when shared on social media. Robustness against such (intentional or unintentional) attacks is another major challenge of forgery detection techniques.
Each individual detection method has its own target use case, strengths and limitations. They often operate under specific assumptions and are trained on a certain manipulation trace or forgery type. As a result, forensic investigators have to run and interpret multiple detection methods, which increases their required skill level and makes the investigation harder. Because of the complementary assumptions and methodologies of existing methods, one can expect a boost in performance when combining multiple methods to maximize the number of utilized forensic clues [1]. Additionally, interpreting a single output requires less skill than interpreting multiple outputs.
Existing fusion methods that combine image forgery detection methods typically have some limitations [6,7,8,9,10,11,12]. First, they may not provide generic frameworks to easily add existing forgery detection methods. This would be particularly interesting when new high-performing methods are developed later, and may, therefore, greatly increase the overall performance of the fusion model. However, some existing fusion methods perform a unique post-processing workflow on each individual method [8], adapt the individual method [7], or target the fusion method on different versions of the same methodology or architecture [6,11,12]. This hurdles the later addition of new individual methods to the fusion framework. Second, existing fusion methods often do not include recent deep-learning based methods that demonstrated better performance than the conventional methods that are used instead. Third, existing fusion methods select complementary methods based on their complementary forgery traces. However, they may lack an objective measurement of their complementariness on evaluation datasets. These limitations of existing fusion methods are further discussed in Section 2.2.
This paper explores the synergy of complementary image forgery detection methods and proposes a novel fusion approach to enhance their performance. The main contributions are the following. First, we objectively measure the complementariness of existing forgery detection and localization methods on several evaluation datasets. Second, we propose a novel fusion method based on a deep-learning architecture that combines the heatmaps of existing forgery localization methods. We take special care to provide a generic fusion approach in which new detection methods can be added in a later stage, without the need to develop a unique post-processing workflow for them. Additionally, we include both conventional and more recent high-performing data-driven detection methods. As such, we tackle the limitations of existing fusion methods.
The rest of this paper is organized as follows. First, Section 2 discusses related image forgery detection and fusion methods. Then, we measure and analyze the complementariness of existing forgery detection methods in Section 3. Subsequently, Section 4 proposes a novel fusion method. The performance of the proposed fusion method is experimentally evaluated in Section 5. Finally, the paper is concluded in Section 6.

2. Related Work

This section first provides an overview of the classes of existing image forgery detection and localization algorithms in Section 2.1, with a specific focus on the algorithms utilized in the proposed fusion method. Additionally, existing fusion methods are discussed in Section 2.2.

2.1. Image Forgery Detection and Localization Methods

Image forgery detection methods can be classified based on the type of forensic clue or artifact that they exploit. The two main forensic types are acquisition artifacts and compression artifacts.
Acquisition artifacts originate from the image capture process. Methods exploiting these artifacts focus on anomalies introduced during the transformation of a physical scene into a digital image. Camera lenses, for instance, possess inherent imperfections. These imperfections are harnessed for forgery detection [13,14]. Furthermore, the demosaicing process, which interpolates data from Color Filter Arrays (CFAs), introduces correlation patterns into the resulting image pixels. Any manipulation can distort this correlation pattern, thereby providing a potential signature of tampering [15]. Moreover, methods exploiting photo-response non-uniformity (PRNU) noise have been developed to identify both camera models and manipulations by analyzing inconsistencies in the unique patterns left by the camera sensor [16]. While exploiting the PRNU noise is effective, it requires multiple images from the same camera, which may not always be feasible in practical scenarios. Noiseprint [17] overcomes this limitation by extracting a camera model fingerprint without extensive data from a specific camera. That is to say, during training, Noiseprint used a large dataset of cameras to learn a camera model fingerprint, but during inference, one image was sufficient. Noiseprint is a machine learning algorithm trained using a Siamese network, and the training set contains only real data, hence there is no bias towards specific forgery types that were seen during training. This is what one calls one-class learning. As such, Noiseprint is a practical and generalizable solution. Similar to Noiseprint, SpliceRadar [18] also uses camera model features in a one-class learning architecture. Moreover, EXIF-SC [19] also uses one-class learning, and aims to determine whether every part of the image content is self-consistent with its EXIF metadata.
A second common approach in image forgery detection involves examining anomalies in JPEG compression artifacts. JPEG is a widely used image compression standard. Mismatches in JPEG blocks and the corresponding grid are signs of manipulation [20,21,22]. Conventional detection methods have relied on statistical analysis of these coding artifacts, leading to the creation of handcrafted models. However, the applicability of these models in real-world scenarios remains a challenge due to their assumptions and limitations. An alternative approach involves double quantization or double JPEG compression artifact-based methods [23,24,25,26]. These methods assume that the authentic region is compressed twice, while the manipulated region is compressed only once. This concept has been explored using both conventional methods and deep-learning approaches, where neural networks are trained to automatically extract relevant features. For example, Comprint extracts a compression fingerprint using deep learning, in which inconsistencies suggest manipulation [11]. This is similar to Noiseprint [17], with the difference that it uses compression traces rather than acquisition artifacts. More recent advancements, such as the Compression Artifact Tracing Network (CAT-Net), have employed deep learning and large training sets to exploit both acquisition and JPEG compression artifacts, operating in both the spatial and Discrete Cosine Transform (DCT) domain, offering improved detection capabilities [27]. However, finding sufficient training datasets that cover and generalize to real-world circumstances is a main unsolved challenge in image forgery detection and localization.
In general, deep-learning-based methods have demonstrated high performance. However, there is still a significant gap to close to reach 100% accuracy, especially under in-the-wild circumstances.

2.2. Fusion Methods

A key challenge of fusion is identifying algorithms that complement each other, leading to improved overall robustness and reliability of localization and detection. Several fusion approaches have been proposed in the literature, and they can be categorized as either feature-level or pixel-level fusion.

2.2.1. Feature-Level Fusion

Feature-level fusion occurs at the same level where image forgery detection and localization algorithms extract suitable low-level features from the image. An example of feature-level fusion is Comprint + Noiseprint [11]. Comprint and Noiseprint are complementary strategies, with Comprint exploiting compression artifacts and Noiseprint targeting camera-based acquisition artifacts. To fuse the methods, their localization algorithm is adapted by concatenating the features (i.e., fingerprints) of both Comprint and Noiseprint, and applying expectation-maximization on the concatenated feature array. Comprint + Noiseprint outperforms the individual Comprint and Noiseprint methods. Although feature-level fusion may perform well, such approaches often encounter challenges related to feature selection and management of a large number of features when combining a significant amount of algorithms [1]. Additionally, it is not straightforward to add other existing methods, as this may require transforming those methods into features in the same latent space. As such, pixel-level fusion methods may provide an easier framework for fusion.

2.2.2. Pixel-Level Fusion

In pixel-level fusion, the output heatmaps of the individual forgery localization methods are combined. For example, Korus and Huang proposed to run forgery localization at multiple scales and combine those results [6]. Similarly, Li et al. fused a statistical feature-based approach with a complementary copy-move detection by combining the corresponding tampering possibility maps [7]. Additionally, Liu and Pun proposed a deep fusion network in which two complementary base networks were trained, and subsequently combined in a fusion network [12]. The base networks are complementary because they are each trained on different forgery traces (noise and compression). These methods have in common that it is not straightforward to add other existing methods, since they adapt the individual methods, or target the fusion method on different versions of the same methodology or architecture.
Iakovidou et al. proposed a knowledge-based fusion approach that combines several selected conventional detection algorithms [8]. These algorithms were selected based on their performance and based on the complementarity of their assumptions. This fusion method first refines the individual localization heatmaps by normalizing, binarizing, connecting, and filtering them. Then, statistical features are extracted from these refined maps to automate the evaluation of their usefulness in a fusion unit. The fusion takes into account estimates of the methods’ interpretability, inter-compatibility, reliability (based on previous experiments), and confidence. The fusion approach primarily focused on conventional detection algorithms, while better-performing data-driven algorithms were not evaluated. Additionally, this fusion approach incorporates certain hard-coded heuristics and assumptions about the selected input algorithms, which limits its practical adaptability to adding new forgery detection algorithms.
Some recent fusion methods perform supervised deep-learning directly on the heatmaps (i.e., the individual methods’ outputs). For example, Operation-wise Attention Fusion (OwAF) [9] improves upon the work of Iakovidou et al. [8], by combining conventional methods in a simple CNN architecture for supervised learning. In numerous cases, the fusion outperforms the individual algorithms, as well as the older fusion algorithm [8]. Similarly, Siopi et al. proposed a multi-stream fusion network that combines two existing complementary conventional detection algorithms (one based on noise and one based on compression) [10]. The input image and the output heatmaps of the selected methods are processed by different encoder stream networks, and the encoding outputs are subsequently fused. These fusion methods make it more practical to add new existing detection methods to the set of input methods. However, a limitation of these models is their lack of generalization to manipulations unseen during training, which is a common problem in supervised learning [28]. Consequently, it occasionally yields worse results compared to the best individual input detection algorithm. Furthermore, the absence of data-driven algorithms in the fusion is a recurring observation among existing fusion methods, despite their demonstrated superior performance over conventional detection algorithms.
In summary, the current state-of-the-art does not offer a universal solution that addresses all challenges in image forgery detection and localization. Existing fusion algorithms exhibit several limitations, as they often rely on predefined heuristics and do not cover recent data-driven algorithms. Additionally, the complementariness of the individual methods is never objectively measured but is only motivated based on the assumptions that they make. Hence, further advancements in this field are needed to develop more robust and effective fusion approaches.

3. Selection & Complementariness of Image Forgery Detection Methods

This section discusses the selected image forgery detection methods and measures their complementariness. Note, that we objectively measure the complementarity, which is novel compared to previous fusion methods that typically select complementary methods solely based on the complementary assumptions that the individual methods make (i.e., they are based on complementary forensic clues).
This section is organized as follows. First, the experimental setup of this analysis is described in Section 3.1. Then, the results of this analysis are reported and discussed in Section 3.2.

3.1. Experimental Setup

3.1.1. Image Forgery Detection Algorithms

In this study, we evaluate a subset of conventional detection algorithms (ADQ1, DCT, BLK, CAGI) that were also chosen in state-of-the-art fusion methods [8,9]. Additionally, we add more recent, high-performing, data-driven algorithms (Noiseprint, Comprint, Comprint + Noiseprint, and CAT-Net). It is noteworthy that Comprint + Noiseprint in itself is already a feature-level fusion method. All selected algorithms are listed in Table 1. Some algorithms are based on JPEG compression artifacts, whereas others on camera-based acquisition artifacts.

3.1.2. Evaluation Datasets

We selected a diverse set of datasets containing manipulated images, on which we evaluated the (complementary) performance of the selected forgery detection methods. These datasets cover a wide range of manipulations, created by both traditional editing software and artificial intelligence (AI) techniques. In total, we selected seven datasets, listed in Table 2, each presenting different challenges and characteristics. Additionally, some subsets of those datasets are used for training (see Section 5.1.1).

3.1.3. Performance Metrics

To evaluate the performance of the image forgery localization methods on the evaluation datasets, we measure the pixel-level forgery localization performance using the F1 score. The F1 score is the harmonic mean of the precision and recall, and is calculated as follows:
F 1 = 1 1 precision + 1 recall = 2 TP 2 TP + FN + FP
In Equation (1), TP, FP and FN represent the number of pixel predictions that are True Positives, False Positives, and False Negatives, respectively. Note, that a higher F1 score (i.e., closer to 1) is better than a lower one (i.e., closer to 0). To classify pixels as either real or fake, the corresponding value in the heatmap needs to be compared to a certain threshold. To eliminate the threshold’s influence on the calculation, the maximum F1 score across all possible thresholds is reported in this paper. Additionally, it is essential to account for cases where the heatmap’s polarity is inverted relative to the ground truth. To address this, we consider both the original and inverted heat maps, and report the maximum F1 score.
Note, that in this section, we only measure the pixel-level localization performance, and we do not (yet) consider the image-level detection performance. When evaluating the proposed fusion method in Section 5, the image-level detection performance metric is described and additionally evaluated.

3.1.4. Theoretical Oracle Fusion & Complementary F1 Delta Metric

To evaluate how complementary a set of forgery localization methods are, we define two new concepts: the theoretical oracle fusion, and the Complementary F1 Delta metric.
The theoretical oracle fusion is the maximum F1 score of all methods, taken for each image separately. In essence, it is an oracle that, for each image in a certain dataset, knows which is the best-performing forgery localization method, and selects that one as its output. In other words, a different algorithm may be selected for each image. As such, the oracle fusion can be seen as a target score of a fusion algorithm on a certain dataset.
The Complementary F1 Delta is a measure of complementariness between two detection methods. It calculates the improvement in F1 score between using one of the two methods as a reference algorithm, on the one hand, and the theoretical oracle fusion of both methods, on the other. When the Complementary F1 Delta is zero, it means that the reference method performs equal or better than the second method for every image in the dataset. A larger Complementary F1 Delta signifies that the two detection methods are more complementary, and hence each shows strengths in different scenarios.

3.2. Performance & Complementariness Results

We measured the pixel-level forgery localization performance (F1 score) of each forgery detection method on each evaluation dataset. Additionally, we measured the Complementary F1 Delta for each combination of two methods. These results are shown in Appendix A, i.e., in Table A1, Table A2, Table A3, Table A4, Table A5, Table A6 and Table A7 for the VIPP, IMD2020, DSO-1, OpenForenics, FaceSwap, Coverage, and NC2016 dataset, respectively. These tables highlight the three largest Complimentary F1 Delta values for each row in bold, underline and italics, respectively. The same highlighting is used for the column with F1 scores.
From the tables in Appendix A, we notice that the recent CAT-Net method performs best on all datasets, except for the DSO-1 dataset where Comprint + Noiseprint performs best. This showcases that these more recent data-driven detection methods perform better than conventional detection methods. Recall that existing fusion methods often limit themselves to combining existing conventional detection methods. Hence, including these newer, high-performing data-driven methods is one of the contributions of our work.
Regarding the complementariness of CAT-Net, we can see the Complementary F1 Delta when combining CAT-Net with each of the other methods in the first row of each table in Appendix A. On all datasets, combining CAT-Net with Comprint + Noiseprint results in the largest localization performance improvement (except for OpenForensics, where combining CAT-Net with Comprint performs slightly better). For example, on the FaceSwap dataset in Table A5, the individual F1 score of CAT-Net is 0.449, and the individual F1 score of Comprint + Noiseprint is 0.412. By combining these two methods using the theoretical oracle fusion, we can increase CAT-Net’s F1 score by 0.130. This demonstrates the complementarity of these two high-performing methods.
Although the tables in Appendix A showcase the complementarity of each two methods, they do not consider the fusion of all methods. Therefore, we additionally show the localization performance of theoretical oracle fusion of all methods, in Table 3. This table shows the F1 scores of each individual method (again), in addition to the best performance of fusing just two methods, and the performance of fusing all methods, using the theoretical oracle. Moreover, the last row of the table shows the performance of the proposed fusion method, which is irrelevant in this section, and is further discussed in Section 5.2.1. Note, that in this table, we highlight the three largest values for each column in bold, underline and italics, respectively. It is no surprise that the first and second largest values of each column are the oracle fusion rows, as these take the image-wise maximum of the individual methods.
From Table 3, we notice that the F1 score can be improved by combining two methods, and even more so when combining all selected methods. For example, CAT-Net has the best individual localization performance on the NC2016 dataset (F1 score of 0.487). The best oracle fusion of 2 methods increases this F1 score to 0.594 (by combining CAT-Net with Comprint + Noiseprint, see Table A7). Moreover, the oracle fusion of all methods brings the performance to an F1 score of 0.626. This is a significant boost in performance, demonstrating the potential of fusing all methods. Hence, in Section 4, we propose a practical fusion method that combines all these detection methods.
In summary, this section demonstrated the potential of fusing existing image forgery detection and localization methods. It should be noted that we analyzed the performance using a theoretical oracle fusion, which cannot be applied in real-world scenarios. Instead, it serves only as a demonstration of fusion potential and aids in measuring the complementariness of existing methods. As such, we motivate combining the selected methods in our proposed fusion method, in Section 4.

4. Proposed Fusion Method

In this section, we propose a novel pixel-level fusion method that combines the outputs of the selected image forgery detection and localization methods listed in Table 1, in addition to the investigated image.
We interpret the problem of fusing multiple heatmaps into a single heatmap as an image-to-image translation task. A widely adopted deep-learning architecture for addressing image-to-image challenges is the Generative Adversarial Network (GAN) [36,37,38,39,40]. In our context, a conditional Generative Adversarial Network (cGAN) serves as the foundation of the proposed fusion algorithm. A GAN consists of two fundamental models: a Generator (G) and a Discriminator (D). These two models engage in a competitive dynamic, with the generator striving to generate the fused heatmaps as close as possible to the ground truth, while the discriminator aims to distinguish the difference between the ground truth and fused heatmaps. In the context of GANs, the goal is to learn a transformation from a random noise vector z to an output image y, denoted as G : z y . However, given our objective of using heatmaps as input, the utilization of a conditional GAN (cGAN) is more fitting. This type of GAN utilizes additional conditional information, x, resulting in G : { x , z } y , in which x is the concatenation of the heatmaps and original image [41].
In the following sections, we delve into the employed training methodologies and architectures (Section 4.1). Subsequently, we discuss how the fusion algorithm is used during inference (Section 4.2).

4.1. Training Methodology

Our proposed fusion method is based on a similar training objective as typical cGANs, such as the Pix2Pix method. Pix2Pix is a widely used model in image-to-image translation [42]. In general, the main objective of a cGAN is to identify the optimal generator ( G * ) capable of generating outputs that closely resemble the distribution of real data (i.e., the ground truth forgery masks). In our proposed method, finding the optimal generator G * is formulated as follows:
G * = arg min G max D L c G A N ( G , D ) + λ L L 2 ( G ) .
In other words, G seeks to minimize the cGAN loss against the adversarial D, which in contrast strives to maximize the loss.
In Equation (2), we combine the typical loss function of a cGAN L c G A N with the L2 loss L L 2 ( G ) . These loss functions are expressed as follows:
L c G A N = E x , y [ log D ( x , y ) ] + E x , z [ log ( 1 D ( x , G ( x , z ) ) ] ,
L L 2 ( G ) = E x , y , z [ | | y G ( x , z ) | | 2 ] .
By including the L2 loss, the generator has an additional goal of producing images that closely resemble the ground truth. Note, that we utilize the L2 loss, which is in contrast to Pix2Pix which uses the L1 loss [42]. In Pix2Pix, the L1 loss was chosen to minimize image blurring. However, blurriness has less impact on performance in our forgery localization task. Instead, using the L2 loss intensifies the penalty for generating noisy heatmaps, amplifying regions with significant errors and disregarding parts with negligible error, due to the quadratic nature of the L2 distance.
It is additionally worth noting that the actual input no longer involves random noise (z), as this approach was found to be ineffective [42]. Instead, noise is introduced through dropout, applied across multiple layers of the generator [42]. This is conducted both during training and inference (the latter is further discussed in Section 4.2).
The generator and discriminator are individually discussed in depth in Section 4.1.1 and Section 4.1.2, respectively.

4.1.1. Generator

The high-level process of training the generator is visualized in Figure 1. The generator loss consists of two parts.
First, it aims to minimize the difference between the generator’s output (fused heatmap) and the discriminator’s expectations. This is achieved by using a sigmoid-cross-entropy loss of the discriminator’s output and an array of ones (which signifies the discriminator’s expectation for every patch to originate from the ground truth). A low score indicates effective deception of the discriminator and vice versa.
Second, the generator aims to minimize the difference between the fused heatmap and the ground truth. This is achieved by using the L2 loss, also known as the Mean Squared Error (MSE), between the fused heatmap and the ground truth. The overall loss is the sum of the sigmoid-cross-entropy loss and the MSE (the latter first multiplied by lambda, which is set to λ = 100 , as conducted in Pix2Pix [42]). This combined loss adapts the generator’s neural network weights during training, causing it to become better at generating good fused heatmaps.
The generator’s architecture is based on U-Net [43], with some modifications. The U-Net architecture involves two main parts: an encoder which iteratively downsamples the previous layers, and a decoder which iteratively upsamples the layers back to the original resolution. Each iterative component contains convolution, batch normalization and Rectified Linear Unit (ReLu) layers. Moreover, the first three upsample components include a dropout layer in between the batch normalization and ReLu. This dropout accounts for randomness, as discussed in Section 4.1. Additionally, skip connections link the corresponding decoder and encoder blocks at the same resolutions. These connections concatenate their channels, allowing information to be shared between them.

4.1.2. Discriminator

The high-level process of training the discriminator is visualized in Figure 2. The discriminator’s loss is separated into two parts.
First, the fused heatmap is presented to the discriminator along with the input images (i.e., the heatmaps and investigated image). The discriminator’s output is then employed to calculate the sigmoid-cross-entropy loss against an array of zeros (which, in contrast to an array of ones, signifies the discriminator’s expectation for every patch to be originating from the fused heatmap, and not the ground truth). This constitutes the generated loss.
Second, the ground truth and the input images are jointly supplied to the discriminator. The discriminator’s output is again utilized to compute the sigmoid-cross-entropy loss, yet this time against an array of ones (signifying the expectation to be originating from the ground truth). This constitutes the real loss. The final discriminator loss is the sum of the generated loss and the real loss.
The discriminator’s architecture is based on a convolutional PatchGAN classifier [42], in order to classify whether image patches originate from either a fused heatmap or the ground truth. This architecture consists of multiple downsampling blocks responsible for creating patches. These downsampling blocks also follow the convolution, batch normalization, and ReLu structure. Additionally, a layer of spectral normalization [44] is introduced after each convolutional layer. This spectral normalization layer re-normalizes the weights of the convolutional layer whenever they are updated, creating a network that mitigates gradient explosion problems, and decreasing the discriminator’s rapid learning rate. Consequently, it allows the generator more time to learn during the initial epochs, where it may struggle.

4.2. Inference Methodology

When applying the model in practice, on real-world images, the fusion model runs in inference mode. The inference process of our fusion algorithm mirrors that of the Pix2Pix architecture. Specifically, only the generator component of the algorithm is utilized.
Note, that our proposed model utilizes the dropout layers of the generator during its inference. This is in contrast to a typical inference protocol, in which dropout layers are not utilized during inference, but only during training. However, in the context of the random nature inherent to (c)GANs, the dropout layer continues to be employed during inference. This was also conducted in Pix2Pix [42].
As such, the fusion algorithm can effectively process new images and produce fused heatmaps that are well-suited to real-world scenarios.

5. Evaluation of Proposed Fusion Method

This section evaluates the proposed fusion method described in Section 4. First, the experimental setup is described in Section 5.1. Then, the results of forgery localization and forgery detection performance are discussed in Section 5.2.

5.1. Experimental Setup

5.1.1. Training Setup

For this study, the OpenForensics [30], FaceSwap [31], and IMD2020 [29] datasets of manipulated images are used for training the model. These datasets were chosen due to their size and wide coverage of diverse forgery types. This leaves the other datasets listed in Table 2 purely for evaluation purposes. In the selected datasets for training, a train/validation/test split is made, distributing the data with an 80/10/10 ratio. Images are randomly selected from the datasets for each split. Additionally, to create a balanced dataset, we included an equal number of training and validation images from each of the datasets. Among the three, FaceSwap is the smallest dataset, having 703 training and 88 validation images. As such, the same quantities of OpenForensics and IMD2020 data are used during training, resulting in a training dataset of 2109 images and a validation dataset of 264 images. Note, that the test datasets are not artificially limited, and remain 10% of the size of the full datasets.
For every image in the dataset, both the image itself and all heat maps produced by the selected input algorithms from Table 1 are utilized as input. It is noteworthy that one of the input algorithms is Comprint + Noiseprint, which is a feature-level fusion method. As such, we combine feature-level and pixel-level fusion. Before using these input images for training, a preprocessing stage is conducted, similar to in Pix2Pix [42]. That is, all heatmaps and the investigated image are first resized to a resolution of 256 × 256 pixels. Afterward, the training set is augmented by resizing the images to 286 × 286 and then randomly cropping them back to the size of 256 × 256. Additionally, random mirroring is implemented by horizontally flipping the image (left to right). All inputs are concatenated at the channel level, generating a composite input size of 256 × 256 × 11 pixels for each image. Eleven channels are the result of concatenating the three RGB channels of the input image with the 8 channels of the single-channel heatmaps.
The actual training exists out of 40 epochs with a batch size of 1, yielding 2109 steps per epoch. We adopt the Adam solver, following the parameters used in Pix2Pix [42]: a learning rate of 0.0002, and momentum parameters β 1 = 0.5 and β 2 = 0.999 .

5.1.2. Evaluation Setup

The datasets used for evaluating the model’s performance are the same as those discussed in Section 3.1.2, and listed in Table 2. However, the datasets used for the training of the model have been swapped for the newly-created test split. The other evaluation datasets remain unchanged.
In addition to the pixel-level forgery localization performance metric (F1 score), we additionally measure the image-level forgery detection performance to evaluate the fusion model. As an image-level forgery detection measure, we utilize the Area Under the ROC Curve (AUC). That is, first, a global statistic is extracted from the heatmap. Next, this global statistic is compared to a threshold, which determines whether the algorithm predicts the image to be forged or pristine. To make the evaluation independent of a certain threshold, the Receiver Operating Characteristic (ROC) curve is used, which illustrates the TP rate against the FP rate at different threshold values for each dataset. Afterward, the ROC is used to compute the AUC, which integrates the ROC curve over all threshold values, yielding a single value from 0 to 1. An AUC of 0.5 indicates random performance. Higher AUC values indicate better detection capabilities. To ensure the evaluation’s independence from the choice of the global statistic, the global AUC score is computed as the maximum AUC among several considered global statistic options, namely the average, median, standard deviation, 99.5th percentile, and 97.5th percentile. This approach mitigates any potential bias that might arise from favoring a particular global statistic.
In addition to comparing the performance of the proposed fusion method to the individual methods’ performance, and the theoretical oracle fusion performance, we also compare it to two simple fusion strategies: Maximum and Minimum. These fusion methods transform the input heatmaps from the image forgery detection methods into a fused heatmap. Each pixel value of a fused heatmap is the maximum (or minimum, respectively) of the corresponding pixel values in the input heatmaps. For these simple fusion methods, we selected only CAT-Net and Comprint + Noiseprint, which are the two best performing forgery detection methods (as discussed in Section 3.2. We do this because we noticed noisy fused heatmaps with worse performance when additionally using the other 6 input methods.
Note, that we do not compare the results of our proposed method to related fusion methods. That is because existing methods fuse other methods than in our work, often only focusing on conventional methods. Moreover, in many state-of-the-art fusion methods, it is not straightforward to add new detection methods as input. Even when it is relatively more easy, it may require re-training those models. Finally, existing fusion methods are typically not available open-source.

5.2. Results

We discuss the performance of forgery localization and detection in Section 5.2.1 and Section 5.2.2, respectively.

5.2.1. Image Forgery Localization Performance

The F1 scores for forgery localization of the proposed fusion method are given in the last row of Table 3. Recall that, in this table, we highlight the three largest values for each column in bold, underline and italics, respectively. Additionally, recall that the oracle fusion models are theoretical and cannot be applied in practice. Instead, they only serve as a theoretical target that demonstrates the fusion potential. The proposed fusion method is not evaluated on the full IMD2020, OpenForensics and FaceSwap datasets, as it was trained on a portion of these datasets. Instead, the F1 score of the test splits of IMD2020, OpenForensics and FaceSwap are given in Table 4.
The proposed fusion algorithm demonstrates exceptional localization performance on the test splits of the IMD2020, OpenForensics, and Faceswap datasets in Table 4. Impressively, it even surpasses the target of theoretical oracle fusion of all methods. Additionally, it also surpasses the Maximum and Minimum fusion methods.
In contrast to the test splits of the datasets that were used during training, evaluating the localization performance on previously unseen datasets performs less impressively, as presented in Table 3. On those datasets, the proposed fusion method outperforms the individual forgery localization methods for the VIPP and NC2016 dataset, but scores worse than the best individual method for the DSO-1 and Coverage dataset. Additionally, the oracle fusion methods scored higher than the proposed fusion methods in all those datasets. This demonstrates the lack of generalization on unseen datasets, which is also an issue in related work that uses supervised learning [9,10].
In our case, the generalization issue may be specifically attributed to the substantial performance gap between CAT-Net and the other input algorithms, which is especially prevalent on the IMD2020 and OpenForensics training datasets, while less outspoken on the FaceSwap training dataset. Consequently, this creates a bias towards the heatmaps generated by the CAT-Net algorithm. This lack of generalization is also an implication of using two-class learning in image forgery localization/detection, in contrast to one-class learning where no fake images are seen during training [1,11,17]. In such two-class learning methods, obtaining a training dataset that contains all possible forgeries is impractical. Although the method does not generalize very well to unseen datasets, it should be noted that the performance of the proposed fusion method is still relatively good. Most notably, the proposed method performs better than all individual input methods on two datasets (VIPP and NC2016). On the other datasets (DSO-1 and Coverage), the proposed fusion method still obtains an F1 score between the top-performing methods, and may hence still help to provide forensic investigators with a single reliable heatmap, rather than a large set of heatmaps from all individual methods. As such, the proposed fusion method can be used in practical applications.
When comparing the proposed fusion method to the simple Maximum and Minimum fusion methods, we make the following observations. For the DSO-1 and NC2016 methods, the simple fusion methods showcase a slightly better performance. However, for the VIPP and Coverage datasets, as well as the test splits of IMD2020, OpenForensics, and Faceswap, the proposed fusion method outperforms the simple fusion methods.
Visual examples of the forgery localization heatmaps can be observed in Figure 3. These examples demonstrate the ability of the fusion method to successfully learn to combine the heatmaps of the individual methods into a more accurate fused heatmap.

5.2.2. Image Forgery Detection Performance

The AUC scores for forgery detection are given in Table 5. Note, that the OpenForensics dataset is not present in the table, as it does not contain real images and can, therefore, not be evaluated on the binary forgery detection task. Additionally, note that we evaluate the full datasets (as in Table 3), and not the test splits of the datasets (as in Table 4).
In Table 5, we notice similar trends as seen in the forgery localization performance discussion in Section 5.2.1. That is, the proposed fusion method obtains relatively high AUC values, but does not perform better than the best individual method. Notably, on DSO-1, a significant gap is observed between the best individual method and the proposed fusion method. This worse performance can again be attributed to the fusion method being biased towards CAT-Net heatmaps, as also discussed in Section 5.2.1. That is, on DSO-1, CAT-Net’s performance is particularly worse than Comprint, Noiseprint, and Comprint + Noiseprint, whereas it is typically the opposite in the other evaluation datasets and in the training datasets. As such, when CAT-Net performs badly, the bias towards CAT-Net has a particularly negative effect on the observed fusion detection performance. Despite the fusion method not performing better than the best individual method, it is still a relatively well-performing method and is positioned in second place for the VIPP, Coverage, and NC2016 datasets. Moreover, in comparison with the simple fusion methods (Maximum and Minimum), the proposed fusion method showcases worse AUC values for the DSO-1 and NC2016 datasets, but better results for the VIPP and Coverage datasets. This is the same observation we made when discussing the image forgery localization performance in Section 5.2.1.
In general, this section demonstrated that the proposed fusion method is a viable approach to combine the outputs of multiple individual methods into a single fused heatmap.

6. Conclusions

This study aims to enhance image forgery detection and localization methods, by uniting existing forgery detection methods.
We evaluated the performance and complementariness of selected existing methods, focusing on both conventional and more recent data-driven methods. We objectively demonstrated the potential to combine multiple detection methods on the basis of the theoretical oracle performance and Complementary F1 Delta metric, which are novel concepts proposed for the first time in this paper.
Moreover, we propose a deep-learning-based fusion method that is trained using cGAN architecture. Our approach shows good results in enhancing forgery localization. Although it sometimes struggles to generalize on unseen datasets, its performance is relatively high and can help forensic investigators with the interpretation of the results of forgery detection and localization methods.
Future work may tackle the generalization issue by utilizing more training data, from a larger variety of datasets with a larger variety of forgery types. Additionally, the generator architecture in our proposed framework could be replaced with cutting-edge alternatives such as vision transformers [45] in order to further refine the performance, which can be inspired by related work that evaluated the generalization of deep-learning models [28]. Finally, we can improve the overall fusion performance by including new high-performing methods, such as TruFor [46] and FOCAL [47], which are relatively easy to include in our proposed fusion framework.
In conclusion, this work contributes to a deeper understanding of image forgery detection and localization. We showcased their complimentariness, as well as both the promising results and limitations of fusion. As such, we further protect our society from misinformation, false evidence, and fraud.

Author Contributions

Conceptualization, H.M. and G.V.W.; methodology, L.D.N. and H.M.; software, L.D.N. and H.M.; formal analysis, L.D.N. and H.M.; writing—original draft preparation, L.D.N. and H.M.; writing—review and editing, P.L. and G.V.W.; supervision, P.L. and G.V.W.; project administration, H.M. and P.L. and G.V.W.; funding acquisition, H.M. and P.L. and G.V.W. All authors have read and agreed to the published version of the manuscript.

Funding

This work was funded in part by the Flemish government (COM-PRESS project, within the relanceplan Vlaamse Veerkracht), IDLab (Ghent University—imec), Research Foundation—Flanders (FWO), Flanders Innovation & Entrepreneurship (VLAIO), and the European Union.

Data Availability Statement

The image forgery detection methods and image forgery datasets used in this proposal are open-source and can be used for research purposes. Links to these methods and datasets can be found in Table 1 and Table 2, respectively.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A. Performance and Complementarity Tables

Table A1. VIPP dataset: forgery localization performance & Complimentary F1 Delta. We highlight the three largest values for each column in bold, underline and italics.
Table A1. VIPP dataset: forgery localization performance & Complimentary F1 Delta. We highlight the three largest values for each column in bold, underline and italics.
Dataset: VIPP [32]
AlgorithmF1Complementary F1 Delta
CAT-NetADQ1BLKCAGIDCTComprComprNoisepr + Noisepr
CAT-Net [27]0.716-0.0390.0490.0190.0450.0370.0910.088
ADQ1 [23]0.5030.246-0.0870.0720.0540.1220.1550.146
BLK [20]0.4310.3340.160-0.1170.0970.1540.2060.175
CAGI [21]0.4380.2990.1440.111-0.0950.1340.1870.180
DCT [22]0.4320.3290.1320.0960.100-0.1470.1980.184
Comprint [11]0.4960.2570.1380.0890.0750.083-0.0950.139
Compr + Noisepr [11]0.5810.2270.0870.0570.0440.0500.011-0.062
Noiseprint [17]0.5560.2480.0900.0500.0620.0600.0790.086-
Table A2. IMD2020 dataset: forgery localization performance & Complimentary F1 Delta. We highlight the three largest values for each column in bold, underline and italics.
Table A2. IMD2020 dataset: forgery localization performance & Complimentary F1 Delta. We highlight the three largest values for each column in bold, underline and italics.
Dataset: IMD2020 [29]
AlgorithmF1Complementary F1 Delta
CAT-NetADQ1BLKCAGIDCTComprCompr + NoiseprNoisepr
CAT-Net [27]0.850-0.0010.0030.0030.0020.0060.0080.006
ADQ1 [23]0.2920.560-0.0640.0870.0740.1380.1730.150
BLK [20]0.2630.5900.093-0.0960.0850.1690.2010.162
CAGI [21]0.2970.5560.0830.062-0.0980.1510.1810.146
DCT [22]0.3130.5380.0520.0350.081-0.1350.1630.134
Comprint [11]0.2970.4660.0400.0410.0580.058-0.0710.087
Compr + Noisepr [11]0.4370.4210.0290.0270.0410.0400.025-0.034
Noiseprint [17]0.3960.4610.0460.0290.0480.0520.0820.075-
Table A3. DSO-1 dataset: forgery localization performance & Complimentary F1 Delta. We highlight the three largest values for each column in bold, underline and italics.
Table A3. DSO-1 dataset: forgery localization performance & Complimentary F1 Delta. We highlight the three largest values for each column in bold, underline and italics.
Dataset: DSO-1 [33]
AlgorithmF1Complementary F1 Delta
CAT-NetADQ1BLKCAGIDCTComprCompr + NoiseprNoisepr
CAT-Net [27]0.675-0.0290.0720.0720.0190.1760.2020.201
ADQ1 [23]0.4200.301-0.0820.1490.0240.3650.4050.396
BLK [20]0.4560.2910.043-0.1310.0140.3490.3860.378
CAGI [21]0.5120.2350.0530.076-0.0130.2910.3310.322
DCT [22]0.3470.3470.0820.1230.179-0.4200.4700.466
Comprint [11]0.7630.0880.0160.0430.0400.004-0.0550.081
Compr + Noisepr [11]0.8130.0630.0080.0290.0300.0030.004-0.038
Noiseprint [17]0.8100.0670.0070.0250.0250.0030.0340.042-
Table A4. OpenForensics dataset: forgery localization performance & Complimentary F1 Delta. We highlight the three largest values for each column in bold, underline and italics.
Table A4. OpenForensics dataset: forgery localization performance & Complimentary F1 Delta. We highlight the three largest values for each column in bold, underline and italics.
Dataset: OpenForensics [30]
AlgorithmF1Complementary F1 Delta
CAT-NetADQ1BLKCAGIDCTComprCompr + NoiseprNoisepr
CAT-Net [27]0.948-0.0010.0010.0020.0010.0120.0120.009
ADQ1 [23]0.4830.466-0.0200.0460.0270.2520.2770.225
BLK [20]0.2630.6860.240-0.0860.1780.3860.4530.411
CAGI [21]0.2940.6570.2350.056-0.1820.3650.4290.389
DCT [22]0.4230.5260.0870.0190.053-0.2990.3290.274
Comprint [11]0.6300.3310.1050.0190.0280.092-0.1080.151
Compr + Noisepr [11]0.7110.2490.0490.0060.0120.0410.028-0.058
Noiseprint [17]0.6710.2860.0370.0030.0120.0250.1100.098-
Table A5. FaceSwap dataset: forgery localization performance & Complimentary F1 Delta. We highlight the three largest values for each column in bold, underline and italics.
Table A5. FaceSwap dataset: forgery localization performance & Complimentary F1 Delta. We highlight the three largest values for each column in bold, underline and italics.
Dataset: FaceSwap [31]
AlgorithmF1Complementary F1 Delta
CAT-NetADQ1BLKCAGIDCTComprCompr + NoiseprNoisepr
CAT-Net [27]0.449-0.0680.0160.0440.0320.0990.1300.105
ADQ1 [23]0.2820.235-0.0230.0480.0240.1270.1730.153
BLK [20]0.1060.3590.199-0.1090.1040.2650.3180.249
CAGI [21]0.1840.3100.1460.031-0.0840.2100.2590.203
DCT [22]0.1910.2900.1150.0190.077-0.1920.2440.192
Comprint [11]0.3510.1990.0570.0200.0430.032-0.0740.125
Compr + Noisepr [11]0.4120.1680.0410.0110.0310.0220.013-0.060
Noiseprint [17]0.3470.2080.0860.0080.0400.0350.1280.125-
Table A6. Coverage dataset: forgery localization performance & Complimentary F1 Delta. We highlight the three largest values for each column in bold, underline and italics.
Table A6. Coverage dataset: forgery localization performance & Complimentary F1 Delta. We highlight the three largest values for each column in bold, underline and italics.
Dataset: Coverage [34]
AlgorithmF1Complementary F1 Delta
CAT-NetADQ1BLKCAGIDCTComprCompr + NoiseprNoisepr
CAT-Net [27]0.573-0.0030.0050.0320.0040.0370.0430.027
ADQ1 [23]0.2110.365-0.0400.0940.0230.1460.1650.128
BLK [20]0.2420.3350.008-0.0730.0100.1230.1410.107
CAGI [21]0.2980.3070.0060.017-0.0090.1040.1180.086
DCT [22]0.2220.3550.0120.0300.085-0.1390.1570.120
Comprint [11]0.3490.2610.0090.0170.0530.013-0.0630.046
Compr + Noisepr [11]0.3680.2480.0080.0150.0490.0110.044-0.030
Noiseprint [17]0.3320.2680.0070.0170.0530.0100.0630.066-
Table A7. NC2016 dataset: forgery localization performance & Complimentary F1 Delta. We highlight the three largest values for each column in bold, underline and italics.
Table A7. NC2016 dataset: forgery localization performance & Complimentary F1 Delta. We highlight the three largest values for each column in bold, underline and italics.
Dataset: NC2016 [35]
AlgorithmF1Complementary F1 Delta
CAT-NetADQ1BLKCAGIDCTComprCompr + NoiseprNoisepr
CAT-Net [27]0.487-0.0170.0340.0550.0100.0940.1070.097
ADQ1 [23]0.2060.298-0.0770.1290.0320.2200.2510.227
BLK [20]0.2330.2870.050-0.0920.0280.1860.2180.192
CAGI [21]0.2930.2500.0390.030-0.0260.1390.1690.146
DCT [22]0.1830.3130.0550.0790.137-0.2390.2660.239
Comprint [11]0.3980.1830.0270.0210.0370.024-0.0600.073
Compr + Noisepr [11]0.4390.1540.0180.0120.0260.0090.019-0.022
Noiseprint [17]0.4130.1710.0210.0130.0300.0100.0590.048-

References

  1. Verdoliva, L. Media Forensics and DeepFakes: An overview. IEEE J. Sel. Top. Signal Process. 2020, 14, 910–932. [Google Scholar] [CrossRef]
  2. Korus, P. Digital image integrity—A survey of protection and verification techniques. Digit. Signal Process. 2017, 71, 1–26. [Google Scholar] [CrossRef]
  3. Salvi, D.; Liu, H.; Mandelli, S.; Bestagini, P.; Zhou, W.; Zhang, W.; Tubaro, S. A Robust Approach to Multimodal Deepfake Detection. J. Imaging 2023, 9, 122. [Google Scholar] [CrossRef] [PubMed]
  4. Wodajo, D.; Atnafu, S. Deepfake Video Detection Using Convolutional Vision Transformer. arXiv 2021, arXiv:2102.11126. [Google Scholar]
  5. Masood, M.; Nawaz, M.; Malik, K.M.; Javed, A.; Irtaza, A.; Malik, H. Deepfakes generation and detection: State-of-the-art, open challenges, countermeasures, and way forward. Appl. Intell. 2023, 53, 3974–4026. [Google Scholar] [CrossRef]
  6. Korus, P.; Huang, J. Multi-Scale Fusion for Improved Localization of Malicious Tampering in Digital Images. IEEE Trans. Image Process. 2016, 25, 1312–1326. [Google Scholar] [CrossRef]
  7. Li, H.; Luo, W.; Qiu, X.; Huang, J. Image forgery localization via integrating tampering possibility maps. IEEE Trans. Inf. Forensics Secur. 2017, 12, 1240–1252. [Google Scholar] [CrossRef]
  8. Iakovidou, C.; Papadopoulos, S.; Kompatsiaris, Y. Knowledge-Based Fusion for Image Tampering Localization. In Proceedings of the Artificial Intelligence Applications and Innovations, 16th IFIP WG 12.5 International Conference, AIAI 2020, Neos Marmaras, Greece, 5–7 June 2020; Maglogiannis, I., Iliadis, L., Pimenidis, E., Eds.; Springer: Cham, Switzerland, 2020; pp. 177–188. [Google Scholar]
  9. Charitidis, P.; Kordopatis-Zilos, G.; Papadopoulos, S.; Kompatsiaris, I. Operation-wise Attention Network for Tampering Localization Fusion. In Proceedings of the 2021 International Conference on Content-Based Multimedia Indexing (CBMI), Lille, France, 28–30 June 2021. [Google Scholar]
  10. Siopi, M.; Kordopatis-Zilos, G.; Charitidis, P.; Kompatsiaris, I.; Papadopoulos, S. A Multi-Stream Fusion Network for Image Splicing Localization. In Proceedings of the International Conference on Multimedia Modeling, Bergen, Norway, 9–12 January 2023; Springer: Cham, Switzerland, 2023; pp. 611–622. [Google Scholar]
  11. Mareen, H.; Vanden Bussche, D.; Guillaro, F.; Cozzolino, D.; Van Wallendael, G.; Lambert, P.; Verdoliva, L. Comprint: Image Forgery Detection and Localization Using Compression Fingerprints. In Proceedings of the Pattern Recognition, Computer Vision, and Image Processing—ICPR 2022 International Workshops and Challenges, Montreal, QC, Canada, 21–25 August 2022; Rousseau, J.J., Kapralos, B., Eds.; Springer: Cham, Switzerland, 2023; pp. 281–299. [Google Scholar] [CrossRef]
  12. Liu, B.; Pun, C.M. Deep fusion network for splicing forgery localization. In Proceedings of the Proceedings of the European Conference on Computer Vision (ECCV) Workshops, Munich, Germany, 8–14 September 2018. [Google Scholar]
  13. Yerushalmy, I.; Hel-Or, H. Digital image forgery detection based on lens and sensor aberration. Int. J. Comput. Vis. 2011, 92, 71–91. [Google Scholar] [CrossRef]
  14. Mayer, O.; Stamm, M.C. Accurate and efficient image forgery detection using lateral chromatic aberration. IEEE Trans. Inf. Forensics Secur. 2018, 13, 1762–1777. [Google Scholar] [CrossRef]
  15. Popescu, A.; Farid, H. Exposing digital forgeries in color filter array interpolated images. IEEE Trans. Signal Process. 2005, 53, 3948–3959. [Google Scholar] [CrossRef]
  16. Lin, X.; Li, C.T. Refining PRNU-based detection of image forgeries. In Proceedings of the 2016 Digital Media Industry Academic Forum (DMIAF), Santorini, Greece, 4–6 July 2016; pp. 222–226. [Google Scholar] [CrossRef]
  17. Cozzolino, D.; Verdoliva, L. Noiseprint: A CNN-based camera model fingerprint. IEEE Trans. Inf. Forensics Secur. 2019, 15, 144–159. [Google Scholar] [CrossRef]
  18. Ghosh, A.; Zhong, Z.; Boult, T.E.; Singh, M. SpliceRadar: A Learned Method For Blind Image Forensics. In Proceedings of the CVPR Workshops, Long Beach, CA, USA, 16–20 June 2019; pp. 72–79. [Google Scholar]
  19. Huh, M.; Liu, A.; Owens, A.; Efros, A.A. Fighting Fake News: Image Splice Detection via Learned Self-Consistency. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018. [Google Scholar]
  20. Li, W.; Yuan, Y.; Yu, N. Passive detection of doctored JPEG image via block artifact grid extraction. Signal Process. 2009, 89, 1821–1829. [Google Scholar] [CrossRef]
  21. Iakovidou, C.; Zampoglou, M.; Papadopoulos, S.; Kompatsiaris, Y. Content-aware detection of JPEG grid inconsistencies for intuitive image forensics. J. Vis. Commun. Image Represent. 2018, 54, 155–170. [Google Scholar] [CrossRef]
  22. Ye, S.; Sun, Q.; Chang, E.C. Detecting Digital Image Forgeries by Measuring Inconsistencies of Blocking Artifact. In Proceedings of the 2007 IEEE International Conference on Multimedia and Expo, Beijing, China, 2–5 July 2007; pp. 12–15. [Google Scholar] [CrossRef]
  23. Lin, Z.; He, J.; Tang, X.; Tang, C.K. Fast, automatic and fine-grained tampered JPEG image detection via DCT coefficient analysis. Pattern Recognit. 2009, 42, 2492–2501. [Google Scholar] [CrossRef]
  24. Bianchi, T.; Piva, A. Image Forgery Localization via Block-Grained Analysis of JPEG Artifacts. IEEE Trans. Inf. Forensics Secur. 2012, 7, 1003–1017. [Google Scholar] [CrossRef]
  25. Bianchi, T.; De Rosa, A.; Piva, A. Improved DCT coefficient analysis for forgery localization in JPEG images. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Prague, Czech Republic, 22–27 May 2011; pp. 2444–2447. [Google Scholar] [CrossRef]
  26. Barni, M.; Nowroozi, E.; Tondi, B. Higher-order, adversary-aware, double JPEG-detection via selected training on attacked samples. In Proceedings of the 2017 25th European Signal Processing Conference (EUSIPCO), Kos, Greece, 28 August–2 September 2017; pp. 281–285. [Google Scholar] [CrossRef]
  27. Kwon, M.J.; Nam, S.H.; Yu, I.J.; Lee, H.K.; Kim, C. Learning JPEG Compression Artifacts for Image Manipulation Detection and Localization. Int. J. Comput. Vis. 2022, 130, 1875–1895. [Google Scholar] [CrossRef]
  28. Coccomini, D.A.; Caldelli, R.; Falchi, F.; Gennaro, C. On the Generalization of Deep Learning Models in Video Deepfake Detection. J. Imaging 2023, 9, 89. [Google Scholar] [CrossRef]
  29. Novozámský, A.; Mahdian, B.; Saic, S. IMD2020: A Large-Scale Annotated Dataset Tailored for Detecting Manipulated Images. In Proceedings of the 2020 IEEE Winter Applications of Computer Vision Workshops (WACVW), Snowmass, CO, USA, 1–5 March 2020; pp. 71–80. [Google Scholar] [CrossRef]
  30. Le, T.; Nguyen, H.H.; Yamagishi, J.; Echizen, I. OpenForensics: Large-Scale Challenging Dataset For Multi-Face Forgery Detection And Segmentation In-The-Wild. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, QC, Canada, 10–17 October 2021. [Google Scholar]
  31. Zhou, P.; Han, X.; Morariu, V.I.; Davis, L.S. Two-Stream Neural Networks for Tampered Face Detection. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Honolulu, HI, USA, 21–26 July 2017. [Google Scholar]
  32. Fontani, M.; Bianchi, T.; De Rosa, A.; Piva, A.; Barni, M. A Framework for Decision Fusion in Image Forensics Based on Dempster–Shafer Theory of Evidence. IEEE Trans. Inf. Forensics Secur. 2013, 8, 593–607. [Google Scholar] [CrossRef]
  33. de Carvalho, T.J.; Riess, C.; Angelopoulou, E.; Pedrini, H.; de Rezende Rocha, A. Exposing Digital Image Forgeries by Illumination Color Classification. IEEE Trans. Inf. Forensics Secur. 2013, 8, 1182–1194. [Google Scholar] [CrossRef]
  34. Wen, B.; Zhu, Y.; Subramanian, R.; Ng, T.T.; Shen, X.; Winkler, S. Coverage—A Novel Database for Copy-Move Forgery Detection. In Proceedings of the IEEE International Conference on Image processing (ICIP), Phoenix, AZ, USA, 25–28 September 2016; pp. 161–165. [Google Scholar]
  35. Guan, H.; Kozak, M.; Robertson, E.; Lee, Y.; Yates, A.N.; Delgado, A.; Zhou, D.; Kheyrkhah, T.; Smith, J.; Fiscus, J. MFC Datasets: Large-Scale Benchmark Datasets for Media Forensic Challenge Evaluation. In Proceedings of the 2019 IEEE Winter Applications of Computer Vision Workshops (WACVW), Waikoloa, HI, USA, 7–11 January 2019; pp. 63–72. [Google Scholar] [CrossRef]
  36. Goodfellow, I.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative adversarial networks. Commun. ACM 2020, 63, 139–144. [Google Scholar] [CrossRef]
  37. Zhu, J.; Park, T.; Isola, P.; Efros, A.A. Unpaired Image-to-Image Translation using Cycle-Consistent Adversarial Networks. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017. [Google Scholar]
  38. Zeng, Y.; Lin, Z.; Yang, J.; Zhang, J.; Shechtman, E.; Lu, H. High-Resolution Image Inpainting with Iterative Confidence Feedback and Guided Upsampling. In Proceedings of the Computer Vision—ECCV 2020: 16th European Conference, Glasgow, UK, 23–28 August 2020. [Google Scholar]
  39. Vitoria, P.; Raad, L.; Ballester, C. ChromaGAN: An Adversarial Approach for Picture Colorization. arXiv 2019, arXiv:1907.09837. [Google Scholar]
  40. Ledig, C.; Theis, L.; Huszar, F.; Caballero, J.; Aitken, A.P.; Tejani, A.; Totz, J.; Wang, Z.; Shi, W. Photo-Realistic Single Image Super-Resolution Using a Generative Adversarial Network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017. [Google Scholar]
  41. Pang, Y.; Lin, J.; Qin, T.; Chen, Z. Image-to-Image Translation: Methods and Applications. IEEE Trans. Multimed. 2021, 24, 3859–3881. [Google Scholar] [CrossRef]
  42. Isola, P.; Zhu, J.; Zhou, T.; Efros, A.A. Image-to-Image Translation with Conditional Adversarial Networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017. [Google Scholar]
  43. Ronneberger, O.; Fischer, P.; Brox, T. U-Net: Convolutional Networks for Biomedical Image Segmentation. In Proceedings of the Medical Image Computing and Computer-Assisted Intervention—MICCAI 2015: 18th International Conference, Munich, Germany, 5–9 October 2015. [Google Scholar]
  44. Miyato, T.; Kataoka, T.; Koyama, M.; Yoshida, Y. Spectral Normalization for Generative Adversarial Networks. arXiv 2018, arXiv:1802.05957. [Google Scholar]
  45. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An Image is Worth 16 × 16 Words: Transformers for Image Recognition at Scale. arXiv 2020, arXiv:2010.11929. [Google Scholar]
  46. Guillaro, F.; Cozzolino, D.; Sud, A.; Dufour, N.; Verdoliva, L. TruFor: Leveraging All-Round Clues for Trustworthy Image Forgery Detection and Localization. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Vancouver, BC, Canada, 17–24 June 2023; pp. 20606–20615. [Google Scholar]
  47. Wu, H.; Chen, Y.; Zhou, J. Rethinking Image Forgery Detection via Contrastive Learning and Unsupervised Clustering. arXiv 2023, arXiv:2308.09307. [Google Scholar]
Figure 1. The training procedure of the generator.
Figure 1. The training procedure of the generator.
Jimaging 10 00004 g001
Figure 2. The training procedure of the discriminator. The dotted box around the discriminator signifies that the weights of these discriminators are shared.
Figure 2. The training procedure of the discriminator. The dotted box around the discriminator signifies that the weights of these discriminators are shared.
Jimaging 10 00004 g002
Figure 3. Examples of forged images and the corresponding heatmaps by the input forgery localization algorithms and the proposed fusion method.
Figure 3. Examples of forged images and the corresponding heatmaps by the input forgery localization algorithms and the proposed fusion method.
Jimaging 10 00004 g003
Table 1. Selection of input algorithms that are evaluated on their performance and complementarity, and used in the proposed fusion method.
Table 1. Selection of input algorithms that are evaluated on their performance and complementarity, and used in the proposed fusion method.
Input Type
AlgorithmArtifactRGBOther
ADQ1 [23]JPEGDCT analysis
DCT [22]JPEGDCT analysis
BLK [20]JPEG-
CAGI [21]JPEG-
Noiseprint [17]camera-based-
Comprint [11]JPEG-
Comprint + Noiseprint [11]camera-based & JPEG-
CAT-Net [27]JPEG & DCTDCT filter
Table 2. Datasets used for the evaluation of image forgery detection methods and training of the proposed fusion method.
Table 2. Datasets used for the evaluation of image forgery detection methods and training of the proposed fusion method.
Dataset#fake#realFormatNotesUsage
IMD2020 [29]2010414PNG & JPEGVarious forgery typesTraining & Evaluation
OpenForensics [30]18,895N/AJPEGSynthetic face swappingTraining & Evaluation
FaceSwap [31]8791651JPEGFace swaps with FaceSwap-appTraining & Evaluation
VIPP [32]6269JPEGUses double JPEG compressionEvaluation
DSO-1 [33]100100PNGOnly splicingEvaluation
Coverage [34]100100TIFOnly copy-movesEvaluation
NC2016 [35]546874JPEGSplicing, copy-moves and inpaintingEvaluation
Table 3. The average forgery localization performance (F1 score) of every method on every dataset, the theoretical improvement of oracle fusion, the simple fusion methods, and the proposed fusion method. We highlight the three largest values for each column in bold, underline and italics.
Table 3. The average forgery localization performance (F1 score) of every method on every dataset, the theoretical improvement of oracle fusion, the simple fusion methods, and the proposed fusion method. We highlight the three largest values for each column in bold, underline and italics.
AlgorithmF1 Score
VIPPIMD2020DSO-1OpenForensicsFaceSwapCoverageNC2016
CAT-Net [27]0.7170.8500.6750.9480.4490.5730.487
ADQ1 [23]0.5030.2920.4200.4830.2820.2110.206
BLK [20]0.4310.2630.4560.2630.1060.2420.233
CAGI [21]0.4370.2970.5120.2940.1840.2980.293
DCT [22]0.4320.3130.3470.4230.1910.2220.183
Comprint [11]0.4960.2970.7630.6300.3510.3490.398
Comprint + Noiseprint [11]0.5810.4370.8130.7110.4120.3680.439
Noiseprint [17]0.5560.3960.8100.6710.3470.3320.413
Best oracle fusion of 2 methods0.8070.8570.8770.9610.5790.6160.594
Oracle fusion of all methods0.8300.8640.9140.9650.6530.6530.626
Maximum0.7000.7760.8220.9240.4350.4880.514
Minimum0.6730.6370.7930.8970.5580.4780.476
Proposed fusion method0.733-0.753--0.5380.510
Table 4. The average forgery localization performance (F1 score) of every method on every test dataset (i.e., the test split of those datasets), the theoretical improvement of oracle fusion, the simple fusion methods, and the proposed fusion method. We highlight the three largest values for each column in bold, underline and italics. Note, that the training splits of these datasets were used during training of the proposed fusion method.
Table 4. The average forgery localization performance (F1 score) of every method on every test dataset (i.e., the test split of those datasets), the theoretical improvement of oracle fusion, the simple fusion methods, and the proposed fusion method. We highlight the three largest values for each column in bold, underline and italics. Note, that the training splits of these datasets were used during training of the proposed fusion method.
AlgorithmF1 Score
IMD2020 TestOpenForensics TestFaceSwap Test
CAT-Net [27]0.8630.9520.448
ADQ1 [23]0.3400.4790.322
BLK [20]0.2740.2560.134
CAGI [21]0.3160.2900.211
DCT [22]0.3360.4180.220
Comprint [11]0.4240.6260.421
Comprint + Noiseprint [11]0.4570.7100.466
Noiseprint [17]0.4080.6670.409
Best oracle fusion of 2 methods0.8730.9610.584
Oracle fusion of all methods0.8830.9650.647
Maximum0.8060.9260.439
Minimum0.6720.8980.580
Proposed fusion method0.8690.9660.814
Table 5. Evaluation of the detection performance (AUC) of the proposed simple fusion algorithms compared to both the input algorithms and the complete fusion threshold of the input algorithms. We highlight the three largest values for each column in bold, underline and italics.
Table 5. Evaluation of the detection performance (AUC) of the proposed simple fusion algorithms compared to both the input algorithms and the complete fusion threshold of the input algorithms. We highlight the three largest values for each column in bold, underline and italics.
AlgorithmAUC
VIPPIMD2020DSO-1FaceSwapCoverageNC2016
CAT-Net [27]0.8070.9330.7860.6610.6970.758
ADQ1 [23]0.6340.6860.6540.7270.5060.603
BLK [20]0.5920.5530.5790.4960.5080.525
CAGI [21]0.6440.6150.6500.4980.5300.558
DCT [22]0.6490.6110.4100.5740.5050.498
Comprint [11]0.6140.6440.9600.5260.5450.566
Comprint + Noiseprint [11]0.6300.6340.9600.5330.5560.597
Noiseprint [17]0.6080.5430.8610.5170.5340.534
Maximum0.7000.8030.8810.5790.6090.672
Minimum0.7340.7500.7750.7380.6180.696
Proposed fusion method0.790-0.792-0.6710.686
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Mareen, H.; De Neve, L.; Lambert, P.; Van Wallendael, G. Harmonizing Image Forgery Detection & Localization: Fusion of Complementary Approaches. J. Imaging 2024, 10, 4. https://doi.org/10.3390/jimaging10010004

AMA Style

Mareen H, De Neve L, Lambert P, Van Wallendael G. Harmonizing Image Forgery Detection & Localization: Fusion of Complementary Approaches. Journal of Imaging. 2024; 10(1):4. https://doi.org/10.3390/jimaging10010004

Chicago/Turabian Style

Mareen, Hannes, Louis De Neve, Peter Lambert, and Glenn Van Wallendael. 2024. "Harmonizing Image Forgery Detection & Localization: Fusion of Complementary Approaches" Journal of Imaging 10, no. 1: 4. https://doi.org/10.3390/jimaging10010004

APA Style

Mareen, H., De Neve, L., Lambert, P., & Van Wallendael, G. (2024). Harmonizing Image Forgery Detection & Localization: Fusion of Complementary Approaches. Journal of Imaging, 10(1), 4. https://doi.org/10.3390/jimaging10010004

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop