Next Article in Journal
Typical Diagnostic Reference Levels of Radiation Exposure on Neonates Under 1 kg in Mobile Chest Imaging in Incubators
Previous Article in Journal
Concealed Weapon Detection Using Thermal Cameras
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Deepfake Media Forensics: Status and Future Challenges

by
Irene Amerini
1,†,
Mauro Barni
2,†,
Sebastiano Battiato
3,†,
Paolo Bestagini
4,†,
Giulia Boato
5,6,†,
Vittoria Bruni
7,†,
Roberto Caldelli
8,9,†,
Francesco De Natale
5,10,†,
Rocco De Nicola
11,†,
Luca Guarnera
3,*,†,
Sara Mandelli
4,†,
Taiba Majid
1,†,
Gian Luca Marcialis
12,†,
Marco Micheletto
12,†,
Andrea Montibeller
5,†,
Giulia Orrù
12,†,
Alessandro Ortis
3,†,
Pericle Perazzo
13,†,
Giovanni Puglisi
14,†,
Nischay Purnekar
2,†,
Davide Salvi
4,†,
Stefano Tubaro
4,†,
Massimo Villari
15,† and
Domenico Vitulano
7,†
add Show full author list remove Hide full author list
1
Department of Computer, Control and Management Engineering, Sapienza University of Rome, 00185 Roma, Italy
2
Department of Information Engineering and Mathematics, University of Siena, 53100 Siena, Italy
3
Department of Mathematics and Computer Science, University of Catania, 95125 Catania, Italy
4
Department of Electronics, Information and Bioengineering (DEIB), Politecnico di Milano, 20133 Milano, Italy
5
Department of Information Engineering and Computer Science, University of Trento, 38123 Trento, Italy
6
Truebees S.r.l., 20900 Monza, Italy
7
Department of Basic and Applied Sciences for Engineering, Sapienza University of Rome, 00185 Roma, Italy
8
CNIT, National Inter-University Consortium for Telecommunications, 50134 Florence, Italy
9
Department of Engineering and Sciences, Universitas Mercatorum, 00186 Rome, Italy
10
CNIT, University of Trento, 38122 Trento, Italy
11
IMT School for Advanced Studies, 55100 Lucca, Italy
12
Department of Electrical and Electronic Engineering, University of Cagliari, 09123 Cagliari, Italy
13
Department of Information Engineering, University of Pisa, 56122 Pisa, Italy
14
Department of Mathematics and Computer Science, University of Cagliari, 09124 Cagliari, Italy
15
MIFT Department, University of Messina, Viale F. Stagno d’Alcontres, 31, 98166 Messina, Italy
*
Author to whom correspondence should be addressed.
All authors contributed equally to this work.
J. Imaging 2025, 11(3), 73; https://doi.org/10.3390/jimaging11030073
Submission received: 16 January 2025 / Revised: 19 February 2025 / Accepted: 24 February 2025 / Published: 28 February 2025

Abstract

:
The rise of AI-generated synthetic media, or deepfakes, has introduced unprecedented opportunities and challenges across various fields, including entertainment, cybersecurity, and digital communication. Using advanced frameworks such as Generative Adversarial Networks (GANs) and Diffusion Models (DMs), deepfakes are capable of producing highly realistic yet fabricated content, while these advancements enable creative and innovative applications, they also pose severe ethical, social, and security risks due to their potential misuse. The proliferation of deepfakes has triggered phenomena like “Impostor Bias”, a growing skepticism toward the authenticity of multimedia content, further complicating trust in digital interactions. This paper is mainly based on the description of a research project called FF4ALL (FF4ALL-Detection of Deep Fake Media and Life-Long Media Authentication) for the detection and authentication of deepfakes, focusing on areas such as forensic attribution, passive and active authentication, and detection in real-world scenarios. By exploring both the strengths and limitations of current methodologies, we highlight critical research gaps and propose directions for future advancements to ensure media integrity and trustworthiness in an era increasingly dominated by synthetic media.

1. Introduction

The rapid advancement of artificial intelligence (AI) has given rise to a new wave of synthetic media, widely known as deepfakes. These are highly realistic images, audio, and videos generated using sophisticated AI techniques such as Generative Adversarial Networks (GANs) [1] and Diffusion Models (DMs) [2]. While offering unprecedented creative possibilities, these technologies have also raised substantial ethical and security concerns, posing risks in domains such as entertainment, politics, and cybersecurity. Deepfakes are often indistinguishable from authentic media, which has led to their misuse in spreading misinformation, impersonation, and other malicious activities. Recent advancements in deepfake technology have enabled highly realistic synthetic media, leading to a surge in fraud cases. A notable example occurred in 2024, when a deepfake video of Elon Musk (https://www.nytimes.com/interactive/2024/08/14/technology/elon-musk-ai-deepfake-scam.html, last accessed on 17 February 2025) was used to promote a cryptocurrency scam, convincing an 82-year-old retiree to invest 690,000, ultimately resulting in a complete financial loss. In 2024, AI technologies were also used for the cyber fraud. In a recent example, the British engineering firm Arup fell victim to a deepfake scam (https://edition.cnn.com/2024/02/04/asia/deepfake-cfo-scam-hong-kong-intl-hnk/index.html, last accessed on 17 February 2025), where attackers impersonated the company’s Chief Financial Officer and other employees during a video conference. A staff member was deceived into authorizing 15 transactions totaling approximately 25.6 million. Unfortunately, deepfake technology has also introduced new threats to democratic processes. In January 2024, during the New Hampshire Democratic primary, an AI-generated robocall impersonating President Joe Biden (https://www.theguardian.com/us-news/article/2024/may/23/biden-robocall-indicted-primary, Last accessed on 17 February 2025) urged voters to abstain from voting, attempting to manipulate electoral participation. This example underscores the growing risk of AI-driven political disinformation. These cases are among the latest attacks leveraging generative AI technology, emphasizing the need for effective and robust detection mechanisms.
One profound consequence of the proliferation of Deepfakes is the emergence of a cognitive phenomenon termed “Impostor Bias” [3]. This bias reflects the tendency of individuals to question the authenticity of multimedia elements, driven by the awareness of AI’s ability to generate realistic forgeries. This skepticism, while protective in some cases, also undermines trust in legitimate media and digital interactions.
To counter these challenges, the field of deepfake detection has gained significant traction. Researchers have developed methods [4] to identify inconsistencies and artifacts imperceptible to the human eye but detectable using machine learning algorithms. These approaches, leveraging Convolutional Neural Networks (CNNs) [5,6], focus on analyzing spatial and temporal anomalies, ranging from pixel-level distortions to inconsistencies in biological or behavioral cues. As deepfake generation techniques evolve, so do detection strategies, creating a dynamic and continuous arms race.
The scope of forensic analysis in the deepfake domain extends beyond detection, encompassing several critical research areas. These include Deepfake Attribution and Recognition, Passive and Active Authentication, and methods designed to operate effectively in realistic, real-world scenarios.
  • Deepfake Attribution and Recognition aims to trace the origins of synthetic content by identifying the specific models used in its creation. This involves analyzing the traces left by generative architectures, enabling the attribution of content to its source.
  • Passive Authentication Methods focus on evaluating the authenticity of media through inherent characteristics, such as statistical irregularities, without requiring additional embedded data. These methods are particularly effective for retrospective analysis.
  • Active Authentication techniques, in contrast, embed verifiable information into media during its creation, such as digital watermarks or cryptographic signatures, allowing for immediate and robust verification.
  • Realistic Scenario Detection addresses the challenges posed by deepfake content in uncontrolled environments, such as low-resolution, compressed, or adversarially manipulated media. This is critical for practical applications where detection systems must operate under diverse and unpredictable conditions.
In this context, the authors of this paper are involved in the FF4ALL initiative (FF4ALL-Detection of Deep Fake Media and Life-Long Media Authentication) (research project line in the context of Spoke 2 “Misinformation and Fakes” funded by the Italian national PNRR and European NextGenerationEU funds), which aims to develop theoretical and practical tools for detecting and combating media counterfeits or deepfakes, tracing their origin and limiting their dissemination. This will be achieved through passive analysis techniques that operate when the content is used or disseminated and active protection methods to be adopted at the time of content creation to facilitate subsequent authentication. Table 1 summarizes the main topics of interest and research techniques on deepfakes. The project is financed within the National Recovery and Resilience Plan, Mission 4 “Education and Research”—financed by the European Union—NextGenerationEU. Further information on the project is available on the website: https://sites.unica.it/ff4all/, accessed on 15 February 2025. The project is structured into multiple Work Packages (WPs), each focusing on specific tasks related to deepfake detection and authentication:
  • WP1—Deepfake Attribution and Recognition
    -
    Task 1.1—Deepfake Fingerprint;
    -
    Task 1.2—Deepfake Attribution;
  • WP2—Passive Deepfake Authentication Methods
    -
    Task 2.1—Deepfake and Biometric Recognition
    -
    Task 2.2—Audio–Video Deepfake;
    -
    Task 2.3—Advanced Methods for Deepfake Detection;
  • WP3—Deepfake Detection Methods in Realistic Scenarios
    -
    Task 3.1—Deepfake Detection of image-videos in the Wild
    -
    Task 3.2—Deepfake and Social Media;
    -
    Task 3.3—Detection of Deepfake Images and Videos in Adversarial Settings;
  • WP4—Active Authentication
    -
    Task 4.1—Active Fingerprinting for Deepfake Detection and Attribution
    -
    Task 4.2—Authentication of Devices for the Acquisition and Processing of Content;
    -
    Task 4.3—Trusted Remote Media Processing on Cloud and Edge Computing Systems
Building upon our previous work [7], this extended version provides a structured comparison of deepfake detection approaches, analyzing their strengths and weaknesses. Additionally, it introduces a more detailed discussion of the deepfake generation process, the methodology behind dataset creation, and an improved presentation of the FF4ALL project’s objectives. Furthermore, this version expands the discussion on open challenges in deepfake forensics, identifying key research gaps and future directions for improving detection robustness, and forensic applicability. In addition, unlike previous studies that focus on isolated aspects of deepfake detection, this work provides an integrated perspective on forensic challenges, real-world constraints, and adaptive methodologies.
In particular, Section 2 provides a brief introduction to the deepfake generation process. Section 3 categorizes and examines current deepfake detection algorithms, providing a detailed analysis of their advantages and disadvantages, while Section 4 introduces the most used deepfake datasets in state of the art. However, as deepfake techniques continue to advance, many of the existing methods struggle to adapt to new forms of manipulation, such as domain shifts or previously unknown attack patterns. To address this limitation, Section 5 investigates continuous learning techniques, which enable detection models to evolve alongside deepfake technologies and adapt to emerging threats.
Although continual learning enhances the adaptability of detection models, it also introduces increased complexity, raising critical questions about the interpretability of these systems. In forensic investigations, explainability is crucial for ensuring the reliability, trustworthiness, and accountability of AI-driven detection tools. Building on this, Section 6 explores methods to improve the interpretability of detection models, ensuring that these tools can be understood and validated by forensic experts.
Beyond the detection of manipulations, however, tracing the origins of synthetic media is equally critical for holding perpetrators accountable and understanding the generative processes behind deepfakes. Section 7 extends the discussion by delving into deepfake attribution and model fingerprinting techniques, which aim to identify the sources of synthetic media and the generative models responsible for their creation.
To provide a broader perspective, Section 8 examines passive deepfake authentication methods, which emphasize non-intrusive approaches to verifying media authenticity. Section 9 focuses on the practical challenges of detecting manipulated content in realistic scenarios, such as highly compressed media commonly found on social networks. Section 10 complements these discussions by addressing active authentication techniques, such as cryptographic signatures, which proactively embed trust and ensure media authenticity at the point of creation.
A general discussion is provided in Section 11, synthesizing the insights from all methodologies and exploring their interconnections. Finally, Section 12 evaluates the strengths and limitations of the methodologies discussed, outlines future directions, and provides a comprehensive perspective on advancements in deepfake media forensics.

2. Deepfake Generation Process

Generative models are a class of machine learning algorithms designed to learn the underlying distribution of a given dataset and generate new samples that resemble the real data. Unlike discriminative models, which focus on distinguishing between different classes or making predictions, generative models aim to capture the data distribution p d a t a and synthesize new samples that share similar characteristics. These models have gained significant attention in various domains, including image synthesis, text generation, music composition, and data augmentation. Two of the most prominent approaches in generative modeling are Generative Adversarial Networks (GANs) and Diffusion Models (DMs). While both methodologies aim at producing high-quality synthetic data, they employ fundamentally different principles and learning paradigms. GANs rely on an adversarial game between two neural networks, while diffusion models leverage a sequential denoising process to learn the data distribution. Some aspects of this framework will be examined in the next subsection.
In general, deepfake creation involves a multi-step pipeline that leverages advanced machine learning techniques to synthesize highly realistic media. Figure 1 shows a generic deepfake creation process, which begins with input data collection, in which source and target videos or images are collected. This is followed by preprocessing, which includes face detection, alignment, and normalization to ensure that the data is suitable for training. The model training phase employs neural network architectures such as autoencoders and GANs (Generative Adversarial Networks) to learn and replicate facial features and expressions. Once the model is trained, the deepfake generation step involves face swapping, blending, and post-processing to produce a seamless and realistic deepfake. The final output is a deepfake image, video, or audio. The process concludes with evaluation, where the quality of the deepfake is assessed, and any artifacts or inconsistencies are identified. This structured workflow highlights the key stages and techniques involved in deepfake creation, providing a comprehensive understanding of the process.

2.1. Generative Adversarial Networks (GANs)

The GAN framework [1] consists of two neural networks trained in a competitive manner (Figure 2a):
The Generator ( G ) , which takes a random noise vector z sampled from a prior distribution (e.g., Gaussian or uniform) and maps it to a data sample x ¯ . The objective of G is to learn the underlying distribution of the training data and produce samples that are indistinguishable from real ones. The Discriminator ( D ) , which acts as a binary classifier, distinguishing between real data from the training set and synthetic data generated by G. Mathematically, the objective function of a GAN can be expressed as a minimax game (Equation (1)):
min G max D V ( D , G ) = E x p d a t a ( x ) [ log D ( x ) ] + E z p z ( z ) [ log ( 1 D ( G ( z ) ) ) ]
where p d a t a ( x ) represents the real data distribution, and p z ( z ) is the prior distribution used to sample the input noise for G. The training process involves an iterative optimization where:
  • D is optimized to maximize its ability to differentiate between real and fake samples.
  • G is optimized to minimize the ability of D to correctly classify generated samples as fake.
Through this adversarial process, the generator progressively improves, producing increasingly realistic samples. GANs have been widely applied in tasks such as image generation (e.g., StyleGAN, BigGAN), super-resolution, and domain adaptation. However, they often suffer from challenges like mode collapse, where G generates only a limited variety of samples, and training instability due to the adversarial nature of optimization.

2.2. Diffusion Models

Diffusion Models (DMs) [8] represent a fundamentally different generative paradigm, inspired by nonequilibrium thermodynamics and probabilistic modeling (Figure 2b). They operate by progressively corrupting the input data through a forward process and then learning to reverse this degradation to reconstruct samples from noise.
The forward process is defined as a fixed Markov chain that gradually adds Gaussian noise to the data (Equation (2)):
q ( x t | x t 1 ) for t = 1 , , T
where x 1 , , x T represent latent variables with the same dimensionality as the original data x 0 . This process eventually transforms the input data into a nearly isotropic Gaussian distribution.
The training objective of diffusion models is to learn the reverse process (Equation (3)):
p θ ( x t 1 | x t ) for t = 1 , , T
which enables the model to generate new samples by starting from pure noise and gradually refining the data through a learned denoising process.
Unlike GANs, which rely on adversarial training, diffusion models optimize a likelihood-based objective, often implemented through denoising score matching or variational inference. This makes them more stable to train and less prone to mode collapse. Recent advances in diffusion models, such as Denoising Diffusion Probabilistic Models (DDPMs) and Score-Based Generative Models, have demonstrated state-of-the-art performance in image synthesis, rivaling and sometimes surpassing GANs in quality and diversity.

3. Deepfake Detection

Ensuring the authenticity of digital content is a critical challenge in multimedia forensics as deepfake technology continues to evolve and produce increasingly realistic synthetic media. Detecting manipulated content is essential to mitigate the risks of misinformation, identity fraud, and media integrity threats [9] while also serving as the foundation for forensic analysis, attribution, and authentication. This section provides a structured overview of detection methodologies, ranging from handcrafted forensic techniques to modern data-driven approaches. Early forensic approaches focus on detecting explicit spatial inconsistencies, such as unnatural lighting, pixel-level anomalies, and compression artifacts [10]. While these methods remain valuable, they often struggle against highly sophisticated manipulations. To overcome these limitations, deep learning models exploit data-driven patterns, learning to recognize subtle inconsistencies imperceptible to the human eye [11]. A comprehensive exploration of these methodologies is presented below, highlighting their principles, technical approaches, limits, and contributions to advancing the field of deepfake detection (Figure 3).

3.1. Undirected Approaches

Undirected approaches harness the adaptability of deep learning models to autonomously learn features from data rather than relying on predefined artifacts. This capability is critical for addressing novel manipulation techniques, which continue to emerge in the evolving landscape of deepfake technology. Deep learning-based techniques often rely on architectures such as Convolutional Neural Networks (CNNs) and autoencoders, which are well-suited for extracting features from spatial and temporal domains [12]. Adaptability in detecting manipulations stems from the ability of these methods to learn hierarchical representations directly from data. Central to this process are latent representations, compact and structured spaces generated during training that capture high-level patterns often imperceptible in raw data. Leveraging these latent spaces enables deep learning models to achieve greater flexibility compared to manually designed features, improving generalization to novel and unseen manipulations. However, their effectiveness heavily depends on their ability to capture the variability of forgery types comprehensively. When the latent space is insufficiently representative, models struggle to generalize beyond the training data. Yan et al. [13] addressed this limitation through Latent Space Data Augmentation (LSDA), a technique designed to enrich the latent space by interpolating diverse samples, thereby enhancing the generalization capabilities of deep learning-based detectors. Another limitation of latent space-based methods is their high computational demands and reliance on extensive labeled datasets, which are resource-intensive and time-consuming to develop [14]. Transfer learning offers a potential solution by enabling models to adapt pre-trained latent representations to new tasks with a reduced need for extensive and fully annotated datasets [15].

3.2. Visual Artifact-Based Detection

The process of creating deepfake content frequently introduces subtle anomalies within the visual data, often referred to as artifacts, such as irregular pixel arrangements, distortions along edges, or inconsistencies in the spectral domain (Figure 4). Methods that focus on visual artifact detection leverage these anomalies as indicators of synthetic manipulation.
Spatial artifacts are one of the earliest and most intuitive indicators of manipulated media. They include unnatural pixel formations, edge inconsistencies, lighting mismatches, and shadow anomalies that arise due to imperfections in generative algorithms [16]. For instance, deepfake algorithms often struggle to maintain consistent lighting across a manipulated face, resulting in subtle but detectable anomalies around critical regions such as the eyes, mouth, and face contours. Frequency domain analysis provides another robust framework for detecting deepfakes by identifying anomalies in the spectral characteristics of images and videos, which are often imperceptible in the spatial domain [17]. Methods based on Fourier transforms and wavelet decomposition have proven effective in uncovering these inconsistencies, leveraging the distinct frequency patterns observed in synthetic media [18,19]. The ability to isolate and analyze frequency-specific features enhances the detection of manipulations, even under challenging conditions such as compression or resolution loss [20].
The ability to isolate and analyze frequency-specific features enhances the detection of manipulations, even under challenging conditions such as compression or resolution loss [20], which are common in deepfake content shared on social media. In fact, lossy compression algorithms, such as JPEG and MPEG, obscure subtle visual cues essential for detection, complicating the identification of deepfakes. At the same time, these compression processes introduce distinctive patterns that specialized models can exploit to reveal discrepancies [21]. For instance, Gao et al. (2024) proposed a High-Frequency Enhancement (HiFE) network that leverages the Discrete Cosine Transform (DCT) and the Discrete Wavelet Transform (DWT) to recover high-frequency details adaptively lost during compression, significantly improving detection performance on highly compressed deepfake content [22].
Lastly, GAN-specific artifacts focus on detecting traces unique to generative adversarial networks. For example, GANs often produce checkerboard patterns or exhibit anomalies in low-frequency regions, which can serve as reliable indicators of manipulation [23]. Although visual artifact-based methods are computationally efficient and effective in many scenarios, their reliance on spatial details may still pose challenges. Integrating frequency-specific analysis could offer a way to overcome these limitations, improving detection robustness against compression and resolution loss.

3.3. Biological Artifact-Based Detection

Biological artifact-based detection methods exploit inconsistencies in human physiology and behavior, aspects that deepfake algorithms struggle to replicate accurately. Facial regions are particularly suitable for this type of analysis due to subtle physiological and behavioral cues [24]. Physiological signals, such as variations in skin tone caused by blood flow or subtle muscle movements, often appear inaccurate or entirely absent in deepfake content. Techniques that analyze these signals, such as remote photoplethysmography (rPPG), can identify deviations from natural biological patterns by examining heart rate dynamics captured in facial videos [25]. Behavioral cues, including eye-blinking frequency, gaze direction, and lip synchronization, provide another layer of analysis. Deepfake algorithms frequently fail to maintain the natural temporal dynamics of these behaviors, leading to detectable anomalies. For instance, inconsistencies in the display of emotions, measured through valence and arousal, have been shown to indicate synthetic content [26]. Similarly, irregular or unnatural blinking frequency in deepfake videos represents a well-documented artifact, as generative models often fail to replicate natural blinking patterns consistently [27]. Facial landmarks also play a critical role in detecting manipulations. Methods that analyze facial features’ geometric relationships and alignments can identify discrepancies typical of face-swapping techniques. Deviations in the proportions or placements of key landmarks serve as reliable indicators of forgery [23]. Biological artifact-based approaches are highly effective because they target features deeply rooted in natural human physiology and behavior, which are difficult for even the most advanced generative models to replicate. However, these methods often require high-resolution input and can struggle when faces are partially occluded or poorly lit.

3.4. Texture and Spatio-Temporal Consistency-Based Detection

Texture and spatio-temporal consistency-based detection focus on fine-grained patterns and temporal coherence within media content. These approaches are particularly effective for video Deepfakes, where the dynamic nature of the content poses additional challenges for manipulation algorithms. Intra-face texture analysis examines inconsistencies across different regions of the same face [28]. For instance, face-swapping manipulations often blend skin textures from disparate sources, creating visible mismatches between inner and outer facial regions. Advanced image processing and deep learning techniques can effectively detect these discrepancies [29]. Temporal texture evolution extends this analysis by evaluating texture consistency over successive video frames [30]. Authentic videos typically display smooth and continuous texture transitions, while manipulated content frequently exhibits abrupt or unnatural changes that expose synthetic origins [31]. Spatio-temporal coherence methods assess the synchronization between various modalities, such as facial movements, gestures, and audio [32]. Discrepancies between speech and lip movements or misalignment between gestures and facial expressions are reliable forgery indicators. Combining these techniques provides a robust framework for identifying even sophisticated video Deepfakes. However, computational intensity and the need for high-quality video input can limit their practicality in real-world applications.

4. State-of-the-Art Deepfake Datasets

In this section, we present an overview of the most widely used datasets in state-of-the-art research on deepfake detection. These datasets span various modalities, including image, audio, video, and multimodal data, reflecting the multifaceted nature of the deepfake detection problem. The datasets vary in composition, with some containing exclusively synthetic data, others comprising only real data, and some offering a combination of both. The choice of dataset depends on the specific objectives of the analysis, as each corpus can be utilized for training, testing, or evaluating the generalization capabilities of deepfake detection models. Cross-dataset evaluation is particularly critical, as it enables the assessment of model robustness and generalization across diverse data sources, a key requirement in multimedia forensics. This ensures that detectors perform reliably not only on data similar to the training set but also on entirely different datasets.

4.1. Image Datasets

WildFake [33] is a large-scale dataset designed to facilitate the detection of AI-generated images. It comprises a diverse collection of fake images sourced from open communities and multiple generative models, ensuring a rich variety of content and styles. The dataset is hierarchically organized, categorizing images based on generative model types, architectures, weights, versions, and release timelines. With over 3.6 million images (2.6 M fake and 1 M real), WildFake provides a comprehensive resource for training and evaluation. Real images are curated from various open datasets to ensure realistic comparisons.
Artifact [34] is a dataset designed to evaluate the generalizability and robustness of synthetic image detectors. It includes a diverse range of images across multiple object categories, such as human faces, animals, vehicles, places, and artworks. The dataset contains over 2.4 million images, with 1.5 million synthetic images generated using 25 different methods, including GANs, diffusion models, and other generative techniques. To simulate real-world conditions, the dataset incorporates social media-related impairments like compression, resizing, and cropping. Artifact introduces a novel multi-class classification approach, distinguishing between real images, fake images from known generators, and fake images from unseen generators. Additionally, it employs a Filter Stride Reduction (FSR) technique to preserve generator artifacts despite image degradation.
DIFF [35] is a large-scale dataset specifically designed for detecting diffusion-generated facial forgeries. It contains over 500,000 images synthesized using 13 state-of-the-art diffusion models under four conditions: Text-to-Image (T2I), Image-to-Image (I2I), Face Swapping (FS), and Face Editing (FE). The dataset is curated using 30,000 textual and visual prompts to ensure high-fidelity and semantically consistent forgeries. Real images from 1070 celebrity identities are included to ensure diversity across gender and age groups. Experiments reveal that both human and automated detectors struggle with these forgeries, underscoring the challenges in detection.
StyleGAN-XL [36] is an extension of the StyleGAN family, designed to generate high-quality images on diverse and unstructured datasets, particularly ImageNet. Unlike its predecessors, StyleGAN [37] and StyleGAN2 [38], which excelled at synthesizing human faces and structured datasets, StyleGAN-XL incorporates Projected GAN training and a progressive growing strategy to enhance stability and scalability. The model achieves state-of-the-art performance in large-scale image synthesis, generating high-resolution images up to 1024 × 1024. It also introduces classifier guidance for improved class-conditional generation. StyleGAN-XL overcomes the scalability issues of previous models, offering better sample diversity, generalization, and image fidelity. The dataset provides new benchmarks for GAN-based image generation and detection, positioning it as a strong alternative to diffusion models.

4.2. Audio Datasets

ASVspoof 2019 [39,40] is a benchmark dataset for synthetic speech detection, containing both real and synthetic audio tracks from 78 speakers (33 male, 45 female) based on the VCTK corpus. The dataset was released for a challenge focused on Automatic Speaker Verification (ASV). Its Logical Access (LA) partition contains deepfake data, which is divided into training, development, and evaluation subsets. The training and development subsets include synthetic speech generated using six algorithms, while the evaluation subset includes 13 techniques, only two of which overlap with the training set. This structure allows for open-set evaluation, assessing detectors on unseen synthesis algorithms. The dataset is highly imbalanced, with synthetic data significantly outnumbering real data, a factor that must be considered during model training.
ASVspoof 2021 [41] is an updated version of ASVspoof 2019, released for a subsequent challenge. It includes three evaluation partitions (LA, PA, and DF), with no new training or development data. The deepfake (DF) partition introduces distortions through lossy codecs commonly used in media storage, simulating real-world conditions.
TIMIT-TTS [42] is a synthetic speech dataset generated using 12 different TTS methods, based on the VidTIMIT corpus. It includes four partitions corresponding to different post-processing pipelines: clean, augmented, Dynamic Time Warping (DTW), and DTW+augmented. Each partition contains 19,780 tracks, totaling nearly 80,000 synthetic speech signals. This dataset can be used independently or combined with VidTIMIT and DeepfakeTIMIT for multimodal research.
AISEC “In-the-Wild” [43] is a synthetic speech dataset designed to replicate real-world conditions. It includes 38 h of audio clips (17 h fake, 21 h real) featuring English-speaking celebrities and politicians. The fake clips are derived from publicly available deepfake content, while real clips are manually collected from sources like podcasts and speeches.
FakeOrReal [44] is a speech dataset containing over 198,000 utterances, including both real and synthetic tracks. The synthetic data are generated using TTS methods, while real data are sourced from open datasets and platforms like TED Talks and YouTube.
Purdue Speech Dataset [45] includes 25,000 synthetic speech tracks generated using five advanced diffusion model-based voice cloning methods (ProDiff, DiffGAN-TTS, ElevenLabs, UnitSpeech, and XTTS). Real speech data are sourced from LJSpeech [46] and LibriSpeech [47].

4.3. Video and Multimodal Datasets

DeepfakeTIMIT [48] is a video deepfake dataset derived from the VidTIMIT corpus [49], which consists of real audio–video recordings of 43 individuals, each reciting short sentences, resulting in a total of 430 videos. The DeepfakeTIMIT dataset focuses exclusively on visual forgeries, leaving the audio component unaltered. The fake video frames are generated using a GAN-based approach adapted from Faceswap [50]. The dataset includes deepfakes of 32 subjects, available in two versions: Low Quality (LQ) and High Quality (HQ), differing in frame resolution. In total, the dataset comprises 640 videos, with 320 videos in each quality category.
FaceForensics++ [51] is a widely used visual-only deepfake dataset containing 5000 videos generated using four distinct deepfake generation methods. The dataset is based on 1000 real YouTube videos and includes two partitions with different levels of compression, quantified by the Quantization Parameter (QP). Specifically, the partitions correspond to QP values of 23 and 40, where a higher QP indicates lower video quality.
DFDC [52] (DeepFake Detection Challenge) is a multimodal deepfake dataset comprising nearly 120,000 videos, of which 100,000 are labeled as fake and the remaining as real. The videos are organized into 50 folders, each containing a set of real videos and their corresponding fake derivatives. Each folder features videos of different subjects. While the dataset does not provide separate labels for the authenticity of individual modalities (audio or video), a video is labeled as fake if either the audio or video component is manipulated. Although the majority of the videos are visual-only forgeries, a subset of videos in folders 45 to 49 includes falsified audio in addition to potential visual manipulations.
FakeAVCeleb [53] is a multimodal deepfake dataset constructed from 500 real videos extracted from the VoxCeleb2 corpus [54]. These real videos serve as the basis for generating approximately 20,000 deepfake videos using various deepfake generation techniques. The subjects in the videos are English speakers with diverse ethnicities and accents. The deepfake video frames are created using Faceswap [50] and FSGAN [55], while the synthetic audio is generated using Real-Time Voice Cloning (RTVC) [56]. To ensure synchronization between the video frames and audio, Wav2Lip [57] is applied. The dataset is divided into four distinct partitions: Real Video–Real Audio, Fake Video–Real Audio, Real Video–Fake Audio, and Fake Video–Fake Audio, enabling comprehensive studies on multimodal deepfake detection.
DeepSpeak 1.0 [58] is a large-scale dataset featuring real and deepfake footage of individuals speaking and gesturing in front of webcams. The dataset includes 17 h of real video footage from 220 diverse individuals and over 26 h of deepfake footage generated using state-of-the-art face-swap and lip-sync techniques. The deepfake videos incorporate both natural and AI-generated voices, providing a realistic representation of advanced deepfake technologies. The authors anticipate releasing future versions of the dataset, incorporating updated deepfake methods to reflect evolving technological advancements.

5. Continual Learning for Deepfake Detection

Deep learning models face significant challenges in handling dynamically evolving data distributions, particularly in areas such as audio deepfake detection. Traditional deep learning approaches assume that training and testing data share a static distribution, which is unrealistic in scenarios where new spoofing attacks emerge regularly [59]. Consequently, these models fail to generalize to unseen data and are highly susceptible to catastrophic forgetting when fine-tuned on new tasks [60]. Catastrophic forgetting occurs because gradient-based optimization overwrites parameters optimized for earlier tasks, thereby degrading the model’s performance on those tasks.
Continual learning (CL) provides a framework to mitigate this issue by enabling models to learn new tasks sequentially without forgetting previously learned ones. The stability–plasticity dilemma is central to CL, as it balances the model’s ability to retain past knowledge (stability) while acquiring new information (plasticity) [61,62]. This balance ensures that the model can adapt to evolving tasks while maintaining performance on earlier ones, making CL essential for dynamic learning scenarios. To achieve this balance, CL leverages strategies such as regularization-based approaches, experience replay, and architectural modifications, ensuring models remain adaptable while mitigating forgetting.

5.1. Replay-Based Methods

Replay-based methods are among the most effective strategies for continual learning, particularly in scenarios with evolving tasks and data distributions. These methods mitigate catastrophic forgetting through the use of a memory buffer that stores past samples, which are revisited periodically during training on new tasks [63]. Incorporating historical data reinforces prior knowledge while allowing the model to adapt to new information. Replay methods vary in their implementation, with some utilizing exact sample storage (e.g., Experience Replay) [64] and others employing generative models to recreate past samples (e.g., Generative Replay) [65]. Selective sampling techniques are often integrated to optimize memory usage, ensuring that samples contributing the most to knowledge retention are prioritized. In audio deepfake detection, replay-based methods enhance the model’s ability to detect both older and emerging deepfake attacks. Reintroducing samples of earlier spoofing techniques during training allows the model to generalize across diverse spoofing patterns and maintain high detection accuracy across evolving scenarios. This strategy ensures a balanced representation of past tasks, supporting robust generalization and effective adaptation in dynamic environments.
Ma et al. [66] introduced the replay based method known as Detecting Fake Without Forgetting (DFWF) method, a robust framework designed to mitigate catastrophic forgetting and enable efficient audio deepfake detection. DFWF integrates Learning without Forgetting (LwF) and Positive Sample Alignment (PSA) to preserve past knowledge while incrementally adapting to new spoofing types. LwF employs a knowledge distillation loss to align the output probabilities of the current and original models, ensuring the retention of prior knowledge and the model’s ability to recognize previously encountered spoofing attacks. PSA focuses on maintaining consistent representations of genuine audio by aligning embeddings across tasks using a cosine similarity loss, enabling the model to effectively differentiate genuine audio from new spoofing attacks. These components are combined into a composite loss function that balances knowledge retention and adaptation. DFWF achieved an Average Equal Error Rate (AvgEER) reduction of 81.63% compared to fine-tuning, maintaining robust performance on both new and old tasks of ASVspoof 2019. Nguyen et al. [67] proposed a Transformer-based model with a continual learning method for incremental training, focusing on unseen spoofing attacks. The approach employs few-shot learning and embedding similarity loss to adapt to new tasks with minimal labeled data. Using the ASVspoof datasets and over 2 million synthetic samples, the model achieved 90–95% AUC on unseen datasets, demonstrating strong generalization capabilities. By combining few-shot learning with replay strategies, this method offers a robust and scalable approach for tackling evolving audio deepfake threats.

5.2. Regularization-Based Methods

Regularization-based continual learning methods mitigate catastrophic forgetting by adding constraints to the loss function, ensuring that updates to the model parameters do not interfere significantly with knowledge learned from previous tasks [68]. These techniques identify and preserve critical parameters through mechanisms such as Elastic Weight Consolidation (EWC) or knowledge distillation, which penalize large deviations in key weights or align output distributions between tasks. This approach allows the model to retain prior knowledge while adapting to new data, balancing stability and plasticity. Dong et al. [69] introduced the Continual Audio Defense Enhancer (CADE) built by integrating regularization techniques with replay strategies to enhance robustness in audio deepfake classification. CADE employs Knowledge Distillation Loss, which ensures that the current model’s predictions remain aligned with its previous outputs, and Positive Sample Alignment (PSA) loss, which maintains consistent embeddings of genuine audio across tasks. Unlike purely replay-based methods, CADE leverages regularization to further mitigate forgetting and improve performance. Evaluations conducted on the ASVspoof2019 dataset demonstrate that CADE achieves lower EERs compared to baseline methods, effectively retaining knowledge of previously seen spoofing attacks while adapting to new ones. This hybrid approach highlights the strength of combining replay and regularization for balanced and scalable continual learning in dynamic environments.

5.3. Architectural-Based Methods

Architectural strategies dynamically modify the structure of neural networks to balance the retention of old knowledge and the acquisition of new information. Methods such as Orthogonal Weight Modification (OWM) prevent interference with previously learned tasks by projecting gradient updates onto an orthogonal subspace, ensuring minimal disruption to past knowledge [70]. This type of technique enhance adaptability and robustness, particularly for sequential learning tasks like audio deepfake detection, by maintaining performance on previously learned tasks.
The authors of [71] developed the Radian Weight Modification (RWM) method, a novel architectural approach for continual learning that categorizes classes based on feature distribution similarity. Compact classes (e.g., genuine audio) are treated with minimal gradient updates to retain stability, while dispersed classes (e.g., fake audio) receive orthogonal updates to prevent interference with prior knowledge. The method also incorporates a self-attention mechanism to dynamically adjust gradient directions, ensuring efficient retention of previous tasks and adaptation to new data. Unlike the RWM method, which focuses on radial modifications of gradient directions, the authors of [72] introduced Dynamic Class Rebalancing (DCR) as part of a framework designed to handle imbalanced class distributions dynamically. The approach integrates SincNet as a teacher model and LightCNN as a student model, combining Feature Distillation (FD) with DCR. FD transfers discriminative features from SincNet to LightCNN, ensuring robust feature representation in the student model, while DCR categorizes classes based on feature similarity and adjusts learning strategies dynamically. This allows for enhanced adaptability to evolving data while retaining prior knowledge. The authors of [73] introduced the EVDA benchmark, designed to evaluate various continual learning methods for audio deepfake detection across diverse datasets and conditions. The benchmark includes datasets such as ASVspoof2015, ASVspoof2019, and FoR, alongside new challenges like real-world scenarios and cross-lingual datasets. Key continual learning methods evaluated include OWM, RWM, and Regularized Adaptive Weight Modification (RAWM). Among these, OWM dynamically adjusts weight updates to minimize interference with previously learned tasks, while RWM optimizes gradients for stability across evolving datasets. The results show that replay-based and architectural methods consistently outperform baseline approaches in maintaining low EER and adapting to new tasks, demonstrating the benchmark’s utility for advancing robust and adaptive audio deepfake detection.
Continual learning approaches provide a robust framework for addressing the dynamic and evolving nature of audio deepfake detection. By leveraging strategies such as replay, regularization, and architectural modifications, these methods effectively mitigate catastrophic forgetting while enabling models to adapt to unseen spoofing attacks. Replay-based methods, like DFWF and Transformer-based models, enhance generalization by revisiting past knowledge and incorporating incremental learning for emerging threats. Regularization techniques, exemplified by CADE, combine knowledge distillation and embedding alignment to preserve stability across tasks while retaining flexibility for new data. Architectural strategies, such as RWM and DCR, dynamically adjust learning pathways, offering adaptability to imbalanced and complex distributions.
Future research in continual learning for audio deepfake detection could explore hybrid frameworks that integrate multiple strategies, such as combining replay with advanced architectural modifications to enhance scalability or applying continual learning techniques only to limited portions of the models [74]. The use of generative models for creating realistic past samples and novel benchmarks, like EVDA, can push the boundaries of evaluating and improving performance in cross-lingual scenarios. Additionally, integrating self-supervised learning for feature extraction and multi-modal data, including audio and visual signals, could further enhance robustness against sophisticated deepfake techniques.

6. Explainability and Interpretability in AI Forensics

While continual learning improves the adaptability of detection models, it also adds complexity, posing significant challenges regarding their interpretability. In forensic investigations, ensuring explainability is essential to maintaining the reliability, credibility, and accountability of deepfake detection systems. In fact, deepfake detection systems must provide clear and interpretable decision processes to be admissible in legal and high-stakes scenarios. Explainable and interpretable AI is therefore not merely a desirable feature, it is essential. The following points illustrate why explainability and interpretability are critical in forensic contexts:
  • Enhanced Trust and Credibility: Forensic evidence underpins legal and judicial decisions, making transparency a non-negotiable requirement. When deepfake detection systems clearly reveal which features, audio segments or images parts contributed to their predictions, forensic experts, legal practitioners, and jurors can better understand and trust the outcomes [75]. This transparency builds confidence in the system and supports the use of its outputs as credible evidence in court.
  • Accountability and Legal Defensibility: In the courtroom, every piece of digital evidence must be defensible under rigorous scrutiny. Black-box models, despite their high predictive performance, often lack the necessary transparency [76]. Explainable AI provides a detailed trace of the decision-making process, allowing forensic analysts to demonstrate exactly how a conclusion was reached [77]. This level of traceability is critical for ensuring that evidence can be defended during cross-examinations and for establishing clear lines of accountability.
  • Identification of Biases and Error Sources: Forensic applications demand the highest standards of accuracy and fairness. By visualizing the specific features or segments that influence model predictions, explainable AI techniques enable analysts to identify potential biases or sources of error [78]. This scrutiny is essential not only for improving the reliability of detection systems but also for ensuring that the evidence is free from hidden biases that could undermine a case.
  • Facilitating Expert Collaboration and Continuous Improvement: Effective forensic analysis is inherently multidisciplinary, involving experts from fields such as audio engineering, computer vision experts, computer science, and law [79]. Interpretable models provide a common language for these experts by clearly explaining the inner workings of the detection system [80]. This shared understanding is vital for ongoing refinement and adaptation of the technology, particularly as new deepfake generation techniques emerge.

6.1. Interpretability and Explainability in Audio Deepfake Detection

Audio deepfake detection has gained significant attention in recent years, driven by the rapid advancements in synthetic audio generation technologies. While detection methods have evolved in accuracy and robustness, their lack of interpretability and explainability limits their broader adoption in high-stakes applications such as forensic analysis, media authentication, and legal proceedings [63]. Explainable Artificial Intelligence (XAI) addresses this gap by making AI models more transparent, enabling users to understand how decisions are made, which features are critical, and why certain predictions are trusted. In the context of audio deepfake detection, interpretability and explainability are crucial not only for technical refinement but also for ensuring trust, accountability, and informed decision-making in forensic applications. In forensic investigations, every decision must be both reproducible and legally defensible, making the clarity provided by explainable AI indispensable. The need for interpretability in audio deepfake detection arises from several factors:
  • Trust and Adoption: Black-box AI systems, while powerful, often fail to gain user trust due to their opaque decision-making processes. In forensic contexts, where evidence may be scrutinized in court, interpretability builds confidence by providing insights into how and why predictions are made [81]. This transparency is essential for the credibility of the evidence, ensuring that forensic experts and legal practitioners can understand and explain the basis of the AI’s conclusions.
  • Debugging and Model Improvement: Understanding model behavior helps researchers identify biases, weaknesses, or errors in the detection process, leading to targeted improvements. In forensic applications, such insights are crucial for continuously refining the system, thereby ensuring that it reliably distinguishes between genuine and synthetic audio even under diverse and challenging conditions.
  • Accountability in Sensitive Applications: In high-stakes settings such as legal proceedings and forensic analysis, explainability ensures that decisions are defensible and grounded in understandable reasoning [82]. This traceability is vital in court, where experts must justify the methods and evidence used in reaching a conclusion. Clear explanations not only support the legal validity of the findings but also facilitate cross-examination by providing a transparent decision-making pathway.
  • Generalization and Robustness: Explainable models are designed to focus on meaningful, interpretable features rather than spurious correlations [83], inherently enhancing their ability to generalize across diverse datasets. These models exhibit promising performance when exposed to new, unseen deepfakes generated by techniques not encountered during training. Preliminary evaluations indicate that emphasizing core audio features such as critical frequency bands and temporal patterns enables the models to maintain reliability and fairness even as the characteristics of synthetic audio vary widely. This robustness is crucial in forensic scenarios, where audio samples differ significantly in quality, origin, and attack strategy. Moreover, the interpretability framework not only clarifies the model’s decision-making process but also serves as a diagnostic tool, enabling researchers to identify and adapt to novel deepfake patterns as they emerge [84]. A systematic analysis of feature attributions across diverse datasets helps identify potential vulnerabilities and address them effectively, reinforcing the model’s applicability in dynamic, real-world settings.

6.2. Methods for Interpretability and Explainability

Various interpretability techniques have been adapted to deepfake detection, focusing on feature importance, visualization, and prototype-based reasoning. Below, we explore these state-of-the-art methods in detail, along with their applications.
  • Feature Attribution Methods:
    Feature attribution techniques identify which parts of the input data contribute most to the model’s predictions. In audio deepfake detection, this often involves analyzing spectrograms or waveforms to highlight critical regions that differentiate real from synthetic audio. In image detection, this typically involves analyzing pixel importance through methods such as saliency maps, Grad-CAM, or SHAP to highlight key regions that influence the classification decision. Lim et al. [85] used layer-wise relevance propagation (LRP) and Deep Taylor Decomposition to explain predictions in spectrogram-based detection models. These techniques highlight frequency bands and temporal regions most influential in the model’s decisions, thereby providing forensic experts with insight into the acoustic features that distinguish genuine audio from deepfakes. Similarly, Yu et al. [86] employed SHAP to calculate feature attribution values in a lightweight machine learning framework. By visualizing these values on spectrograms, they identified critical high-frequency amplitude patterns and harmonics, thus improving both transparency and trust.
  • Attention Mechanisms for Explainability:
    Attention mechanisms provide a natural form of interpretability by highlighting parts of the input data that the model focuses on during decision-making. Channing et al. [87] implemented an attention roll-out mechanism for Transformer-based classifiers, visualizing attention weights across audio segments. This approach not only reveals the model’s focus but also pinpoints critical regions in the audio data that contribute most to classification, thereby enhancing both accuracy and transparency in forensic examinations.
  • Prototype-Based Interpretability:
    Prototype-based methods enhance interpretability by associating model decisions with specific, interpretable prototypes. Ilyas et al. [88] introduced prototype learning to align discriminative features with interpretable prototypes. Although primarily applied to visual deepfake detection, this approach can be adapted to audio. By associating specific characteristics (e.g., pitch or timbre for audio, texture, edges, or facial landmarks for images) with prototypes representing real or fake categories, forensic experts can more easily understand and communicate the model’s reasoning.
  • Lightweight and Explainable Frameworks:
    In resource-constrained environments, lightweight frameworks with built-in explainability offer a practical solution. Bisogni et al. [89] utilized hand-crafted features such as spectral centroid and Mel-Frequency Cepstral Coefficients (MFCCs) combined with SHAP to create interpretable models that achieve robust detection performance. These frameworks balance accuracy and transparency, making them suitable for real-time applications and forensic scenarios where timely, explainable decisions are critical.

7. Deepfake Attribution and Recognition

Deepfake Fingerprint and Attribution

The progression of deepfake generation technologies has necessitated sophisticated techniques to ensure accountability and mitigate malicious misuse. Researchers have focused on deepfake attribution, which involves identifying the specific model or architecture responsible for generating synthetic content, and ownership verification, aimed at protecting intellectual property. Deepfake attribution, also known as Deepfake Model Recognition, refers to the suite of methodologies designed to identify the specific generative model responsible for producing synthetic data [90,91]. This task involves not only recognizing the general architecture, such as GANs or diffusion models, but also attempting to estimate the unique weights of the model [92] that detail the instance of the architecture used to generate the deepfake. Current state-of-the-art (SOTA) methods have proven to be very effective in detecting deepfake content generated by GAN architectures [93,94,95] and Diffusion Models (DMs) [96,97], demonstrating the ability to specialize not only in recognizing the architecture but also in identifying the specific pattern within the creation process. Initial investigations have also been carried out in the audio domain, recognizing the generator used to generate speech deepfake signals [98].
Figure 5 provides a detailed conceptual overview of the process of detecting, classifying, and recognizing real and synthetic (deepfake) images, with a specific emphasis on identifying the generative model responsible for creating fake images. It begins with a dataset of facial images, which includes both genuine photographs and those generated by artificial intelligence. The first step involves determining whether an image is real or a deepfake. At this stage, the system focuses on differentiating between authentic content and those synthesized using generative models. For images classified as real, the process ends here. However, if an image is flagged as a deepfake, it undergoes further scrutiny to identify the specific generative technology behind it (GAN, DM, …). Next, an attempt is made to recognize the type of model used, such as GAN-based architectures (e.g., StyleGAN, StarGAN, or CycleGAN) or other generative frameworks like Stable Diffusion or DALL-E. Once the architecture is identified, the focus shifts to recognizing the precise instance of the model that produced the image. For example, if the architecture is identified as StyleGAN2, the next step is to determine the specific StyleGAN2 instance used in the image generation process. This capability is critical for forensic investigations, as it allows experts to trace the origins of synthetic content, offering insights into the tools and methodologies used.
In the context of the deepfake attribution task, several innovative SOTA approaches were proposed. Ning Yu et al. [99,100] introduced the concept of GAN fingerprints, revealing that generative adversarial networks (GANs) leave stable and unique traces in their output. These fingerprints allow for fine-grained attribution, enabling the identification of the exact GAN model used for synthesis. Their studies demonstrated the robustness of these fingerprints against adversarial perturbations, highlighting their potential in securing intellectual property and forensic investigations. Aiming to address real-world challenges, Yang et al. [101] explored architecture-level attribution in scenarios where models are fine-tuned or retrained. They proposed DNA-Det, which focuses on globally consistent fingerprints tied to the architecture, as opposed to model-specific weights. This method is particularly useful when dealing with privately trained or modified models, ensuring broader applicability in forensic settings. Sun et al. [102] introduced Contrastive Pseudo Learning (CPL) to address open-world attribution challenges. By integrating a Global–Local Voting module and soft pseudo-labeling, CPL aligns features across known and novel forgery types, enhancing model interpretability and security. A significant step forward in this field is the ability to distinguish between instances of the same architecture. For example, Guarnera et al. [90] demonstrated that the combination of a ResNET-18 model [103] with a metric learning approach [104] can achieve remarkable accuracy in distinguishing between 100 different instances of StyleGAN2-ADA [105]. This work highlights the feasibility of identifying the unique parameterization of a model instance. Furthermore, their approach showed promising results in preliminary tests involving models other than StyleGAN2-ADA, suggesting a potential for generalization across architectures.
The advent of open-world scenarios, where models encounter novel and unseen forgeries, has led to the development of benchmarks like OW-DFA++ by Sun et al. [106]. This benchmark integrates labeled and unlabeled datasets to evaluate model performance in diverse and dynamic contexts. Their proposed Multi-Perspective Sensory Learning (MPSL) framework employs multi-scale global–local feature alignment and confidence-adaptive pseudo-labeling to enhance attribution accuracy in such settings.
A robust solution for deepfake model recognition holds immense significance for intellectual property protection [107]. Such a system would enable the attribution of a specific synthetic image or video to its model owner, thereby addressing concerns over ownership and accountability in synthetic media. However, to achieve this level of precision, new strategies, and tailored metrics are required [108], particularly for cases involving high similarity between models trained with slight variations in data or hyperparameters.
In forensic applications, deepfake attribution parallels the role of camera source identification in traditional forensic investigations. Just as identifying the source camera helps trace the origin of a photograph, deepfake model recognition aims to trace synthetic media back to the specific model instance within a given architecture. This analogy underscores the critical need for advanced techniques to reliably attribute digital media to its generative source. The development of these techniques not only facilitates authenticity verification in digital forensics but also ensures accountability in cases of misuse or malicious intent. Emerging challenges in this domain include distinguishing between models trained with similar datasets, identifying models subjected to fine-tuning, and maintaining robustness against adversarial attacks [109] aimed at obfuscating model fingerprints. To address these challenges, future research can explore the integration of self-supervised learning, adversarial training, and ensemble methods that exploit multiple complementary approaches to model attribution.

8. Passive Deepfake Authentication Methods

In the modern era, where video calls have become a fundamental tool for global communication, ensuring the authenticity of audio and video streams is paramount. The rise of deepfake technology presents a substantial threat to the integrity of digital communication. To address this challenge, multimedia forensics researchers have developed various methods for detecting deepfakes, categorized based on the modality that they analyze.

8.1. Audio-Only Deepfake Detection

The rapid advancements in synthetic speech generation have increased interest in speech deepfake detection. Consequently, the scientific community has introduced a range of detection techniques employing diverse strategies and approaches [110]. Some methods focus on low-level features, looking for artifacts introduced by the generators at the signal level, while others analyze higher-level features that capture more complex aspects of speech.
An example of an artifact-based approach is presented in [111], where channel pattern noise analysis is used to secure Automatic Speaker Verification (ASV) systems against physical attacks. Similarly, the authors of [112,113,114] exploit bicoherence features based on the assumption that a genuine recording has more significant non-linearity than a fake one. Alternatively, the authors of [115,116] propose end-to-end networks trained for extracting deep features from speech, while [117] uses Mel-Frequency Cepstral Coefficient (MFCC) features as input of a Support Vector Machine (SVM) classifier. To enhance the practicality of existing detection methods in real-world scenarios, new strategies have been proposed in [118,119,120].
Detection approaches based on semantic features operate under the assumption that, while deepfake generators can synthesize low-level aspects of the signals, they are unable to replicate more intricate high-level features, and we can investigate such characteristics to discriminate real and fake data. For instance, the authors of [121] exploit classic audio features inherited from the Music Information Retrieval (MIR) community to detect speech deepfakes. Similarly, the authors of [122] leverage the lack of emotional content in synthetic voices generated via Text-to-Speech (TTS) techniques to recognize them, while those of [123] combine ASV and prosody features for the same task.
Recent trends in the speech deepfake detection field include the exploration of explainable AI (XAI) methodologies [124,125,126], the study of singing voice deepfakes [127], and the integration of pre-trained self-supervised models. These models, originally designed for Automatic Speech Recognition (ASR), are repurposed as feature extractors to create embeddings from input speech, which are subsequently used by synthetic speech detectors [128,129,130,131].

8.2. Video-Only Deepfake Detection

Techniques for detecting video-based deepfakes leverage visual content using a variety of approaches, including manual feature analysis and deep learning-based feature extraction.
Early forgery detection methods primarily depend on handcrafted features such as facial landmarks [27,132,133], optical flow [134], and various digital image processing techniques designed to enhance the visibility of artifacts [135]. While effective initially, these methods face limitations as deepfake generation technologies produce increasingly realistic and high-quality videos. Consequently, researchers have begun applying Deep Neural Networks (DNNs), which offer powerful feature extraction capabilities and support implicit learning of complex patterns. These methods have significantly improved the accuracy and reliability of detection processes.
As an example, the authors of [51,136] are pioneers in using DNNs to extract deep features from video frames. In [137], Convolutional Neural Networks (CNNs) and Long Short-Term Memory (LSTM) models are combined to detect fake videos generated using face-swapping techniques. The authors of [138] consider an ensemble of CNNs to detect video face manipulations, while those of [139] introduce multi-head attention and fine-grained classification to detect deepfake videos, showing that the approach is robust to low-quality videos. Liu et al. [140] analyze the frequency domain signal of deepfake videos and utilize the phase spectrum to obtain more information. Finally, the authors of [141] provide a semantic approach to deepfake detection, making use of a biological signal called photoplethysmography (PPG), an optical technique that can detect subtle changes resulting in skin color due to blood in peripheral circulation through the face.
These evolving strategies demonstrate the growing sophistication and versatility of video deepfake detection techniques, addressing both technical and physiological aspects of forgery.

8.3. Audio–Video Deepfake Detection

In recent years, there has been an increasing interest in the development of multimodal deepfake detection methods that can simultaneously analyze multiple modalities to achieve accurate and robust results. This comes from the fact that traditional deepfake detection methods may fall short as they often focus on either audio or video data in isolation. However, deepfakes may involve sophisticated manipulations of both audio and video streams, making them harder to detect with monomodal methods. By examining multiple modalities, multimodal detectors can identify inconsistencies or artifacts across audio and video streams, enhancing their detection capabilities. For example, a deepfake video might feature realistic facial expressions but unnatural background sounds or mismatched lip movements.
An example of such an approach is provided in [142], where the authors leverage the incongruity between emotional cues portrayed by audio and visual modalities. Conversely, the authors of [143] integrate temporal data from image sequences, audio, and video frames. Moreover, the results of [144] show that an ensemble of audio and visual baselines outperforms monomodal counterparts. The authors of [145] replace the standard MFCC features with an embedding of a deep neural-network trained for automatic speech recognition, and then incorporate mouth landmarks. In [146], the authors establish a mapping between audio and video frames by analyzing the changes in the lip opening degree, while in [32,147,148], the authenticity of a speaker is verified by detecting anomalous correspondences between their facial movements and what they say. Finally, the authors of [149] show that it is possible to perform multimodal analysis even when considering separate detectors individually trained on monomodal tasks.
Despite the great effort of the multimedia forensics community, a series of challenges remain. Concerning multimodal solutions, the need for audio–video deepfake datasets is becoming more than an urgent necessity [42,53], as the majority of current efforts have focused on developing monomodal datasets. Moreover, with the rise of large language models, there is potential to adapt similar reasoning frameworks for audio-visual analysis, which could open new avenues for research and innovation in Deepfake detection. Finally, explainability remains a significant challenge. Current methods often lack the transparency required for legal and forensic applications, making their use in courtrooms problematic. Enhancing the interpretability of these systems is essential for ensuring their reliability and acceptance in high-stakes scenarios.

9. Deepfakes Detection Method on Realistic Scenarios

9.1. Deepfake Detection of Multimedia in the Wild

The rapid advancement of deepfakes and AI-generated media has introduced significant challenges for detection systems operating in real-world environments, while Deep Learning (DL) models have demonstrated effectiveness in controlled settings, their performance often deteriorates in practical applications where uncontrolled variables, such as compression, adversarial manipulations, and evolving generative techniques, come into play [150,151]. This highlights the need for detection models that not only excel in recognizing current threats but can adapt continuously as new forms of deepfake media emerge.
A critical obstacle to achieving this lies in the tendency of DL models to become outdated when exposed to unseen data distributions, commonly referred to as data drift. The inability to generalize across evolving threats leads to gaps in detection, undermining the reliability of forensic systems. Furthermore, as models are retrained to address new threats, they risk forgetting previously learned patterns, a phenomenon known as catastrophic forgetting [152]. This creates a cycle where models must repeatedly start from scratch, limiting their long-term effectiveness.
To mitigate these issues, Continual Learning (CL) frameworks are increasingly adopted to ensure that detection models retain prior knowledge while adapting to novel manipulations [62]. CL enables systems to evolve alongside deepfake generation techniques, fostering resilience and sustained performance in diverse environments. However, even with continual adaptation, the lack of interpretability in DL models poses a barrier to their deployment in sensitive areas such as forensic analysis, legal proceedings, and media authentication. The black-box nature of these systems makes it difficult to justify or explain decisions, reducing trust and acceptance in high-stakes applications [153].
Addressing this, the integration of explainable AI (XAI) within deepfake detection pipelines is becoming a priority [154,155]. By enhancing transparency and providing insights into model behavior, XAI ensures that deepfake detection systems can be audited, understood, and trusted by non-technical stakeholders. This dual focus on adaptability and interpretability forms the foundation for next-generation detection frameworks that can operate effectively across dynamic and adversarial landscapes.
The path forward involves developing end-to-end deepfake detection systems that combine continual learning with interpretable decision-making processes. These systems must not only recognize deepfakes with high accuracy but also articulate the rationale behind their classifications, ensuring accountability and fostering confidence in their outputs. Leveraging Machine Learning Operation (MLOp) pipelines, future detection models can achieve continuous updates and seamless integration into real-world workflows, enhancing their robustness against evolving threats [156,157].

9.2. Deepfakes and Social Media

A highly challenging “real-world” scenario involves detecting deepfake multimedia shared on social networks [158,159,160]. To address bandwidth and storage constraints [158], social networks apply aggressive data compression and resizing. However, while these processes reduce the size of multimedia files, they also diminish the forensic features crucial for distinguishing real from fake content [158,161,162].
An initial study on GAN-generated images [159] shared via Twitter highlights the adverse effects of social network compression on deepfake detectors. Specifically, although the visual quality of shared images remained intact, the presence of forensic traces was significantly reduced. These effects were further analyzed in detail in [158].
In [160], the authors examine the challenges and advancements in media forensics as applied to social networks. The study addresses growing concerns regarding the authenticity and reliability of digital media shared on these platforms, focusing on challenges that impact source attribution algorithms [163] and multimedia verification [161]. For multimedia verification, particular attention was given to assessing whether multimedia content aligns with its descriptive text. The authors of [160] discuss emerging challenges such as the proliferation of deepfakes and the use of bots to spread disinformation. They emphasize the need for advanced forensic tools to counter these sophisticated manipulation techniques.
In [158], the authors assembled a large and diverse dataset, visible in Figure 6, comprising 80k fake images generated with StyleGAN models and 70k real images sourced from several state-of-the-art datasets [37]. The study also provides insights into the extent and impact of compression applied by Twitter, Facebook, and Telegram. Furthermore, the authors demonstrated how this dataset could be utilized to fine-tune new detectors, preserving their accuracy on social network-compressed images while avoiding “catastrophic forgetting loss” [164].
Interestingly, the authors of [165] showed that while social networks degrade the forensic artifacts used by real vs. fake detectors, they introduce new traces that can help reconstruct the multimedia’s life cycle and identify the social networks where it was shared. Although the life cycle of multimedia does not directly reveal whether the content is real or fake, it aids in recovering versions closer to the original, uncompressed file, thereby enhancing detection accuracy.
In the last few years, fake images have increasingly been used on social networks to impersonate humans with realistic profiles and engage in various inauthentic activities [166], generating a growing interest in the topic and the development of methods able to detect these images.
In their work, Yang et al. [166] presented a systematic analysis of fake accounts that use GAN-generated faces on X (formerly Twitter) and a heuristic based on the eye coordinates in GAN-generated images for identifying such accounts. The authors [166] observed that GAN-generated images of faces tend to place the eyes in approximately constant coordinates. Astonishingly, the authors, following their heuristic, estimate a total of 10,000 active accounts using GAN-generated images of faces per day on X.
Similarly, Ricker et al. [167] provided a systematic, large-scale study of AI-generated profile images on X. They propose a multi-stage pipeline based on a ResNet50 [168] to distinguish real from fake images. In their study, Ricker et al. [167] estimate a total number of accounts using GAN-generated faces as profile pictures very close to the estimate provided by [166].
Maier and Riess [162] extended the studies on detecting fake GAN-generated images on social networks to include images generated by diffusion models. In their paper, the authors propose a Bayesian neural network (BNN) for detecting synthetic images. The BNN excels in identifying out-of-distribution inputs and achieves performance comparable to state-of-the-art detectors while offering more reliable predictions.
Accordingly, Kumar et al. [169] study deepfake implementation methods and their detection using GAN-based deep convolution models. They also provide a comparative analysis of the proposed GAN with other models using Inception Score (IS) and Fréchet Inception Distance (FID) metrics.
Despite recent progress in deepfake detection on social networks, the problem is far from being solved. Indeed, social network image processing algorithms are constantly updated, and fake image detectors, trained on real and fake images shared on social networks during the last few years, need to be fine-tuned according to these updates [158].
To make things worse, Lago et al. [170] showed how fake images are becoming increasingly indistinguishable to humans from real images.
This makes it necessary to develop methods that are constantly updated through continual- and few-shot learning [62,171].
Finally, preliminary studies on videos shared on social networks [172] reveal similar effects to those observed with images. One such study [21,172] investigated the impact of social network compression on FaceForensics [172] videos shared on Facebook and YouTube. The findings align with those for images [172] and introduce a new dataset of shared videos for fine-tuning real vs. fake detectors.

9.3. Detection of Deepfake Images and Videos in Adversarial Setting

An additional problem affecting virtually all the deepfake forensic techniques developed so far is that such techniques are thought to operate in a benign setting, that is, by neglecting the possible efforts made by an adversary to mislead the forensic analysis. Yet, recent studies [173,174] have shown how easy is to generate adversarial contents capable of deceiving image and video processing techniques based on DL, when the adversary is informed about the details of the tools employed by the analyst. Some works [175,176] have also studied the transferability of adversarial examples to networks different than those targeted by the attack, opening the way to the development of powerful attacks even when the attacker is unaware, or only partially aware, of the techniques used by the analyst. This includes recent techniques based on Contrastive Language-Image Pretraining (CLIP) features employing a vision transformers as backbone. In [177], for instance, it is shown that methods based on CNNs and CLIP features share the same vulnerabilities against adversarial attacks, even if the kind of perturbations introduced within the attacked images in the two cases is quite different. For this reason, attacks against CNN-based techniques usually do not transfer to methods based on vision transformers and CLIP features, and vice versa.
Often, it is not even necessary that the adversary applies sophisticated attacks relying on the full or partial knowledge of the to-be-attacked system. By relying on the lack of robustness and the generalization capabilities of the forensic tools outlined in Section 9.1, the adversary may simply process the deepfake content in such way to prevent a correct forensic analysis, or at least degrade its performance to a point that makes it unusable. Some examples of this kind of attack, often referred to as laundering attacks, include the application of moderate to strong lossy compression, geometric processing of images and videos, noise addition, histogram stretching, and many others.
Understanding and ensuring the security of deepfake forensic tools is a crucial problem if such tools have to be used under the intrinsically adversarial conditions typical of multimedia forensics applications. For this reason, several efforts have been made to defend against adversarial attacks [178,179], both in the realm of computer vision applications and multimedia forensics. Still, no general effective solutions have been found yet [180]. Among the solutions developed so far, adversarial training [181] has received some consensus and has proven to at least mitigate the effectiveness of adversarial attacks in computer vision applications. As argued in [182], adversarial training forces DL models to focus on robust, possibly semantic, features, which are inherently more difficult to attack. Whether such a beneficial effect of adversarial training also applies to deepfake forensic applications is still an open problem. It is not clear, in fact, if in multimedia forensics the equivalent of semantic computer vision features exist.
With regard to laundering attacks, the solutions proposed so far are similar to those already discussed in Section 9.1, given that, ultimately, the effectiveness of laundering attacks can be drastically reduced by improving the robustness and generalization capabilities of the forensic tools. A common approach to do so involves the use of data augmentation techniques that enrich the training set with processed samples, thus improving the robustness against the processing operators used for data augmentation. Yet, accounting for all possible kinds of processing during training is clearly unfeasible. Among the solutions proposed so far, one is the possibility of identifying a kind of worst possible laundering attack, an approach we have included in the training procedure. An example of such an approach is the inclusion within the data augmentation pool of a Print and Scan (P&S) simulator [183]. The P&S process, in fact, introduces various distortions that tend to eliminate or significantly diminish the subtle statistical traces often relied upon by synthetic image detectors. In [183], the authors contend that training a detector to maintain its performance on images that have undergone P&S compels it to depend on robust features that remain discernible even after substantial transformations. Given the impracticality of creating a large dataset of physical P&S images, in [183] the authors utilize a CycleGAN to simulate the P&S process. The CycleGAN P&S simulator was initially trained on image patches (patch P&S) and then fine-tuned on full face images (full P&S). This approach was tested on a synthetic image detection task, using both common processing and P&S simulation in the augmentation list. Validation was carried out on natural image datasets like CelebA-HQ and FFHQ (N), alongside synthetic datasets generated by StyleGAN2 (SG2), Latent Diffusion (LT), and Taming Transformers (TTs). Experimental results, summarized in Table 2, demonstrate that detectors trained with P&S-augmentation achieve greater robustness to unseen processing techniques, including JPEG2000 compression, WebP, filtering, and color transformations, without sacrificing performance on the processing tasks used during augmentation.
Another example of the worst case attack approach is given in [184], where a double JPEG detector is trained on a selected pool of images subject to a set of worst case laundering attacks belonging to different processing categories.
Despite all the efforts made, even for laundering attacks, a definitive solution has not been devised yet, thus adding yet another point to the to-do list of multimedia forensic researchers.

10. Active Authentication

10.1. Active Deepfake Detection

Deepfake detection has been approached so far by means of passive techniques [185,186,187] that operate after the fact, meaning they are applied after the forged content has been created, distributed, and potentially manipulated. Conversely, active methods work proactively by pre-processing media and embedding a fingerprint to facilitate subsequent analysis. This is obviously an important change in the paradigm and it foresees the capacity to intervene at the stage of the content generation process and, above all, the active cooperation of the entity that trained the media-generation network. This last option is not so easy to achieve but it can be motivated by the fact that entities and organizations that have spent effort in training models for image/video generation should be willing to protect them. Moreover, they should also have an interest in tracking the multimedia content they are going to produce through the interactive services they make available to the common end-users. In addition, there will also be the need to understand if such fingerprinted contents are used as a basis to construct new trained models for deepfake creation. So it could be crucial to individuate if a particular deepfake document (e.g., an image) has been produced by a specific generation system in order to perform source attribution and authenticity assessment. One of the basic ideas behind active fingerprinting for active deepfake detection was presented in [100]. The method can be divided into two main steps. The first is to train a model, constituted by a couple encoder–decoder ( E n c / D e c ), that, given an image I as input, is able to embed a watermark w (a string of n bits) into a reconstructed image I . The training process has to minimize a loss function composed of two terms, a mean squared error between the input image and the watermarked-reconstructed one ( L M S E ) and a binary cross entropy ( L B C E ) between the input watermark w and the decoded one w . When this network is trained, it is adopted for the second step, which is the actual active fingerprinting. The encoder ( E n c ) is now used to embed a specific watermark w i in a dataset of images so to obtain a watermarked dataset that contains images marked with w i . Such a dataset is then used to train a certain generative network ( G A N i composed of a generator G e n i and a discriminator D i s i ) that will learn to generate images with the watermark w i embedded. After that, if such trained image generator G e n i will be used to create a deepfake content (e.g., a deepfake image), it will be possible, by taking the decoder D e c , to extract the watermark w i from such image in order to check its provenance and, at the same time, to determine that the content is not pristine. The same process can be extended to N different kinds of generation systems ( 1 i N ) to perform fingerprinting of the diverse deepfake generative models. Obviously, these kinds of approaches need to be tested and improved to satisfy diverse requirements to become actually efficient; in fact, it is fundamental to balance the need to grant robustness against possible image processing with the request to preserve image quality and to provide a sufficient capacity in information hiding. All this opens the way to possible new research opportunities in the field of active deepfake detection.

10.2. Efficient Media Origin Authentication

Customary deepfake detection methods, both passive and active, are subject to false positives and false negatives, whose rate highly depends on the employed method and the goodness of the training data. False positives are due to various factors, such as the complexity of the content, the quality of the training data, or the intrinsic limitations of the detection algorithm itself. False negatives could happen if the deepfake is very well made, or in general if the detection method fails to recognize certain patterns or features that indicate a deepfake. On the other hand, cryptographic signatures are “almost perfect” from this point of view, in the sense that false negatives (i.e., authentic signatures which are not recognized so) are zero and false positives (i.e., fake signatures that are taken as authentic) are considered computationally infeasible to forge. This suggests that cryptographic signatures could be used fruitfully to detect deepfakes (or better, the absence of deepfakes) with perfect precision. In this direction, the work-in-progress standard JPEG Trust [188] by the Joint Photographic Experts Group (JPEG) aims to establish trust in digital media by addressing aspects of authenticity, provenance, and integrity. The JPEG Trust will provide a framework for establishing trust in media through secure annotation of media assets throughout their life cycle, using cryptography as a key component. Cryptographically signing media allows deepfakes to be repudiated by the interested person, since signatures will be absent or invalid. Such an anti-fake signature should allow “good” manipulation of the original file (at least cropping) and disallow “bad” manipulation, but also be space efficient to save bandwidth on web servers once the media file is disseminated. Unfortunately, customary signature schemes like ECDSA do not have these properties.
To address this challenge, a solution could rely on novel aggregatable signatures, such as the Boneh–Lynn–Shacham (BLS) signature [189,190], which has been successfully used in blockchain technologies like Ethereum 2.0 to optimize storage (https://github.com/ethereum/consensus-specs/blob/dev/specs/phase0/beacon-chain.md#bls-signatures, accessed on 15 February 2025). The BLS signature scheme makes use of a novel form of cryptography called pairing-based cryptography, which allows for a plethora of new functionalities like attribute-based encryption [191,192,193]. Aggregatable signatures could be employed in JPEG (possibly within the JPEG Trust standard itself) in such a way to permit benign alterations of the image like cropping while preventing malicious tampering, without increasing too much the bandwidth occupation on web servers.
The system is composed of three entities: Signer, Publisher, and Verifier. The Signer represents the interested person (or an agency that manages their public reputation), which is the only entity in possession of the private key. The Signer will sign the individual JPEG blocks that make up the image, and the signatures produced will be stored as metadata within the JPEG file. The JPEG file, along with the related BLS signatures, will then be passed to the Publisher, which represents a news agency interested in publishing photographic material through the web. The Publisher can possibly crop the original image, by removing the JPEG blocks and the related signatures that are outside a given cropping rectangle. The remaining signatures (those relative to the blocks inside the cropping rectangle) can be aggregated into a single signature by the Publisher, leveraging the aggregability properties of BLS. This allows the Publisher to drastically reduce the size of the JPEG file, therefore saving a lot of bandwidth on the web server. Finally, the Verifier represents the final user, which wants to verify the origin of the (possibly cropped) image. The Verifier extracts the aggregated signature embedded in the JPEG file downloaded from the Publisher’s web site and verifies it by means of the Signer’s public key.

10.3. Trusted Remote Media Processing on Cloud and Edge Computing Systems

In the analysis of new techniques and future challenges, it is important to mention the new use of Distributed Systems in the context of AI, one of the main modern outcomes of which is the creation of the Transformer paradigm [194]. Here, OpenAI has also benefited from this, having leveraged the power of the cloud in conjunction with Big data in creating ChatGPT, based initially on GPT-3 then on the GPT-4 architecture [195,196].
In this work, we focus on emerging media production contexts, i.e., systems based on the IoT (Internet of Things), which is playing an important role increating the Digital Smart Data Cities of the near future. Here, citizens can interact with the environment and benefit from a plethora of advanced services, such as video surveillance, intelligent traffic lightning, air quality detection, and new advanced fire and flooding management systems. From a technological point of view, sensors and actuators able to automate new services are strategic in their configuration, and optimization, hence, even more digital infrastructures need to adapt their behavior to the specific needs of the context; this represents a big challenge both in terms of system design and security. Deepfake media detection in these scenarios represents a challenge due to the nature of possible manipulations of future digital content in citizens’ daily lives; hence, a more holistic approach should be considered where media production happens, since Edge Computing Systems [197] need to be taken into account. In just 20 years, with the objective of increasing system response and reduce communication latency, computation moved from mainframes and computing rooms towards Cloud Computing, Fog Computing, and lastly, Edge Computing. A Federated Cloud–Edge infrastructure is considered, where different administrative domains are in place and where Artificial Intelligence and Machine Learning software artifacts, in the assertion of Federated Learning even at the Edge, help to distribute intelligence in this scenario.
Multimedia acquisition devices based on the IoT generate an unprecedented amount of data, leading to more frequent instances of manipulated and fake data [198], hence, the need to develop cloud-based video big data analytics frameworks. A distributed approach in video recording and elaboration systems, such as video surveillance systems based on IP cameras, is highly recommended to overcome the maximum storage or throughput limitation of network video rRecorders installed on single machines. To perform such a variety of tasks, and to be able to modify a device’s behavior on-demand, the Function as a Service (FaaS) computational paradigm has generally been adopted. FaaS allows several minimal applications to be defined and one or more instances of these applications to be run on the same device at the same time. The FaaS framework relies on two configuration approaches: a local configuration file, generally YAML, or a secure remote server. However, both come with limitations: a local file configuration requires direct access to the device, physically or through a secure connection, to modify it; while, although a remote server can store and send updated configuration files, it might be vulnerable to well-known cyberattacks suck as Man-in-the-middle (MITM) or Distributed Denial of Service (DDoS), making it unusable and unreliable. To overcome such limitations, it is possible to benefit from three technologies that have been increasingly recognized to be able to address information access problems and system trustworthiness in different application domain: Federated Learning [199], the Blockchain, and the IPFS (InterPlanetary File System) starting at the Edge:
  • Federated Learning is a decentralized approach to training machine learning models. In traditional machine learning, data are centralized in the Cloud, where a single model is trained on the entire dataset. Federated Learning, conversely, allows the training of machine learning models across multiple decentralized devices or servers that hold local data samples without exchanging them. Moreover, Federated Learning at the Edge refers to the application of Federated Learning techniques on Edge devices, such as IoT devices, or Edge Servers. This approach combines the benefits of Federated Learning, which ensures data privacy and reduces communication costs, with the advantages of Edge Computing Systems, which enables data processing and model training to occur closer to where the data are generated, hence, fake media might not exit from the Edge. New approaches can be adopted like the Federated Learning system deployed into Web Browsers for enlarging the the Cloud–Edge–Client Continuum capabilities [200,201].
  • The use of the Blockchain, supported by the flexibility and robustness of smart contracts, allows the combination of the well-known FaaS paradigm with the intrinsic features of data non-repudiation and immutability, replacing the service configuration with a smart contract, guaranteeing protection against distributed cyber-attacks [202].
  • The IPFS is a distributed system for storing and accessing files. Since the block size of the Blockchain does not allow storing files, these can be uploaded to this special file storage, which produces a unique hash value to be used as a key to access its content [203].

11. Discussion

The advances and challenges discussed in this paper highlight the evolving landscape of deepfake detection, authentication, and forensic analysis. As synthetic media generation continues to improve, detection methodologies must adapt to maintain their effectiveness in real-world scenarios. This section, devoted to a generic discussion, summarizes the key points presented in the paper, evaluating current detection strategies, identifying existing limitations, and outlining future research directions to improve the robustness, efficiency, and ethical use of deepfake detection technologies.

11.1. Evaluating Deepfake Detection Strategies: Challenges, Advances, and Future Directions

The findings and methodologies presented in this comprehensive review highlight the dynamic interplay between the rapid advancements in deepfake generation technologies and the equally rapidly evolving detection strategies. Table 3 provides a structured comparison of detection methods analyzed in this paper, outlining their advantages and weaknesses while referencing the sections where they are examined in detail. As manipulation techniques become increasingly sophisticated, detection methods must continually adapt to address new forensic challenges, each requiring different approaches and levels of robustness. In the short term, one of the most pressing challenges remains the availability of diverse and high-quality datasets for training and evaluating deepfake detection models. Existing datasets often fail to represent the full spectrum of real-world manipulations, limiting the generalization capabilities of forensic models. A possible solution could be expanding dataset diversity to include a broader range of manipulations and real-world distortions. However, this approach alone is insufficient, as new deepfake techniques constantly emerge, making it impractical to cover all possible variations. Instead, future efforts should prioritize adaptive strategies, such as synthetic augmentation, domain adaptation, and meta-learning, to improve the generalization capabilities of detection models beyond their training distributions.
A key mid-term challenge is improving the adaptability and resilience of detection models, while forensic-based techniques leveraging frequency-domain analysis and spatio-temporal coherence can detect subtle artifacts, their reliance on high-quality input data and substantial computational resources limits their practicality in real-world applications. Machine learning-based approaches, particularly deep learning methods, have demonstrated strong performance on benchmark datasets, but they remain vulnerable to adversarial manipulations and suffer from high computational costs. Another major limitation across detection strategies is their sensitivity to media compression. Social media platforms, where deepfake content is frequently disseminated, apply aggressive compression that often degrades forensic traces. Future research should focus on developing robust detection frameworks that integrate compression resilience without compromising detection accuracy. Continual learning frameworks present a promising solution, allowing models to evolve alongside new threats while mitigating catastrophic forgetting. However, the lack of universally accepted benchmarks for evaluating these methods remains a significant bottleneck, hindering their large-scale deployment. Additionally, the integration of explainable AI into forensic pipelines is becoming increasingly important, providing interpretability to detection decisions and improving their usability in real-world forensic investigations. Future works should balance model transparency with detection performance, ensuring that forensic tools are both interpretable and robust.
In the long term, an effective deepfake mitigation strategy must extend beyond detection models and incorporate proactive defense mechanisms such as active authentication. Embedding traceable fingerprints at the content creation stage provides a safeguard against manipulation, but its efficacy is dependent on widespread adoption by content platforms and resilience against tampering. Furthermore, regulatory frameworks must evolve to address deepfake forensics’ ethical and legal implications, ensuring a balance between content authenticity verification and fundamental rights such as privacy and freedom of expression. Given the rapid advancements in generative AI, the long-term success of deepfake detection will require a coordinated effort between researchers, policymakers, and industry stakeholders to develop scalable and adaptive solutions that align with both technological advancements and societal needs.

11.2. Ff4ll’S Goals in Deepfake Detection and Media Authentication

The FF4LL project is designed to address the evolving challenges of deepfake detection and media authenticity by providing comprehensive solutions that integrate detection, attribution, and authentication strategies. Given the increasing sophistication of generative models and their potential for abuse, FF4LL aims to bridge the gap between cutting-edge research and practical dissemination, ensuring that forensic tools remain robust and available in different real-world applications.
The overall goals of FF4LL can be categorized into three main points:
  • Advancing deepfake detection through multimodal approaches: The project aims to develop robust and scalable detection frameworks capable of countering image- and audio-based forgeries. Leveraging machine learning, frequency analysis, and adversarial robustness techniques, FF4LL aims to improve detection accuracy while reducing computational overhead by addressing key limitations in this field;
  • Ensuring media authenticity through active and passive authentication: FF4LL emphasizes the integration of proactive authentication methods to incorporate traceable fingerprints at the time of content creation. This approach, combined with device authentication and secure cloud-based media processing, ensures that the integrity of content can be verified even before forensic analysis is required, enhancing trust in digital media ecosystems;
  • Impact on cybersecurity, digital communication, and public trust: The implications of deepfake proliferation go beyond media forensics to cybersecurity, disinformation control, and public perception of digital content. FF4LL’s contributions to secure authentication and forensic analysis aim to mitigate fraud, identity theft, and disinformation campaigns by providing governments, journalists, and digital platforms with the tools to safeguard digital communication channels.
By integrating these approaches, FF4LL provides a framework that aligns with both current challenges and future advances in synthetic media generation. The project’s interdisciplinary collaboration among AI researchers, forensic experts, and many others ensures that deepfake detection and authentication technologies evolve in parallel with emerging threats, promoting a more secure and trusted digital landscape.

11.3. Ethical Concerns and Governance in Deepfake Detection

Deepfake detection technology, while critical to safeguarding the integrity of digital media, raises significant ethical issues that require careful consideration. One of the most pressing issues is privacy. The effectiveness of these tools relies on the analysis of facial features, voice patterns, and other biometric data, which are inherently sensitive. The collection and storage of this information presents substantial privacy risks, especially in the event of unauthorized access. These data, if misused, could lead to identity theft, surveillance, or a serious breach of personal privacy. Therefore, while the intent of using deepfake detection technology may be to protect people from deception, it also exposes them to potential violations of their biometric security. Beyond privacy, the issue of disinformation poses a significant ethical dilemma. Deepfakes have already demonstrated their ability to manipulate public opinion by creating realistic but false narratives. Even with advances in detection technologies, the damage caused by misleading content can be rapid and widespread before it is actually exposed. Moreover, a paradox emerges: the evolution of detection tools drives the development of more sophisticated deepfakes. Malicious actors, learning from detection algorithms, can create increasingly deceptive content. This not only reduces public trust in digital media but also compromises the entire information system. To address these ethical challenges, it is essential to develop comprehensive standards and policies that govern the use of deepfake detection technology. A primary measure is the implementation of strict data privacy standards. These standards should ensure that biometric data are securely stored, anonymized, and used only for detection purposes. Compliance with international privacy standards, such as the General Data Protection Regulation (GDPR) or the California Consumer Privacy Act (CCPA), is critical to maintain transparency and protect users’ rights. It is also necessary to implement robust consent mechanisms that allow individuals to control their biometric data, including the right to request deletion. In terms of privacy protection, it is essential to ensure transparency and accountability to reduce the risks associated with deepfake detection technology. It is important that deepfake content is clearly labeled and detection tools are used so that users know whether a piece of media is authentic or not. In addition, the creation of independent institutions to monitor these technologies would help ensure fairness and prevent abuse. Thus, although deepfake detection technology plays a key role in countering digital deception, its ethical implications must be carefully managed. Privacy concerns, the risk of misinformation, and the potential for misuse underscore the need for balanced and comprehensive policies. By implementing strict data privacy standards and promoting transparency and responsibility, society can take advantage of the benefits of deepfake detection technology while minimizing ethical risks. This balanced approach will be critical to safeguarding both individual rights and the integrity of digital information.
To counter the disinformation risks posed by deepfakes, FF4LL researchers focus on robust detection strategies that remain effective even when synthetic content is highly compressed and manipulated for adversarial purposes. These techniques reduce the rapid dissemination of false information before it can significantly influence public opinion. In addition, FF4LL studies explainable AI (XAI) approaches to improve the interpretability and forensic reliability of deepfake detection, ensuring that decisions are transparent and defensible in legal and cybersecurity contexts. By balancing technological advances with privacy safeguards, transparency, and ethical considerations, FF4LL aims to strengthen public trust in digital media while minimizing the ethical risks associated with deepfake detection. Through the collaboration of AI experts, forensic and legal experts, the project is helping to create a safer, more responsible and ethical media landscape.

12. Conclusions

This study presents a comprehensive analysis of the current landscape of deepfake media forensics, focusing on detection, attribution, and authentication techniques. The work highlights the critical challenges posed by synthetic media in controlled and real-world scenarios, as well as the technological advancements developed to counteract these threats. While significant progress has been made, the field faces persistent challenges related to scalability, real-time applicability, and robustness in dynamic environments. Scalability remains a pressing issue, as many current state-of-the-art techniques require high computational resources and extensive labeled datasets, which limit their adoption in large-scale or resource-constrained settings. Addressing this challenge requires the development of lightweight, energy-efficient algorithms, potentially leveraging model pruning, quantization, or distillation to maintain performance while reducing computational demands. Real-time applicability is another critical hurdle. Many detection algorithms, particularly those employing deep learning, suffer from latency issues, making them unsuitable for applications such as live media verification or streaming platforms. To overcome this limitation, future research should focus on optimizing inference times by using edge computing frameworks and designing architectures tailored for low-latency environments, such as transformer-based models adapted for efficiency. The robustness of detection systems in dynamic environments is equally crucial. As generative technologies evolve, new types of deepfake manipulations constantly emerge, challenging the adaptability of current detectors. Adversarial attacks, in particular, pose a significant threat, as they can exploit weaknesses in detection systems to evade scrutiny. Enhancing the generalization capabilities of detection models is vital and can be achieved through continual learning frameworks, adversarial training, and extensive data augmentation strategies that simulate diverse real-world conditions. These approaches can help models retain knowledge of previously encountered manipulations while adapting to novel threats. Beyond technical advancements, addressing ethical considerations is essential to ensure the responsible use of these technologies. Standardized guidelines should be established to tackle privacy concerns, prevent misuse, and promote transparency in the development and application of detection tools. Prioritizing initiatives like explainable AI (XAI) can enhance the interpretability and trustworthiness of detection systems, particularly in forensic and legal contexts where their outputs may significantly influence critical decisions. In conclusion, while the advancements in deepfake media forensics are commendable, a concerted focus on technical innovation and the establishment of robust ethical frameworks is essential to address the challenges posed by synthetic media and ensure the integrity and authenticity of digital content in an increasingly complex technological landscape.

Author Contributions

Conceptualization, L.G., A.O., S.B., R.C., P.P., M.M., G.L.M., G.O., G.P., I.A., T.M., P.B., S.M., D.S., S.T., M.V., M.B., N.P., A.M., G.B., F.D.N., R.D.N., D.V. and V.B.; methodology, L.G., A.O., S.B., R.C., P.P., M.M., G.L.M., G.O., G.P., I.A., T.M., P.B., S.M., D.S., S.T., M.V., M.B., N.P., A.M., G.B., F.D.N., R.D.N., D.V. and V.B.; formal analysis, L.G., A.O., S.B., R.C., P.P., M.M., G.L.M., G.O., G.P., I.A., T.M., P.B., S.M., D.S., S.T., M.V., M.B., N.P., A.M., G.B., F.D.N., R.D.N., D.V. and V.B.; investigation, L.G., A.O., S.B., R.C., P.P., M.M., G.L.M., G.O., G.P., I.A., T.M., P.B., S.M., D.S., S.T., M.V., M.B., N.P., A.M., G.B., F.D.N., R.D.N., D.V. and V.B.; writing—original draft preparation, L.G., A.O., S.B., R.C., P.P., M.M., G.L.M., G.O., G.P., I.A., T.M., P.B., S.M., D.S., S.T., M.V., M.B., N.P., A.M., G.B., F.D.N., R.D.N., D.V. and V.B.; writing—review and editing, L.G., A.O., S.B., R.C., P.P., M.M., G.L.M., G.O., G.P., I.A., T.M., P.B., S.M., D.S., S.T., M.V., M.B., N.P., A.M., G.B., F.D.N., R.D.N., D.V. and V.B.; visualization, L.G., A.O., S.B., R.C., P.P., M.M., G.L.M., G.O., G.P., I.A., T.M., P.B., S.M., D.S., S.T., M.V., M.B., N.P., A.M., G.B., F.D.N., R.D.N., D.V. and V.B.; supervision, L.G., A.O., S.B., R.C., P.P., M.M., G.L.M., G.O., G.P., I.A., T.M., P.B., S.M., D.S., S.T., M.V., M.B., N.P., A.M., G.B., F.D.N., R.D.N., D.V. and V.B. All authors have read and agreed to the published version of the manuscript.

Funding

This study has been partially supported by SERICS (PE00000014) under the MUR National Recovery and Resilience Plan funded by the European Union—NextGenerationEU.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data are contained within the article.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
FaaSFunction as a Service
MITMMan-in-the-Middle
DDoSDistributed Denial of Service
XAIExplainable Artificial Intelligence
GANsGenerative Adversarial Networks
DMsDiffusion Models
SOTAState of the art

References

  1. Goodfellow, I.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative Adversarial Nets. Adv. Neural Inf. Process. Syst. 2014, 27. [Google Scholar]
  2. Ho, J.; Jain, A.; Abbeel, P. Denoising Diffusion Probabilistic Models. Adv. Neural Inf. Process. Syst. 2020, 33, 6840–6851. [Google Scholar]
  3. Casu, M.; Guarnera, L.; Caponnetto, P.; Battiato, S. GenAI Mirage: The Impostor Bias and the Deepfake Detection Challenge in the Era of Artificial Illusions. Forensic Sci. Int. Digit. Investig. 2024, 50, 301795. [Google Scholar] [CrossRef]
  4. Guarnera, L.; Giudice, O.; Guarnera, F.; Ortis, A.; Puglisi, G.; Paratore, A.; Bui, L.M.; Fontani, M.; Coccomini, D.A.; Caldelli, R.; et al. The Face Deepfake Detection Challenge. J. Imaging 2022, 8, 263. [Google Scholar] [CrossRef] [PubMed]
  5. Wang, S.Y.; Wang, O.; Zhang, R.; Owens, A.; Efros, A.A. CNN-Generated Images are Surprisingly Easy to Spot …for Now. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 8695–8704. [Google Scholar]
  6. Guarnera, L.; Giudice, O.; Battiato, S. Level Up the Deepfake Detection: A Method to Effectively Discriminate Images Generated by GAN Architectures and Diffusion Models. In Proceedings of the Intelligent Systems Conference, Guilin, China, 26–27 October 2024; Springer: Berlin/Heidelberg, Germany, 2024; pp. 615–625. [Google Scholar]
  7. Amerini, I.; Barni, M.; Battiato, S.; Bestagini, P.; Boato, G.; Bonaventura, T.S.; Bruni, V.; Caldelli, R.; De Natale, F.; De Nicola, R.; et al. Deepfake Media Forensics: State of the Art and Challenges Ahead. arXiv 2024, arXiv:2408.00388. [Google Scholar]
  8. Sohl-Dickstein, J.; Weiss, E.; Maheswaranathan, N.; Ganguli, S. Deep Unsupervised Learning Using Nonequilibrium Thermodynamics. In Proceedings of the International Conference on Machine Learning, PMLR, Lille, France, 7–9 July 2015; pp. 2256–2265. [Google Scholar]
  9. Gong, L.Y.; Li, X.J. A Contemporary Survey on Deepfake Detection: Datasets, Algorithms, and Challenges. Electronics 2024, 13, 585. [Google Scholar] [CrossRef]
  10. Shelke, N.A.; Kasana, S.S. A Comprehensive Survey on Passive Techniques for Digital Video Forgery Detection. Multimed. Tools Appl. 2021, 80, 6247–6310. [Google Scholar] [CrossRef]
  11. Mallet, J.; Dave, R.; Seliya, N.; Vanamala, M. Using Deep Learning to Detecting Deepfakes. In Proceedings of the 2022 9th International Conference on Soft Computing & Machine Intelligence (ISCMI), Toronto, ON, Canada, 26–27 November 2022; pp. 1–5. [Google Scholar]
  12. Thing, V.L. Deepfake Detection with Deep Learning: Convolutional Neural Networks Versus Transformers. In Proceedings of the 2023 IEEE International Conference on Cyber Security and Resilience (CSR), Venice, Italy, 31 July–2 August 2023; pp. 246–253. [Google Scholar]
  13. Yan, Z.; Luo, Y.; Lyu, S.; Liu, Q.; Wu, B. Transcending Forgery Specificity with Latent Space Augmentation for Generalizable Deepfake Detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 16–22 June 2024; pp. 8984–8994. [Google Scholar]
  14. Lanzino, R.; Fontana, F.; Diko, A.; Marini, M.R.; Cinque, L. Faster Than Lies: Real-time Deepfake Detection using Binary Neural Networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 16–22 June 2024; pp. 3771–3780. [Google Scholar]
  15. Khan, S.A.; Valles, D. Deepfake Detection Using Transfer Learning. In Proceedings of the 2024 IEEE 15th Annual Ubiquitous Computing, Electronics & Mobile Communication Conference (UEMCON), Yorktown Heights, NY, USA, 17–19 October 2024; pp. 556–562. [Google Scholar]
  16. Chai, L.; Bau, D.; Lim, S.N.; Isola, P. What Makes Fake Images Detectable? Understanding Properties that Generalize. In Proceedings of the Computer Vision—ECCV 2020: 16th European Conference, Glasgow, UK, 23–28 August 2020, Proceedings, Part XXVI; Springer: Berlin/Heidelberg, Germany, 2020; pp. 103–120. [Google Scholar]
  17. Durall, R.; Keuper, M.; Keuper, J. Watch Your Up-Convolution: CNN-Based Generative Deep Neural Networks are Failing to Reproduce Spectral Distributions. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 7890–7899. [Google Scholar]
  18. Tan, C.; Zhao, Y.; Wei, S.; Gu, G.; Liu, P.; Wei, Y. Frequency-Aware Deepfake Detection: Improving Generalizability through Frequency Space Domain Learning. In Proceedings of the AAAI Conference on Artificial Intelligence, Vancouver, BC, Canada, 20–27 February 2024; Volume 38, pp. 5052–5060. [Google Scholar]
  19. Wolter, M.; Blanke, F.; Heese, R.; Garcke, J. Wavelet-Packets for Deepfake Image Analysis and Detection. Mach. Learn. 2022, 111, 4295–4327. [Google Scholar] [CrossRef]
  20. Concas, S.; Perelli, G.; Marcialis, G.L.; Puglisi, G. Tensor-Based Deepfake Detection in Scaled and Compressed Images. In Proceedings of the 2022 IEEE International Conference on Image Processing (ICIP), Bordeaux, France, 16–19 October 2022; pp. 3121–3125. [Google Scholar]
  21. Hu, J.; Liao, X.; Wang, W.; Qin, Z. Detecting Compressed Deepfake Videos in Social Networks Using Frame-Temporality Two-Stream Convolutional Network. IEEE Trans. Circuits Syst. Video Technol. 2021, 32, 1089–1102. [Google Scholar] [CrossRef]
  22. Gao, J.; Xia, Z.; Marcialis, G.L.; Dang, C.; Dai, J.; Feng, X. DeepFake Detection Based on High-Frequency Enhancement Network for Highly Compressed Content. Expert Syst. Appl. 2024, 249, 123732. [Google Scholar] [CrossRef]
  23. Sun, X.; Wu, B.; Chen, W. Identifying Invariant Texture Violation for Robust Deepfake Detection. arXiv 2020, arXiv:2012.10580. [Google Scholar]
  24. Patil, K.; Kale, S.; Dhokey, J.; Gulhane, A. Deepfake Detection Using Biological Features: A Survey. arXiv 2023, arXiv:2301.05819. [Google Scholar]
  25. Hernandez-Ortega, J.; Tolosana, R.; Fierrez, J.; Morales, A. Deepfakeson-Phys: Deepfakes Detection Based on Heart Rate Estimation. arXiv 2020, arXiv:2010.00400. [Google Scholar]
  26. Mittal, T.; Bhattacharya, U.; Chandra, R.; Bera, A.; Manocha, D. Emotions Do not Lie: An Audio-Visual Deepfake Detection Method Using Affective Cues. In Proceedings of the 28th ACM International Conference on Multimedia, Seattle, WA, USA, 12–16 October 2020; pp. 2823–2832. [Google Scholar]
  27. Li, Y.; Chang, M.C.; Lyu, S. In Ictu Oculi: Exposing AI-Generated Fake Face Videos by Detecting Eye Blinking. arXiv 2018, arXiv:1806.02877. [Google Scholar]
  28. Concas, S.; La Cava, S.M.; Casula, R.; Orrù, G.; Puglisi, G.; Marcialis, G.L. Quality-based Artifact Modeling for Facial Deepfake Detection in Videos. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 16–22 June 2024; pp. 3845–3854. [Google Scholar]
  29. Gao, J.; Micheletto, M.; Orrù, G.; Concas, S.; Feng, X.; Marcialis, G.L.; Roli, F. Texture and Artifact Decomposition for Improving Generalization in Deep-Learning-Based DeepFake Detection. Eng. Appl. Artif. Intell. 2024, 133, 108450. [Google Scholar] [CrossRef]
  30. Yu, Y.; Ni, R.; Yang, S.; Ni, Y.; Zhao, Y.; Kot, A.C. Mining Generalized Multi-timescale Inconsistency for Detecting Deepfake Videos. Int. J. Comput. Vis. 2024, 1–17. [Google Scholar] [CrossRef]
  31. Amin, M.A.; Hu, Y.; Hu, J. Analyzing Temporal Coherence for Deepfake Video Detection. Electron. Res. Arch. 2024, 32, 2621–2641. [Google Scholar] [CrossRef]
  32. Zhou, Y.; Lim, S.N. Joint Audio-Visual Deepfake Detection. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 10–17 October 2021; pp. 14800–14809. [Google Scholar]
  33. Hong, Y.; Zhang, J. Wildfake: A large-scale challenging dataset for ai-generated images detection. arXiv 2024, arXiv:2402.11843. [Google Scholar]
  34. Rahman, M.A.; Paul, B.; Sarker, N.H.; Hakim, Z.I.A.; Fattah, S.A. Artifact: A large-scale dataset with artificial and factual images for generalizable and robust synthetic image detection. In Proceedings of the IEEE International Conference on Image Processing (ICIP), Kuala Lumpur, Malaysia, 8–11 October 2023. [Google Scholar]
  35. Cheng, H.; Guo, Y.; Wang, T.; Nie, L.; Kankanhalli, M. Diffusion facial forgery detection. In Proceedings of the ACM International Conference on Multimedia, Bari, Italy, 15–18 April 2024. [Google Scholar]
  36. Sauer, A.; Schwarz, K.; Geiger, A. Stylegan-xl: Scaling stylegan to large diverse datasets. In Proceedings of the ACM SIGGRAPH, Vancouver, BC, Canada, 7–11 August 2022. [Google Scholar]
  37. Karras, T.; Laine, S.; Aila, T. A Style-Based Generator Architecture for Generative Adversarial Networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019. [Google Scholar]
  38. Karras, T.; Laine, S.; Aittala, M.; Hellsten, J.; Lehtinen, J.; Aila, T. Analyzing and improving the image quality of stylegan. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020. [Google Scholar]
  39. Todisco, M.; Wang, X.; Vestman, V.; Sahidullah, M.; Delgado, H.; Nautsch, A.; Yamagishi, J.; Evans, N.; Kinnunen, T.; Lee, K.A. ASVspoof 2019: Future horizons in spoofed and fake audio detection. In Proceedings of the International Speech Communication Association Conference (INTERSPEECH), Graz, Austria, 15–19 September 2019. [Google Scholar]
  40. Wang, X.; Yamagishi, J.; Todisco, M.; Delgado, H.; Nautsch, A.; Evans, N.; Sahidullah, M.; Vestman, V.; Kinnunen, T.; Lee, K.A.; et al. ASVspoof 2019: A large-scale public database of synthesized, converted and replayed speech. Comput. Speech Lang. 2020, 64, 101114. [Google Scholar] [CrossRef]
  41. Yamagishi, J.; Wang, X.; Todisco, M.; Sahidullah, M.; Patino, J.; Nautsch, A.; Liu, X.; Lee, K.A.; Kinnunen, T.; Evans, N.; et al. ASVspoof 2021: Accelerating progress in spoofed and deepfake speech detection. In Proceedings of the Automatic Speaker Verification and Spoofing Countermeasures Challenge, Online, 16 September 2021. [Google Scholar]
  42. Salvi, D.; Hosler, B.; Bestagini, P.; Stamm, M.C.; Tubaro, S. TIMIT-TTS: A Text-to-Speech Dataset for Multimodal Synthetic Media Detection. IEEE Access 2023, 11, 50851–50866. [Google Scholar] [CrossRef]
  43. Müller, N.M.; Czempin, P.; Dieckmann, F.; Froghyar, A.; Böttinger, K. Does audio deepfake detection generalize? In Proceedings of the International Speech Communication Association Conference (INTERSPEECH), Incheon, Republic of Korea, 18–22 September 2022.
  44. Reimao, R.; Tzerpos, V. FOR: A dataset for synthetic speech detection. In Proceedings of the IEEE International Conference on Speech Technology and Human–Computer Dialogue (SpeD), Timisoara, Romania, 10–12 October 2019. [Google Scholar]
  45. Bhagtani, K.; Yadav, A.K.S.; Bestagini, P.; Delp, E.J. Are Recent Deepfake Speech Generators Detectable? In Proceedings of the ACM Workshop on Information Hiding and Multimedia Security, Baiona, Spain, 24–26 June 2024.
  46. Ito, K.; Johns, L. The LJSpeech Dataset. 2017. Available online: https://keithito.com/LJ-Speech-Dataset/ (accessed on 15 January 2025).
  47. Panayotov, V.; Chen, G.; Povey, D.; Khudanpur, S. Librispeech: An ASR corpus based on public domain audio books. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Brisbane, Australia, 19–24 April 2015. [Google Scholar]
  48. Korshunov, P.; Marcel, S. Deepfakes: A new threat to face recognition? Assessment and detection. arXiv 2018, arXiv:1812.08685. [Google Scholar]
  49. Sanderson, C. The VidTIMIT database. Technical Report, IDIAP. 2002. Available online: https://www.google.com/url?sa=t&source=web&rct=j&opi=89978449&url=https://publications.idiap.ch/downloads/reports/2002/com02-06.pdf&ved=2ahUKEwid9vKeueWLAxVuvokEHTziMbQQFnoECBUQAQ&usg=AOvVaw0kiW0MZbBb8yeSzteosaGP (accessed on 15 January 2025).
  50. Korshunova, I.; Shi, W.; Dambre, J.; Theis, L. Fast face-swap using convolutional neural networks. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017. [Google Scholar]
  51. Rossler, A.; Cozzolino, D.; Verdoliva, L.; Riess, C.; Thies, J.; Nießner, M. FaceForensics++: Learning to Detect Manipulated Facial Images. In Proceedings of the IEEE/CVF International Conference on Computer Vision (CVPR), Long Beach, CA, USA, 15–20 June 2019. [Google Scholar]
  52. Dolhansky, B.; Bitton, J.; Pflaum, B.; Lu, J.; Howes, R.; Wang, M.; Ferrer, C.C. The deepfake detection challenge (DFDC) dataset. arXiv 2020, arXiv:2006.07397. [Google Scholar]
  53. Khalid, H.; Tariq, S.; Kim, M.; Woo, S.S. FakeAVCeleb: A Novel Audio-Video Multimodal Deepfake Dataset. In Proceedings of the Conference on Neural Information Processing Systems Datasets and Benchmarks Track (Round 2), Virtual, 6 December 2021. [Google Scholar]
  54. Chung, J.S.; Nagrani, A.; Zisserman, A. VoxCeleb2: Deep Speaker Recognition. In Proceedings of the INTERSPEECH, Hyderabad, India, 2–6 September 2018. [Google Scholar]
  55. Nirkin, Y.; Keller, Y.; Hassner, T. Fsgan: Subject agnostic face swapping and reenactment. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Republic of Korea, 27 October–2 November 2019. [Google Scholar]
  56. Jia, Y.; Zhang, Y.; Weiss, R.; Wang, Q.; Shen, J.; Ren, F.; Nguyen, P.; Pang, R.; Lopez Moreno, I.; Wu, Y.; et al. Transfer learning from speaker verification to multispeaker text-to-speech synthesis. arXiv 2018, arXiv:1806.04558. [Google Scholar]
  57. Prajwal, K.; Mukhopadhyay, R.; Namboodiri, V.P.; Jawahar, C. A lip sync expert is all you need for speech to lip generation in the wild. In Proceedings of the ACM International Conference on Multimedia, Seattle, WA, USA, 12–16 October 2020. [Google Scholar]
  58. Barrington, S.; Bohacek, M.; Farid, H. DeepSpeak Dataset v1.0. arXiv 2024, arXiv:2408.05366. [Google Scholar]
  59. Almutairi, Z.; Elgibreen, H. A Review of Modern Audio Deepfake Detection Methods: Challenges and Future Directions. Algorithms 2022, 15, 155. [Google Scholar] [CrossRef]
  60. Aleixo, E.L.; Colonna, J.G.; Cristo, M.; Fernandes, E. Catastrophic Forgetting in Deep Learning: A Comprehensive Taxonomy. J. Braz. Comput. Soc. 2024, 30, 175–211. [Google Scholar] [CrossRef]
  61. De Lange, M.; Aljundi, R.; Masana, M.; Parisot, S.; Jia, X.; Leonardis, A.; Slabaugh, G.; Tuytelaars, T. A Continual Learning Survey: Defying Forgetting in Classification Tasks. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 44, 3366–3385. [Google Scholar]
  62. Wang, L.; Zhang, X.; Su, H.; Zhu, J. A Comprehensive Survey of Continual Learning: Theory, Method and Application. IEEE Trans. Pattern Anal. Mach. Intell. 2024, 46, 5362–5383. [Google Scholar] [CrossRef]
  63. Wani, T.M.; Qadri, S.A.A.; Wani, F.A.; Amerini, I. Navigating the Soundscape of Deception: A Comprehensive Survey on Audio Deepfake Generation, Detection, and Future Horizons. Found. Trends® Priv. Secur. 2024, 6, 153–345. [Google Scholar] [CrossRef]
  64. Neves, D.E.; do Patrocínio Júnior, Z.K.G.; Ishitani, L. Advances and Challenges in Learning from Experience Replay. Artif. Intell. Rev. 2024, 58, 54. [Google Scholar] [CrossRef]
  65. Kang, S.; Shi, Z.; Zhang, X. Poisoning Generative Replay in Continual Learning to Promote Forgetting. In Proceedings of the International Conference on Machine Learning, PMLR, Honolulu, HI, USA, 23–29 July 2023; pp. 15769–15785. [Google Scholar]
  66. Ma, H.; Yi, J.; Tao, J.; Bai, Y.; Tian, Z.; Wang, C. Continual Learning for Fake Audio Detection. arXiv 2021, arXiv:2104.07286. [Google Scholar]
  67. Nguyen Le, T.D.; Teh, K.K.; Dat Tran, H. Continuous Learning of Transformer-based Audio Deepfake Detection. arXiv 2024, arXiv:2409.05924. [Google Scholar]
  68. Zhang, Y.; Lin, W.; Xu, J. Joint Audio-Visual Attention with Contrastive Learning for More General Deepfake Detection. ACM Trans. Multimed. Comput. Commun. Appl. 2024, 20, 1–23. [Google Scholar] [CrossRef]
  69. Dong, F.; Tang, Q.; Bai, Y.; Wang, Z. Advancing Continual Learning for Robust Deepfake Audio Classification. arXiv 2024, arXiv:2407.10108. [Google Scholar]
  70. Li, X.; Yang, Z.; Guo, P.; Cheng, J. An Intelligent Transient Stability Assessment Framework with Continual Learning Ability. IEEE Trans. Ind. Informatics 2021, 17, 8131–8141. [Google Scholar] [CrossRef]
  71. Zhang, X.; Yi, J.; Wang, C.; Zhang, C.Y.; Zeng, S.; Tao, J. What to Remember: Self-Adaptive Continual Learning for Audio DeepFake Detection. In Proceedings of the AAAI Conference on Artificial Intelligence, Vancouver, BC, Canada, February 20–27 2024; Volume 38, pp. 19569–19577. [Google Scholar]
  72. Wani, T.M.; Amerini, I. Audio Deepfake Detection: A Continual Approach with Feature Distillation and Dynamic Class Rebalancing. In International Conference on Pattern Recognition; Antonacopoulos, A., Chaudhuri, S., Chellappa, R., Liu, C.L., Bhattacharya, S., Pal, U., Eds.; Springer: Cham, Switzerland, 2025; pp. 211–227. [Google Scholar]
  73. Zhang, X.; Yi, J.; Tao, J. EVDA: Evolving Deepfake Audio Detection Continual Learning Benchmark. arXiv 2024, arXiv:2405.08596. [Google Scholar]
  74. Salvi, D.; Negroni, V.; Bondi, L.; Bestagini, P.; Tubaro, S. Freeze and Learn: Continual Learning with Selective Freezing for Speech Deepfake Detection. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Hyderabad, India, 6–11 April 2025. [Google Scholar]
  75. Hydara, E.; Kikuchi, M.; Ozono, T. Empirical Assessment of Deepfake Detection: Advancing Judicial Evidence Verification through Artificial Intelligence. IEEE Access 2024, 12, 151188–151203. [Google Scholar] [CrossRef]
  76. Rudin, C. Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead. Nat. Mach. Intell. 2019, 1, 206–215. [Google Scholar] [CrossRef] [PubMed]
  77. Solanke, A.A. Explainable digital forensics AI: Towards mitigating distrust in AI-based digital forensics analysis using interpretable models. Forensic Sci. Int. Digit. Investig. 2022, 42, 301403. [Google Scholar] [CrossRef]
  78. Albahri, A.S.; Duhaim, A.M.; Fadhel, M.A.; Alnoor, A.; Baqer, N.S.; Alzubaidi, L.; Albahri, O.S.; Alamoodi, A.H.; Bai, J.; Salhi, A.; et al. A systematic review of trustworthy and explainable artificial intelligence in healthcare: Assessment of quality, bias risk, and data fusion. Inf. Fusion 2023, 96, 156–191. [Google Scholar] [CrossRef]
  79. National Research Council; Policy and Global Affairs; Division on Engineering and Physical Sciences; Committee on Science, Technology, and Law; Committee on Applied and Theoretical Statistics; Committee on Identifying the Needs of the Forensic Sciences Community. Strengthening Forensic Science in the United States: A path forward; National Academies Press: Washington, DC, USA, 2009. [Google Scholar]
  80. Hong, S.R.; Hullman, J.; Bertini, E. Human factors in model interpretability: Industry practices, challenges, and needs. Proc. ACM Hum.-Comput. Interact. 2020, 4, 1–26. [Google Scholar] [CrossRef]
  81. Jain, A. Enhancing Forensic Analysis of Digital Evidence Using Machine Learning: Techniques, Applications, and Challenges. Int. J. Innov. Res. Multidiscip. Perspect. Stud. (IJIRMPS) 2024, 18, 1–8. [Google Scholar]
  82. Richmond, K.M.; Muddamsetty, S.M.; Gammeltoft-Hansen, T.; Olsen, H.P.; Moeslund, T.B. Explainable AI and law: An evidential survey. Digit. Soc. 2024, 3, 1. [Google Scholar] [CrossRef]
  83. Sletten, A.H.d.S. Leveraging Explainability Maps for Group-Unsupervised Robustness to Spurious Correlations. Master’s Thesis, UiT Norges Arktiske Universitet, Tromsø, Norway, 2023. [Google Scholar]
  84. Nailwal, S.; Singhal, S.; Singh, N.T.; Raza, A. Deepfake Detection: A Multi-Algorithmic and Multi-Modal Approach for Robust Detection and Analysis. In Proceedings of the 2023 International Conference on Research Methodologies in Knowledge Management, Artificial Intelligence and Telecommunication Engineering (RMKMATE), Chennai, India, 1–2 November 2023; pp. 1–8. [Google Scholar]
  85. Lim, S.Y.; Chae, D.K.; Lee, S.C. Detecting DeepFake Voice Using Explainable Deep Learning Techniques. Appl. Sci. 2022, 12, 3926. [Google Scholar] [CrossRef]
  86. Yu, N.; Chen, L.; Leng, T.; Chen, Z.; Yi, X. An Explainable DeepFake of Speech Detection Method with Spectrograms and Waveforms. J. Inf. Secur. Appl. 2024, 81, 103720. [Google Scholar] [CrossRef]
  87. Channing, G.; Sock, J.; Clark, R.; Torr, P.; de Witt, C.S. Toward Robust Real-World Audio Deepfake Detection: Closing the Explainability Gap. arXiv 2024, arXiv:2410.07436. [Google Scholar]
  88. Ilyas, H.; Javed, A.; Malik, K.M. ConvNext-PNet: An Interpretable and Explainable Deep-Learning Model for DeepFakes Detection. In Proceedings of the 2024 IEEE International Joint Conference on Biometrics (IJCB), Buffalo, NY, USA, 15–18 September 2024; pp. 1–9. [Google Scholar]
  89. Bisogni, C.; Loia, V.; Nappi, M.; Pero, C. Acoustic Features Analysis for Explainable Machine Learning-Based Audio Spoofing Detection. Comput. Vis. Image Underst. 2024, 249, 104145. [Google Scholar] [CrossRef]
  90. Guarnera, L.; Giudice, O.; Nießner, M.; Battiato, S. On the Exploitation of Deepfake Model Recognition. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshop, New Orleans, LA, USA, 19–20 June 2022; pp. 61–70. [Google Scholar]
  91. Pontorno, O.; Guarnera, L.; Battiato, S. On the Exploitation of DCT-Traces in the Generative-AI Domain. In Proceedings of the 2024 IEEE International Conference on Image Processing (ICIP), Abu Dhabi, United Arab Emirates, 27–30 October 2024; pp. 3806–3812. [Google Scholar] [CrossRef]
  92. Asnani, V.; Yin, X.; Hassner, T.; Liu, X. Reverse engineering of generative models: Inferring model hyperparameters from generated images. IEEE Trans. Pattern Anal. Mach. Intell. 2023, 45, 15477–15493. [Google Scholar] [CrossRef] [PubMed]
  93. Guarnera, L.; Giudice, O.; Battiato, S. Deepfake Detection by Analyzing Convolutional Traces. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, Seattle, WA, USA, 14–19 June 2020; pp. 666–667. [Google Scholar]
  94. Guarnera, L.; Giudice, O.; Battiato, S. Fighting Deepfake by Exposing the Convolutional Traces on Images. IEEE Access 2020, 8, 165085–165098. [Google Scholar] [CrossRef]
  95. Giudice, O.; Guarnera, L.; Battiato, S. Fighting Deepfakes by Detecting GAN DCT Anomalies. J. Imaging 2021, 7, 128. [Google Scholar] [CrossRef]
  96. Guarnera, L.; Giudice, O.; Battiato, S. Mastering Deepfake Detection: A Cutting-Edge Approach to Distinguish GAN and Diffusion-Model Images. Acm Trans. Multimed. Comput. Commun. Appl. 2024, 20, 1–24. [Google Scholar] [CrossRef]
  97. Pontorno, O.; Guarnera, L.; Battiato, S. DeepFeatureX Net: Deep Features eXtractors based Network for Discriminating Synthetic from Real Images. In International Conference on Pattern Recognition; Springer: Cham, Switzerland, 2025; pp. 177–193. [Google Scholar]
  98. Salvi, D.; Bestagini, P.; Tubaro, S. Exploring the synthetic speech attribution problem through data-driven detectors. In Proceedings of the IEEE International Workshop on Information Forensics and Security (WIFS), Rome, Italy, 2–5 December 2024. [Google Scholar]
  99. Yu, N.; Davis, L.S.; Fritz, M. Attributing Fake Images to GANs: Learning and Analyzing GAN Fingerprints. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 7556–7566. [Google Scholar]
  100. Yu, N.; Skripniuk, V.; Abdelnabi, S.; Fritz, M. Artificial Fingerprinting for Generative Models: Rooting Deepfake Attribution in Training Data. In Proceedings of the 2021 IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, BC, Canada, 11–17 October 2021; pp. 14448–14457. [Google Scholar] [CrossRef]
  101. Yang, T.; Huang, Z.; Cao, J.; Li, L.; Li, X. DeepFake Network Architecture Attribution. In Proceedings of the AAAI Conference on Artificial Intelligence, Online, 22 February–1 March 2022; Volume 36, pp. 4662–4670. [Google Scholar]
  102. Sun, Z.; Chen, S.; Yao, T.; Yin, B.; Yi, R.; Ding, S.; Ma, L. Contrastive Pseudo Learning for Open-World DeepFake Attribution. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Paris, France, 2–3 October 2023; pp. 20882–20892. [Google Scholar]
  103. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  104. Liu, E.Y.; Guo, Z.; Zhang, X.; Jojic, V.; Wang, W. Metric Learning from Relative Comparisons by Minimizing Squared Residual. In Proceedings of the 2012 IEEE 12th International Conference on Data Mining, Brussels, Belgium, 10 December 2012; pp. 978–983. [Google Scholar]
  105. Karras, T.; Aittala, M.; Hellsten, J.; Laine, S.; Lehtinen, J.; Aila, T. Training Generative Adversarial Networks with Limited Data. In Advances in Neural Information Processing Systems; Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H., Eds.; Curran Associates, Inc.: San Jose, CA, USA, 2020; Volume 33, pp. 12104–12114. [Google Scholar]
  106. Sun, Z.; Chen, S.; Yao, T.; Yi, R.; Ding, S.; Ma, L. Rethinking Open-World DeepFake Attribution with Multi-perspective Sensory Learning. Int. J. Comput. Vis. 2025, 133, 628–651. [Google Scholar] [CrossRef]
  107. Leotta, R.; Giudice, O.; Guarnera, L.; Battiato, S. Not with My Name! Inferring Artists’ Names of Input Strings Employed by Diffusion Models. In International Conference on Image Analysis and Processing; Springer: Cham, Switzerland, 2023; pp. 364–375. [Google Scholar]
  108. Huang, Z.; Li, B.; Cai, Y.; Wang, R.; Guo, S.; Fang, L.; Chen, J.; Wang, L. What Can Discriminator Do? Towards Box-Free Ownership Verification of Generative Adversarial Networks. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Paris, France, 2–3 October 2023; pp. 5009–5019. [Google Scholar]
  109. Guarnera, L.; Guarnera, F.; Ortis, A.; Battiato, S.; Puglisi, G. Evasion Attack on Deepfake Detection via DCT Trace Manipulation. In International Conference on Pattern Recognition; Springer: Berlin/Heidelberg, Germany, 2024. [Google Scholar]
  110. Cuccovillo, L.; Papastergiopoulos, C.; Vafeiadis, A.; Yaroshchuk, A.; Aichroth, P.; Votis, K.; Tzovaras, D. Open Challenges in Synthetic Speech Detection. In Proceedings of the IEEE International Workshop on Information Forensics and Security (WIFS), Rome, Italy, 2–5 December 2024. [Google Scholar]
  111. Wang, Z.; Wei, G.; He, Q. Channel Pattern Noise-Based Playback Attack Detection Algorithm for Speaker Recognition. In Proceedings of the IEEE International Conference on Machine Learning and Cybernetics (ICMLC), Guilin, China, 10–13 July 2011. [Google Scholar]
  112. Malik, H. Securing Voice-Driven Interfaces Against Fake (Cloned) Audio Attacks. In Proceedings of the IEEE Conference on Multimedia Information Processing and Retrieval (MIPR), San Jose, CA, USA, 28– 30 March 2019. [Google Scholar]
  113. Borrelli, C.; Bestagini, P.; Antonacci, F.; Sarti, A.; Tubaro, S. Synthetic Speech Detection Through Short-Term and Long-Term Prediction Traces. EURASIP J. Inf. Secur. 2021, 2021, 2. [Google Scholar] [CrossRef]
  114. Mari, D.; Salvi, D.; Bestagini, P.; Milani, S. All-for-One and One-For-All: Deep learning-based feature fusion for Synthetic Speech Detection. In Proceedings of the European Conference on Machine Learning and Knowledge Discovery in Databases Workshops (ECML PKDD), Turin, Italy, 18–22 September 2023. [Google Scholar]
  115. Tak, H.; Patino, J.; Todisco, M.; Nautsch, A.; Evans, N.; Larcher, A. End-to-end anti-spoofing with RawNet2. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Toronto, ON, Canada, 6–11 June 2021. [Google Scholar]
  116. Jung, J.w.; Heo, H.S.; Tak, H.; Shim, H.j.; Chung, J.S.; Lee, B.J.; Yu, H.J.; Evans, N. AASIST: Audio Anti-Spoofing using Integrated Spectro-Temporal Graph Attention Networks. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Singapore, 22–27 May 2022. [Google Scholar]
  117. Hamza, A.; Javed, A.R.; Iqbal, F.; Kryvinska, N.; Almadhor, A.S.; Jalil, Z.; Borghol, R. Deepfake Audio Detection via MFCC Features using Machine Learning. IEEE Access 2022, 10, 134018–134028. [Google Scholar] [CrossRef]
  118. Salvi, D.; Bestagini, P.; Tubaro, S. Reliability Estimation for Synthetic Speech Detection. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Rhodes Island, Greece, 4–10 June 2023. [Google Scholar]
  119. Yadav, A.K.S.; Bhagtani, K.; Salvi, D.; Bestagini, P.; Delp, E.J. FairSSD: Understanding Bias in Synthetic Speech Detectors. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Seattle, WA, USA, 17–18 June 2024. [Google Scholar]
  120. Negroni, V.; Salvi, D.; Mezza, A.I.; Bestagini, P.; Tubaro, S. Leveraging Mixture of Experts for Improved Speech Deepfake Detection. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Hyderabad, India, 6–11 April 2025. [Google Scholar]
  121. Sahidullah, M.; Kinnunen, T.; Hanilçi, C. A Comparison of Features for Synthetic Speech Detection. In Proceedings of the International Speech Communication Association Conference (INTERSPEECH), Dresden, Germany, 6–10 September 2015. [Google Scholar]
  122. Conti, E.; Salvi, D.; Borrelli, C.; Hosler, B.; Bestagini, P.; Antonacci, F.; Sarti, A.; Stamm, M.C.; Tubaro, S. Deepfake Speech Detection Through Emotion Recognition: A Semantic Approach. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Singapore, 22–27 May 2022. [Google Scholar]
  123. Attorresi, L.; Salvi, D.; Borrelli, C.; Bestagini, P.; Tubaro, S. Combining Automatic Speaker Verification and Prosody Analysis for Synthetic Speech Detection. In Proceedings of the International Conference on Pattern Recognition (ICPR), Montreal, QC, Canada, 21–25 August 2022. [Google Scholar]
  124. Ge, W.; Todisco, M.; Evans, N. Explainable DeepFake and Spoofing Detection: An Attack Analysis Using SHapley Additive ExPlanations. In Proceedings of the International Speech Communication Association Conference (INTERSPEECH), Incheon, Republic of Korea, 18–22 September 2022. [Google Scholar]
  125. Salvi, D.; Bestagini, P.; Tubaro, S. Towards Frequency Band Explainability in Synthetic Speech Detection. In Proceedings of the European Signal Processing Conference (EUSIPCO), Helsinki, Fnland, 4–8 September 2023. [Google Scholar]
  126. Salvi, D.; Balcha, T.S.; Bestagini, P.; Tubaro, S. Listening Between the Lines: Synthetic Speech Detection Disregarding Verbal Content. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing Workshops (ICASSPW), Seoul, Republic of Korea, 14–19 April 2024. [Google Scholar]
  127. Gohari, M.; Salvi, D.; Bestagini, P.; Adami, N. Audio Features Investigation for Singing Voice Deepfake Detection. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Hyderabad, India, 6–11 April 2025. [Google Scholar]
  128. Salvi, D.; Yadav, A.K.S.; Bhagtani, K.; Negroni, V.; Bestagini, P.; Delp, E.J. Comparative Analysis of ASR Methods for Speech Deepfake Detection. In Proceedings of the 58th Asilomar Conference on Signals, Systems, and Computers, Pacific Grove, CA, USA, 27–30 October 2024. [Google Scholar]
  129. Kawa, P.; Plata, M.; Czuba, M.; Szymanski, P.; Syga, P. Improved DeepFake Detection Using Whisper Features. In Proceedings of the Conference of the International Speech Communication Association (INTERSPEECH), Dublin, Ireland, 20–24 August 2023. [Google Scholar]
  130. Guo, Y.; Huang, H.; Chen, X.; Zhao, H.; Wang, Y. Audio Deepfake Detection With Self-Supervised Wavlm Furthermore, Multi-Fusion Attentive Classifier. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Seoul, Republic of Korea, 14–19 April 2024. [Google Scholar]
  131. Salvi, D.; Bestagini, P.; Tubaro, S. Synthetic speech detection through audio folding. In Proceedings of the ACM International Workshop on Multimedia AI against Disinformation (MAD), Thessaloniki, Greece, 12 June 2023. [Google Scholar]
  132. Yang, X.; Li, Y.; Lyu, S. Exposing DeepFakes Using Inconsistent Head Poses. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Brighton, UK, 12–17 May 2019. [Google Scholar]
  133. Matern, F.; Riess, C.; Stamminger, M. Exploiting Visual Artifacts to Expose Deepfakes and Face Manipulations. In Proceedings of the IEEE Winter Applications of Computer Vision Workshops (WACVW), Waikoloa Village, HI, USA, 7–11 January 2019. [Google Scholar] [CrossRef]
  134. Amerini, I.; Galteri, L.; Caldelli, R.; Del Bimbo, A. DeepFake Video Detection Through Optical Flow-Based CNN. In Proceedings of the IEEE/CVF International Conference on Computer Vision (CVPR), Long Beach, CA, USA, 15–20 June 2019. [Google Scholar]
  135. Koopman, M.; Rodriguez, A.M.; Geradts, Z. Detection of DeepFake Video Manipulation. In Proceedings of the Irish Machine Vision and Image Processing Conference (IMVIP), Belfast, Northern Ireland, 29–31 August 2018. [Google Scholar]
  136. Afchar, D.; Nozick, V.; Yamagishi, J.; Echizen, I. MesoNet: A Compact Facial Video Forgery Detection Network. In Proceedings of the IEEE International Workshop on Information Forensics and Security (WIFS), Hong Kong, China, 11–13 December 2018. [Google Scholar]
  137. Güera, D.; Delp, E.J. DeepFake Video Detection Using Recurrent Neural Networks. In Proceedings of the IEEE International Conference on Advanced Video and Signal Based Surveillance (AVSS), Auckland, New Zealand, 27–30 November 2018. [Google Scholar]
  138. Bonettini, N.; Cannas, E.D.; Mandelli, S.; Bondi, L.; Bestagini, P.; Tubaro, S. Video Face Manipulation Detection Through Ensemble of CNNs. In Proceedings of the International Conference on Pattern Recognition (ICPR), Milan, Italy, 10–15 January 2021. [Google Scholar]
  139. Zhao, H.; Zhou, W.; Chen, D.; Wei, T.; Zhang, W.; Yu, N. Multi-Attentional DeepFake Detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 19–25 June 2021. [Google Scholar]
  140. Liu, H.; Li, X.; Zhou, W.; Chen, Y.; He, Y.; Xue, H.; Zhang, W.; Yu, N. Spatial-Phase Shallow Learning: Rethinking Face Forgery Detection in Frequency Domain. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021. [Google Scholar]
  141. Ciftci, U.A.; Demir, I.; Yin, L. FakeCatcher: Detection of Synthetic Portrait Videos Using Biological Signals. IEEE Trans. Pattern Anal. Mach. Intell. (TPAMI), 2020; online ahead of print. [Google Scholar]
  142. Hosler, B.; Salvi, D.; Murray, A.; Antonacci, F.; Bestagini, P.; Tubaro, S.; Stamm, M.C. Do Deepfakes Feel Emotions? A Semantic Approach to Detecting Deepfakes via Emotional Inconsistencies. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021. [Google Scholar]
  143. Lomnitz, M.; Hampel-Arias, Z.; Sandesara, V.; Hu, S. Multimodal Approach for DeepFake Detection. In Proceedings of the IEEE Applied Imagery Pattern Recognition Workshop (AIPR), Washington DC, DC, USA, 13–15 October 2020. [Google Scholar]
  144. Khalid, H.; Kim, M.; Tariq, S.; Woo, S.S. Evaluation of an Audio-Video Multimodal DeepFake Dataset Using Unimodal and Multimodal Detectors. In Proceedings of the Workshop on Synthetic Multimedia-Audiovisual DeepFake Generation and Detection, Virtual, 24 October 2021. [Google Scholar]
  145. Korshunov, P.; Halstead, M.; Castan, D.; Graciarena, M.; McLaren, M.; Burns, B.; Lawson, A.; Marcel, S. Tampered Speaker Inconsistency Detection with Phonetically Aware Audio-Visual Features. In Proceedings of the International Conference on Machine Learning (ICML), Long Beach, CA, USA, 10–15 June 2019. [Google Scholar]
  146. Agarwal, S.; Farid, H.; El-Gaaly, T.; Lim, S.N. Detecting Deep-Fake Videos from Appearance and Behavior. In Proceedings of the IEEE International Workshop on Information Forensics and Security (WIFS), New York, NY, USA, 6–11 December 2020. [Google Scholar]
  147. Agarwal, S.; Hu, L.; Ng, E.; Darrell, T.; Li, H.; Rohrbach, A. Watch Those Words: Video Falsification Detection Using Word-Conditioned Facial Motion. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision (WACV), Waikoloa, HI, USA, 2–7 January 2023. [Google Scholar]
  148. Oorloff, T.; Koppisetti, S.; Bonettini, N.; Solanki, D.; Colman, B.; Yacoob, Y.; Shahriyari, A.; Bharaj, G. AVFF: Audio-Visual Feature Fusion for Video Deepfake Detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 17–21 June 2024. [Google Scholar]
  149. Salvi, D.; Liu, H.; Mandelli, S.; Bestagini, P.; Zhou, W.; Zhang, W.; Tubaro, S. A Robust Approach to Multimodal Deepfake Detection. J. Imaging 2023, 9, 122. [Google Scholar] [CrossRef] [PubMed]
  150. Maiano, L.; Benova, A.; Papa, L.; Stockner, M.; Marchetti, M.; Convertino, G.; Mazzoni, G.; Amerini, I. Human Versus Machine: A Comparative Analysis in Detecting Artificial Intelligence-Generated Images. IEEE Secur. Priv. 2024, 22, 77–86. [Google Scholar] [CrossRef]
  151. Barni, M.; Campisi, P.; Delp, E.J.; Doërr, G.; Fridrich, J.; Memon, N.; Pérez-González, F.; Rocha, A.; Verdoliva, L.; Wu, M. Information Forensics and Security: A Quarter-Century-Long Journey. IEEE Signal Process. Mag. 2023, 40, 67–79. [Google Scholar] [CrossRef]
  152. Kirkpatrick, J.; Pascanu, R.; Rabinowitz, N.; Veness, J.; Desjardins, G.; Rusu, A.A.; Milan, K.; Quan, J.; Ramalho, T.; Grabska-Barwinska, A.; et al. Overcoming Catastrophic Forgetting in Neural Networks. Proc. Natl. Acad. Sci. USA 2017, 114, 3521–3526. [Google Scholar] [CrossRef] [PubMed]
  153. Şahin, E.; Arslan, N.N.; Özdemir, D. Unlocking the Black Box: An In-Depth Review on Interpretability, Explainability, and Reliability in Deep Learning. Neural Comput. Appl. 2024, 37, 859–965. [Google Scholar]
  154. Maheshwari, R.U.; Paulchamy, B. Securing Online Integrity: A Hybrid Approach to DeepFake Detection and Removal Using Explainable AI and Adversarial Robustness Training. Automatika 2024, 65, 1517–1532. [Google Scholar] [CrossRef]
  155. Khoo, B.; Phan, R.C.W.; Lim, C.H. DeepFake Attribution: On the Source Identification of Artificially Generated Images. Wiley Interdiscip. Rev. Data Min. Knowl. Discov. 2022, 12, e1438. [Google Scholar] [CrossRef]
  156. Paleyes, A.; Urma, R.G.; Lawrence, N.D. Challenges in Deploying Machine Learning: A Survey of Case Studies. ACM Comput. Surv. 2022, 55, 1–29. [Google Scholar] [CrossRef]
  157. Semola, R.; Lomonaco, V.; Bacciu, D. Continual-Learning-as-a-Service (claas): On-demand Efficient Adaptation of Predictive Models. arXiv 2022, arXiv:2206.06957. [Google Scholar]
  158. Boato, G.; Pasquini, C.; Stefani, A.L.; Verde, S.; Miorandi, D. TrueFace: A Dataset for the Detection of Synthetic Face Images from Social Networks. In Proceedings of the 2022 IEEE International Joint Conference on Biometrics (IJCB), Abu Dhabi, United Arab Emirates, 10–13 October 2022; pp. 1–7. [Google Scholar] [CrossRef]
  159. Marra, F.; Gragnaniello, D.; Cozzolino, D.; Verdoliva, L. Detection of GAN-Generated Fake Images over Social Networks. In Proceedings of the 2018 IEEE Conference on Multimedia Information Processing and Retrieval (MIPR), Miami, FL, USA, 10–12 April 2018; pp. 384–389. [Google Scholar]
  160. Pasquini, C.; Amerini, I.; Boato, G. Media Forensics on Social Media Platforms: A Survey. EURASIP J. Inf. Secur. 2021, 2021, 4. [Google Scholar] [CrossRef]
  161. Verdoliva, L. Media Forensics and DeepFakes: An Overview. IEEE J. Sel. Top. Signal Process. 2020, 14, 910–932. [Google Scholar] [CrossRef]
  162. Maier, A.; Riess, C. Reliable Out-of-Distribution Recognition of Synthetic Images. J. Imaging 2024, 10, 110. [Google Scholar] [CrossRef] [PubMed]
  163. Lukáš, J.; Fridrich, J.; Goljan, M. Detecting Digital Image Forgeries Using Sensor Pattern Noise. In Proceedings of the Security, Steganography, and Watermarking of Multimedia Contents VIII, SPIE, San Jose, CA, USA, 15 January 2006; Volume 6072, pp. 362–372. [Google Scholar]
  164. French, R.M. Catastrophic Forgetting in Connectionist Networks. Trends Cogn. Sci. 1999, 3, 128–135. [Google Scholar] [CrossRef] [PubMed]
  165. Baracchi, D.; Boato, G.; De Natale, F.; Iuliani, M.; Montibeller, A.; Pasquini, C.; Piva, A.; Shullani, D. Towards Open-World Multimedia Forensics Through Media Signature Encoding. IEEE Access 2024, 12, 59930–59952. [Google Scholar] [CrossRef]
  166. Yang, K.C.; Singh, D.; Menczer, F. Characteristics and Prevalence of Fake Social Media Profiles with AI-Generated Faces. arXiv 2024, arXiv:2401.02627. [Google Scholar] [CrossRef]
  167. Ricker, J.; Assenmacher, D.; Holz, T.; Fischer, A.; Quiring, E. AI-Generated Faces in the Real World: A Large-Scale Case Study of Twitter Profile Images. In Proceedings of the 27th International Symposium on Research in Attacks, Intrusions and Defenses, Padua, Italy, 30 September–2 October 2024; pp. 513–530. [Google Scholar]
  168. Koonce, B.; Koonce, B. ResNet 50. In Convolutional Neural Networks with Swift for TensorFlow: Image Recognition and Dataset Categorization; Apress: Berkeley, CA, USA, 2021; pp. 63–72. [Google Scholar]
  169. Kumar, M.; Sharma, H.K. A GAN-Based Model of Deepfake Detection in Social Media. Procedia Comput. Sci. 2023, 218, 2153–2162. [Google Scholar]
  170. Lago, F.; Pasquini, C.; Böhme, R.; Dumont, H.; Goffaux, V.; Boato, G. More Real Than Real: A Study on Human Visual Perception of Synthetic Faces [Applications Corner]. IEEE Signal Process. Mag. 2021, 39, 109–116. [Google Scholar] [CrossRef]
  171. Cozzolino, D.; Poggi, G.; Corvi, R.; Nießner, M.; Verdoliva, L. Raising the Bar of AI-Generated Image Detection with CLIP. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 16–22 June 2024; pp. 4356–4366. [Google Scholar]
  172. Marcon, F.; Pasquini, C.; Boato, G. Detection of Manipulated Face Videos Over Social Networks: A Large-Scale Study. J. Imaging 2021, 7, 193. [Google Scholar] [CrossRef] [PubMed]
  173. Szegedy, C.; Zaremba, W.; Sutskever, I.; Bruna, J.; Erhan, D.; Goodfellow, I.; Fergus, R. Intriguing Properties of Neural Networks. arXiv 2013, arXiv:1503.02531. [Google Scholar]
  174. Goodfellow, I.J.; Shlens, J.; Szegedy, C. Explaining and Harnessing Adversarial Examples. arXiv 2014, arXiv:1412.6572. [Google Scholar]
  175. Papernot, N.; McDaniel, P.; Goodfellow, I.; Jha, S.; Celik, Z.B.; Swami, A. Practical Black-Box Attacks Against Machine Learning. In Proceedings of the 2017 ACM on Asia Conference on Computer and Communications Security, Abu Dhabi, United Arab Emirates, 2–6 April 2017; pp. 506–519. [Google Scholar]
  176. Barni, M.; Kallas, K.; Nowroozi, E.; Tondi, B. On the Transferability of Adversarial Examples Against CNN-Based Image Forensics. In Proceedings of the ICASSP—IEEE International Conference on Acoustics, Speech and Signal Processing, Brighton, UK, 12–17 May 2019; pp. 8286–8290. [Google Scholar]
  177. Rosa, V.D.; Guillaro, F.; Poggi, G.; Cozzolino, D.; Verdoliva, L. Exploring the Adversarial Robustness of CLIP for AI-generated Image Detection. In Proceedings of the IEEE International Workshop on Information Forensics and Security (WIFS), Rome, Italy, 2–5 December 2024. [Google Scholar]
  178. Bountakas, P.; Zarras, A.; Lekidis, A.; Xenakis, C. Defense Strategies for Adversarial Machine Learning: A Survey. Comput. Sci. Rev. 2023, 49, 100573. [Google Scholar] [CrossRef]
  179. Liang, H.; He, E.; Zhao, Y.; Jia, Z.; Li, H. Adversarial Attack and Defense: A Survey. Electronics 2022, 11, 1283. [Google Scholar] [CrossRef]
  180. Athalye, A.; Carlini, N.; Wagner, D. Obfuscated Gradients Give a False Sense of Security: Circumventing Defenses to Adversarial Examples. In Proceedings of the PMLR, International Conference on Machine Learning, Stockholm, Sweden, 10–15 July 2018; pp. 274–283. [Google Scholar]
  181. Madry, A.; Makelov, A.; Schmidt, L.; Tsipras, D.; Vladu, A. Towards Deep Learning Models Resistant to Adversarial Attacks. arXiv 2017, arXiv:1706.06083. [Google Scholar]
  182. Tsipras, D.; Santurkar, S.; Engstrom, L.; Turner, A. Robustness May Be at Odds with Accuracy. arXiv 2018, arXiv:1805.12152. [Google Scholar]
  183. Purnekar, N.; Abady, L.; Tondi, B.; Barni, M. Improving the Robustness of Synthetic Images Detection by Means of Print and Scan Augmentation. In Proceedings of the Information Hiding & Multimedia Security Conference (IH&MMSEC), Baiona, Spain, 24–26 June 2024. [Google Scholar]
  184. Barni, M.; Nowroozi, E.; Tondi, B. Higher-Order, Adversary-Aware, Double JPEG Detection via Selected Training on Attacked Samples. In Proceedings of the 25th European Signal Processing Conference (EUSIPCO), Kos Island, Greece, 28 August–2 September 2017; pp. 281–285. [Google Scholar]
  185. Coccomini, D.A.; Zilos, G.K.; Amato, G.; Caldelli, R.; Falchi, F.; Papadopoulos, S.; Gennaro, C. MINTIME: Multi-Identity Size-Invariant Video Deepfake Detection. IEEE Trans. Inf. Forensics Secur. 2024, 19, 6084–6096. [Google Scholar] [CrossRef]
  186. Ciamarra, A.; Caldelli, R.; Becattini, F.; Seidenari, L.; Del Bimbo, A. Deepfake Detection by Exploiting Surface Anomalies: The Surfake Approach. In Proceedings of the 2024 IEEE/CVF Winter Conference on Applications of Computer Vision Workshops (WACVW), Waikoloa, HI, USA, 1–6 January 2024; pp. 1024–1033. [Google Scholar] [CrossRef]
  187. Marra, F.; Saltori, C.; Boato, G.; Verdoliva, L. Incremental Learning for the Detection and Classification of GAN-Generated Images. In Proceedings of the 2019 IEEE International Workshop on Information Forensics and Security (WIFS), Delft, The Netherlands, 9–12 December 2019; pp. 1–6. [Google Scholar]
  188. Temmermans, F.; Caldwell, S.; Papadopoulos, S.; Pereira, F.; Rixhon, P. Towards an International Standard to Establish Trust in Media Production, Distribution and Consumption. In Proceedings of the 2023 24th International Conference on Digital Signal Processing (DSP), Island of Rhodes, Greece, 11–13 June 2023; pp. 1–5. [Google Scholar]
  189. Boneh, D.; Lynn, B.; Shacham, H. Short Signatures from the Weil Pairing. In Advances in Cryptology—ASIACRYPT 2001; Boyd, C., Ed.; Springer: Berlin/Heidelberg, Germany, 2001; pp. 514–532. [Google Scholar]
  190. Boneh, D.; Gorbunov, S.; Wahby, R.S.; Wee, H.; Wood, C.A.; Zhang, Z. BLS Signatures. Internet-Draft draft-irtf-cfrg-bls-signature-05, Internet Engineering Task Force. 202, Work in Progress. Available online: https://datatracker.ietf.org/doc/draft-irtf-cfrg-bls-signature/ (accessed on 15 January 2025).
  191. Goyal, V.; Pandey, O.; Sahai, A.; Waters, B. Attribute-Based Encryption for Fine-Grained Access Control of Encrypted Data. In Proceedings of the 13th ACM Conference on Computer and Communications Security, Alexandria, VA, USA, 30 October–3 November 2006; CCS ’06. pp. 89–98. [Google Scholar] [CrossRef]
  192. Rasori, M.; Perazzo, P.; Dini, G. ABE-Cities: An Attribute-Based Encryption System for Smart Cities. In Proceedings of the 2018 IEEE International Conference on Smart Computing (SMARTCOMP), Sicily, Italy, 18–20 June 2018; pp. 65–72. [Google Scholar]
  193. Sicari, S.; Rizzardi, A.; Dini, G.; Perazzo, P.; La Manna, M.; Coen-Porisini, A. Attribute-Based Encryption and Sticky Policies for Data Access Control in a Smart Home Scenario: A Comparison on Networked Smart Object Middleware. Int. J. Inf. Secur. 2021, 20, 695–713. [Google Scholar] [CrossRef]
  194. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, L.; Polosukhin, I. Attention Is All You Need. arXiv 2017, arXiv:1706.03762. [Google Scholar] [CrossRef]
  195. OpenAI. Available online: https://openai.com/ (accessed on 24 December 2022).
  196. OpenAI Blog. Available online: https://openai.com/blog/chatgpt (accessed on 24 December 2022).
  197. Catalfamo, A.; Celesti, A.; Fazio, M.; Randazzo, G.; Villari, M. A Platform for Federated Learning on the Edge: A Video Analysis Use Case. In Proceedings of the 2022 IEEE Symposium on Computers and Communications (ISCC), Rhodes, Greece, 30 June–3 July 2022; pp. 1–7. [Google Scholar] [CrossRef]
  198. Garofalo, M.; Catalfamo, A.; Colosi, M.; Villari, M. Federated Objective: Assessing Client Truthfulness in Federated Learning. In Proceedings of the International Conference on Big Data (IEEE BigData 2024), Washington DC, USA, 15–16 December 2024. [Google Scholar]
  199. Abdelgaber, Y.E.; Ahmed, Y.A.; Salem, M.A.M.; Salem, M.A.G. Federated Learning for Resource Management in Edge Computing. In Proceedings of the 2023 Eleventh International Conference on Intelligent Computing and Information Systems (ICICIS), Cairo, Egypt, 25–27 November 2025; pp. 102–109. [Google Scholar] [CrossRef]
  200. Mario Colosi, Alessio Catalfamo, M.G.; Villari., M. Enabling Flower for Federated Learning in Web Browsers in the Cloud-Edge- Client Continuum. In Proceedings of the 2024 17th IEEE/ACM International Conference on Utility and Cloud Computing (UCC 2024), Sharjah, United Arab Emirates, 16–19 December 2024.
  201. Marco Garofalo, Mario Colosi, A.C.; Villari., M. Web-Centric Federated Learning over the Cloud-Edge Continuum Leveraging ONNX and WASM. In Proceedings of the 2024 IEEE Symposium on Computers and Communications (ISCC), Paris, France, 26–29 June 2024; pp. 1–7. [CrossRef]
  202. Catalfamo, A.; Ruggeri, A.; Celesti, A.; Fazio, M.; Villari, M. A Microservices and Blockchain Based One Time Password (MBB-OTP) Protocol for Security-Enhanced Authentication. In Proceedings of the 2021 IEEE Symposium on Computers and Communications (ISCC), Athens, Greece, 5–8 September 2021; pp. 1–6. [Google Scholar] [CrossRef]
  203. Lukaj, V.; Martella, F.; Fazio, M.; Galletta, A.; Celesti, A.; Villari, M. Gateway-Based Certification Approach to Include IoT Nodes in a Trusted Edge/Cloud Environment. In Proceedings of the 2023 IEEE/ACM 23rd International Symposium on Cluster, Cloud and Internet Computing Workshops (CCGridW), Bangalore, India, 1–4 May; pp. 237–241.
Figure 1. Generic diagram of the deepfake creation process.
Figure 1. Generic diagram of the deepfake creation process.
Jimaging 11 00073 g001
Figure 2. (a) The GAN framework consists of a Generator (G) that creates synthetic data x ¯ from random noise z, aiming to learn the data distribution p d a t a , and a Discriminator (D) that distinguishes between real and generated data. Both models are trained simultaneously in an adversarial manner; (b) The Diffusion Model [8] uses a fixed Markov chain to add Gaussian noise to data, approximating the posterior distribution q ( x t | x t 1 ) for t = 1 , , T . The goal is to learn the reverse process p θ ( x t 1 | x t ) to generate data by reversing the noise-adding chain, where x 1 , , x T are latent variables with the same dimensionality as x 0 .
Figure 2. (a) The GAN framework consists of a Generator (G) that creates synthetic data x ¯ from random noise z, aiming to learn the data distribution p d a t a , and a Discriminator (D) that distinguishes between real and generated data. Both models are trained simultaneously in an adversarial manner; (b) The Diffusion Model [8] uses a fixed Markov chain to add Gaussian noise to data, approximating the posterior distribution q ( x t | x t 1 ) for t = 1 , , T . The goal is to learn the reverse process p θ ( x t 1 | x t ) to generate data by reversing the noise-adding chain, where x 1 , , x T are latent variables with the same dimensionality as x 0 .
Jimaging 11 00073 g002
Figure 3. Taxonomy of Deepfake Detection Methods. Detection techniques are categorized into four primary groups: general network-based approaches, methods focusing on visual artifacts, approaches targeting biological inconsistencies, and techniques leveraging texture and spatio-temporal consistency. Each category addresses specific characteristics of manipulated content.
Figure 3. Taxonomy of Deepfake Detection Methods. Detection techniques are categorized into four primary groups: general network-based approaches, methods focusing on visual artifacts, approaches targeting biological inconsistencies, and techniques leveraging texture and spatio-temporal consistency. Each category addresses specific characteristics of manipulated content.
Jimaging 11 00073 g003
Figure 4. Example of deepfake detection features. The full face (a) exhibits visual artifacts, such as unnatural pixel formations around the facial features (e.g., glasses and skin edges). These are indicative of irregularities introduced by generative algorithms or compression distortions. The zoomed-in region (b) around the mouth shows potential texture inconsistencies, such as unnatural blending of lip textures and a lack of smooth transitions typical of natural skin and lip patterns.
Figure 4. Example of deepfake detection features. The full face (a) exhibits visual artifacts, such as unnatural pixel formations around the facial features (e.g., glasses and skin edges). These are indicative of irregularities introduced by generative algorithms or compression distortions. The zoomed-in region (b) around the mouth shows potential texture inconsistencies, such as unnatural blending of lip textures and a lack of smooth transitions typical of natural skin and lip patterns.
Jimaging 11 00073 g004
Figure 5. A conceptual pipeline for deepfake detection and model recognition, illustrating the process of identifying whether an image is real or synthetic, determining the generative architecture, and tracing the specific model instance used for creation.
Figure 5. A conceptual pipeline for deepfake detection and model recognition, illustrating the process of identifying whether an image is real or synthetic, determining the generative architecture, and tracing the specific model instance used for creation.
Jimaging 11 00073 g005
Figure 6. The TrueFace Dataset [158], comprising 80k fake images generated with StyleGAN models and 70k real images, of which 60k have been also shared on three distinct social networks.
Figure 6. The TrueFace Dataset [158], comprising 80k fake images generated with StyleGAN models and 70k real images, of which 60k have been also shared on three distinct social networks.
Jimaging 11 00073 g006
Table 1. Comparison of topics in deepfake research, highlighting their focus, techniques, proactivity, practical applications, and constraints addressed.
Table 1. Comparison of topics in deepfake research, highlighting their focus, techniques, proactivity, practical applications, and constraints addressed.
AspectDeepfake Detection (Section 3, Section 5 and Section 6)Attribution and Recognition (Section 7)Passive Methods (Section 8)Realistic Scenarios (Section 9)Active Authentication (Section 10)
FocusDetecting fake mediaIdentifying creators/toolsForensic analysis of contentDetection methods in real-world contextsEmbedding authenticity proactively
TechniquesML, DL, handcrafted featuresSource tracing, watermarkingArtifact analysis, temporal signalsCompression-aware, robustness frameworksWatermarking, blockchain, cryptography
ProactivityReactiveReactiveReactiveContext-adaptiveProactive
Practical Use CaseIdentifying deepfakesTracing fake originsLow-resource environmentsAddressing low-quality, compressed contentEnsuring authenticity before distribution
Constraints AddressedFeature engineering, ML/DL gapsAccountabilityLack of active authenticationReal-world deployment challengesPre-distribution authenticity assurance
Table 2. Accuracy of synthetic image detectors when faced with image processing operations that were not part of the augmentation set, comprising JPEG2000 and WebP compression, Median (Med), Gaussian (GF), and Wiener (WF) filtering, brightness and contrast transformations (CT), hue modification in the HSV space (HSV), and P&S simulation trained on patches (P&S P).
Table 2. Accuracy of synthetic image detectors when faced with image processing operations that were not part of the augmentation set, comprising JPEG2000 and WebP compression, Median (Med), Gaussian (GF), and Wiener (WF) filtering, brightness and contrast transformations (CT), hue modification in the HSV space (HSV), and P&S simulation trained on patches (P&S P).
TaskModelAugmentJPEG 2000WebPMedGFWFCTHSVP&S P
N vs. SG2XNetConventional0.870.720.820.910.560.980.860.50
+ patch P&S0.930.920.930.940.900.980.930.81
+ full P&S0.940.840.860.970.920.970.910.50
ResNetConventional0.940.920.910.960.960.970.930.95
+ patch P&S0.960.960.950.970.960.970.930.97
+ full P&S0.950.950.920.960.950.960.940.96
N vs. DMXNetConventional0.580.670.690.980.540.990.990.50
+ patch P&S0.770.950.780.840.780.990.990.89
+ full P&S0.630.930.750.860.590.990.980.50
N vs. TTXNetConventional0.910.960.861.00.911.00.990.5
+ patch P&S0.990.990.960.990.990.990.990.97
+ full P&S0.940.990.900.990.980.990.990.50
Table 3. Comparison of deepfake detection methods discussed in this paper, along with their strengths, weaknesses and the Sections where they are examined.
Table 3. Comparison of deepfake detection methods discussed in this paper, along with their strengths, weaknesses and the Sections where they are examined.
MethodDescriptionStrengthsWeaknessesRelevant Sections
Forensic-Based DetectionUses handcrafted features to detect anomalies in digital content (e.g., artifacts, lighting inconsistencies).Interpretable; Can detect handcrafted manipulations; Computationally efficient.Struggles with advanced AI-generated content; Sensitive to compression artifacts.Section 3.2, Section 3.4, Section 8 and Section 9
Machine Learning-Based DetectionEmploys deep learning models like CNNs and Transformers to learn patterns from large datasets.High accuracy on benchmark datasets; Can generalize well if trained on diverse data.Requires large labeled datasets; Prone to adversarial attacks and dataset bias.Section 3.1, Section 8 and Section 9
Biometric-Based DetectionLeverages physiological and behavioral cues (e.g., blinking, heart rate, gaze tracking) to detect manipulations.Difficult for deepfake models to replicate; Based on real human characteristics.Requires high-resolution input; Can fail under occlusions or low-light conditions.Section 3.3
Continual Learning for Deepfake DetectionEnables models to continuously adapt to new types of deepfake manipulations over time.Improves model adaptability; Reduces catastrophic forgetting; Enables learning from evolving deepfake techniques.Requires large-scale continual training; Susceptible to domain shifts; High computational cost.Section 5 and Section 9
Deepfake Attribution and Model FingerprintingUses techniques to analyze residual patterns in synthetic media to determine the source generative model.Traces deepfake origins; Identifies generative model fingerprints; Useful for forensic tracking.Limited effectiveness if the attacker modifies or fine-tunes the generator; Relies heavily on dataset biases.Section 7
Methods for Explainability in Deepfake DetectionFocuses on making deepfake detection models more interpretable and explainable to improve forensic reliability.Enhances trust in detection models; Provides interpretability for forensic investigations; Supports model debugging.No standard framework for explainability; May reduce model accuracy due to interpretability constraints; Black-box AI models remain dominant.Section 6 and Section 9
Active AuthenticationUses proactive security mechanisms such as watermarking and cryptographic signatures to ensure authenticity.Provides proactive protection; Can grant deepfake detection and generation source attributionDepends on adoption by platforms; Can be circumvented if attackers remove metadata.Section 10
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Amerini, I.; Barni, M.; Battiato, S.; Bestagini, P.; Boato, G.; Bruni, V.; Caldelli, R.; De Natale, F.; De Nicola, R.; Guarnera, L.; et al. Deepfake Media Forensics: Status and Future Challenges. J. Imaging 2025, 11, 73. https://doi.org/10.3390/jimaging11030073

AMA Style

Amerini I, Barni M, Battiato S, Bestagini P, Boato G, Bruni V, Caldelli R, De Natale F, De Nicola R, Guarnera L, et al. Deepfake Media Forensics: Status and Future Challenges. Journal of Imaging. 2025; 11(3):73. https://doi.org/10.3390/jimaging11030073

Chicago/Turabian Style

Amerini, Irene, Mauro Barni, Sebastiano Battiato, Paolo Bestagini, Giulia Boato, Vittoria Bruni, Roberto Caldelli, Francesco De Natale, Rocco De Nicola, Luca Guarnera, and et al. 2025. "Deepfake Media Forensics: Status and Future Challenges" Journal of Imaging 11, no. 3: 73. https://doi.org/10.3390/jimaging11030073

APA Style

Amerini, I., Barni, M., Battiato, S., Bestagini, P., Boato, G., Bruni, V., Caldelli, R., De Natale, F., De Nicola, R., Guarnera, L., Mandelli, S., Majid, T., Marcialis, G. L., Micheletto, M., Montibeller, A., Orrù, G., Ortis, A., Perazzo, P., Puglisi, G., ... Vitulano, D. (2025). Deepfake Media Forensics: Status and Future Challenges. Journal of Imaging, 11(3), 73. https://doi.org/10.3390/jimaging11030073

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop