Next Article in Journal
Optical Coherence Tomography (OCT) Findings in Post-COVID-19 Healthcare Workers
Previous Article in Journal
Efficient Multi-Material Volume Rendering for Realistic Visualization with Complex Transfer Functions
Previous Article in Special Issue
Deepfake Media Forensics: Status and Future Challenges
 
 
Due to scheduled maintenance work on our database systems, there may be short service disruptions on this website between 10:00 and 11:00 CEST on June 14th.
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Evaluating Features and Variations in Deepfake Videos Using the CoAtNet Model

1
Computer Science Department, Faculty of Computing and Information Technology, King Abdulaziz University, Jeddah 21589, Saudi Arabia
2
School of Computer Science, University of Sheffield, Regent’s Court, Sheffield S1 4DP, UK
3
Institutional Advancements, King Abdullah University of Science and Technology, Thuwal 23955, Saudi Arabia
*
Author to whom correspondence should be addressed.
J. Imaging 2025, 11(6), 194; https://doi.org/10.3390/jimaging11060194
Submission received: 23 April 2025 / Revised: 23 May 2025 / Accepted: 4 June 2025 / Published: 12 June 2025

Abstract

:
Deepfake video detection has emerged as a critical challenge in the realm of artificial intelligence, given its implications for misinformation and digital security. This study evaluates the generalisation capabilities of the CoAtNet model—a hybrid convolution–transformer architecture—for deepfake detection across diverse datasets. Although CoAtNet has shown exceptional performance in several computer vision tasks, its potential for generalisation in cross-dataset scenarios remains underexplored. Thus, in this study, we explore CoAtNet’s generalisation ability by conducting an extensive series of experiments with a focus on discovering features and variations in deepfake videos. These experiments involve training the model using various input and processing configurations, followed by evaluating its performance on widely recognised public datasets. To the best of our knowledge, our proposed approach outperforms state-of-the-art models in terms of intra-dataset performance, with an AUC between 81.4% and 99.9%. Our model also achieves outstanding results in cross-dataset evaluations, with an AUC equal to 78%. This study demonstrates that CoAtNet achieves the best AUC for both intra-dataset and cross-dataset deepfake video detection, particularly on Celeb-DF, while also showing strong performance on DFDC.

1. Introduction

Videos are an extremely common multimedia form that can be conveniently transferred to different social media platforms, such as WhatsApp, YouTube, Instagram, and Facebook. Multimedia content, including videos, can be readily modified via modern editing tools [1]. This is perhaps the start of a slippery slope regarding the “authenticity” of such content. Modifications may have positive aesthetic or presentational goals, but, of course, some modifications may be intended to mislead. Where videos are concerned, such matters took a significant turn in 2017 when a Reddit account called “Deepfake” posted synthetic pornographic videos generated using a Deep Neural Network (DNN). The account’s name, combining “Deep” (from deep learning) and “fake”, caught on and now refers to hyper-realistic images, speech, and videos generated using Generative Adversarial Networks (GANs) that render the identification of their authenticity difficult for humans [2,3].
While digitally synthesising faces or manipulating a real face requires a significant volume of source data, such data are now publicly available. Contemporary deep learning techniques, such as autoencoders (AEs) and Generative Adversarial Networks (GANs), eliminate several manual editing processes [4].
Multiple mobile applications, websites, and software programs have been made publicly available, allowing for the production of high-level synthesised media. These resources require no prior training. Applications such as FakeApp [5], DFaker [6], Faceswap-GAN [7], Faceswap [8], and DeepFaceLab [9] have been used to create the deepfakes contained in deepfake datasets or videos circulated on the Internet that involve celebrities, such as former president Barack Obama [10] and actor Tom Cruise [11].
Although there is concern regarding deepfake technology, it also has creative and productive applications [4,12,13,14]. For example, it can be used in education, criminal forensics, virtually trying on clothes while shopping, 3D modelling industrial applications, entertainment [15], acting [16], film production, and video dubbing [17]. In education, deepfakes can enhance engagement by transforming teachers into familiar characters or animating historical figures for classroom interaction. In privacy and healthcare, they can help to de-identify patients in videos while preserving useful visual information and serve as virtual counsellors, and are especially effective for individuals with social anxiety. Additionally, AI characters can play roles in preserving culture and history by bringing historical artworks and figures to life. These technologies can enrich storytelling, therapy, and cultural preservation efforts [18].
At the same time, deepfakes have raised significant concerns, particularly due to the potential for their abuse and misuse [2,19,20]. Deepfake videos can misrepresent a person’s views and actions, which could result in serious political, social, financial, and legal issues. Deepfakes pose wide-ranging threats if used harmfully: the manipulation of the stock market, political discourse, or elections; targeting celebrities with revenge porn; creating fake news and spreading misinformation; financial fraud; and creating fake social media accounts to incite violence or direct the public to specific perspectives. These fake multimedia have serious consequences, such as misinforming the public, harming a personal or business reputation, affecting political perspectives, and being maliciously used as evidence in court.
Various laws seek to protect individuals against the misuse of deepfake technology. For example, in the USA, the DEEPFAKES Accountability Act (H.R. 5586) [21] establishes civil remedies for victims of harmful deepfake content, giving victims the right to initiate civil actions against individuals who create or distribute deepfake material that causes harm. This empowers victims to seek damages and injunctive relief, offering a legal avenue to address and mitigate the impacts of malicious deepfake content. There are also similar regulations in Canada [22] and China [23].
Deepfake misuse leads to doubts about available videos and concerns about people’s privacy [24]. Moreover, it poses an issue for security and ethics, as visual media can no longer be considered trustworthy content [25]. Consequently, there is a great demand for methods to verify that videos are genuinely what they appear to be.
As the public’s interest in deepfake technology grows, so will the number of relevant studies. Over the last three years, tremendous progress has been made in developing detection technologies. The academic community, research groups, and commercial companies worldwide are undertaking relevant studies to mitigate the negative effects of such a problem [26].
Most research to date has involved training and evaluating detection models using a restricted dataset. The training dataset will contain application instances of the same faking techniques, perhaps in addition to the same real-world environments. Models developed in this way often exhibit good generalisation performance on “unseen” examples, but they can radically underperform when such data instance assumptions are relaxed. For example, a model trained using a dataset that includes fakes made using three different techniques may be useless at detecting fakes that are created using a fourth technique. In a sense, models are developed to carry out generalisations in an intra-dataset manner, but they do not generalise reliably across datasets (i.e., when applied to different datasets).
In this study, we investigate the generalisation capability of the CoAtNet model for deepfake video detection across multiple datasets. The CoAtNet model combines aspects of convolutional networks and specific vision-focused attention networks known as Vision Transformers. While CoAtNet has demonstrated exceptional performance in various computer vision tasks, its effectiveness in distinguishing real from manipulated videos—particularly in cross-dataset scenarios—remains underexplored. Our research systematically evaluates CoAtNet’s performance using benchmark deepfake datasets, including FaceForensics++ [27], DFDC [28], Celeb-DF [29], and FaceShifter [30], to assess its robustness and adaptability to unseen data. The main contributions of this study can be summarised as follows:
  • We evaluate the generalisation ability of the CoAtNet model in deepfake videos for synthesised faces and discover different features and variations.
  • Our study proposes an improved CoAtNet model (CoAtNet16A) that ensures better generalisation.
  • We investigate the detection effect of CoAtNet16A using different frame selection strategies, including a single middle frame, fifteen random frames, fifteen optical flow frames, cosine similarity keyframes, and facial landmark keyframes.
Our method—using CoAtNet with a voting-based approach that integrates predictions from single frames, random frames, and optical flow frames—achieved outstanding performances on the FF++ dataset, with an AUC of 0.9996, surpassing leading methods. In cross-dataset evaluations, our model demonstrated superior results on the Celeb-DF dataset with an AUC of 0.76 and on the DFDC dataset with an AUC of 0.68 (the fake images in these datasets were created using different manipulation techniques).
The remainder of this study is organised as follows: Section 2 reviews related research on deepfake generation and detection, including CNN-based approaches, Vision Transformer (ViT)-based approaches, and the CoAtNet model. Section 3 illustrates the two-stage proposed methodology and the details of the experiments. Section 4 outlines the comparison of performance evaluations for both intra-dataset and cross-dataset contexts. Finally, Section 5 provides the conclusions, limitations, and future research directions.

2. Related Research

In recent years, extensive research has been conducted to address the growing challenge of deepfake detection, driven by the rapid advancements in deepfake generation techniques. Early detection methods primarily relied on machine learning techniques and then on Convolutional Neural Networks (CNNs) due to their strong ability to capture spatial features from images and videos. However, with the emergence of more sophisticated and realistic deepfakes, researchers have explored advanced architectures such as Vision Transformers (ViTs), which excel at modelling global dependencies within visual data. Recently, CNN/ViT hybrid models such as CoAtNet have been introduced. This section summarises deepfake generation techniques, highlighting the strengths and weaknesses of both CNNs and ViTs, and explains how the CoAtNet model combines them.

2.1. Deepfake Generation

There are several techniques to generate hyper-realistic images, videos, and audio. However, the most used techniques are variations or combinations of deep learning architectures, such as Encoder–Decoder networks and Generative Adversarial Networks (GANs) [31]. Encoder–Decoder (ED) networks consist of an encoder that extracts latent features from an image and a decoder that reconstructs the image from these features [26]. On the other hand, a GAN comprises two competing neural networks: a generator G and a discriminator D. G produces fake samples to deceive D, while D learns to distinguish between real samples and fake samples. The repetition of this scenario results in G developing better samples (i.e., they increasingly cannot be distinguished from those of the real samples) [32].
In deepfake generation studies, Lyu [2] categorised the manipulation types into three categories—head puppetry, face swapping, and lip syncing—as shown in Figure 1. Head puppetry (also called facial re-enactment [33]) involves changing the target’s entire head and upper shoulder according to the head of the source person to give the same appearance as the target. Face swapping is the process where the target’s faces are swapped with synthesised faces from the source, maintaining facial expressions. Lip syncing creates a fake video by altering the target’s lips to be consistent with speech chosen by the attacker (i.e., it “puts words into the target’s mouth”).

2.2. Deepfake Detection

The literature reveals a progression from early heuristic-based techniques to sophisticated architectures designed for the deepfake detection task. This section reviews the state-of-the-art studies on deepfake detection, highlighting key methodologies, datasets, and challenges in this evolving threat.
Afchar et al. [34] were the first to detect deepfake videos without using traditional image forensics techniques. They proposed the MesoNet model, a CNN architecture with a few layers focusing on the images’ mesoscopic properties (smaller semantic details) to analyse video frames. Two different types of architecture were used: Meso4 and MesoInception4. Another study carried out by Nguyen, Yamagishi, and Echizen [35] investigated the utilisation of Capsule Networks for detecting fake images and videos. Capsule Networks, recognised for their proficiency in discerning spatial hierarchies within datasets, present a promising alternative to conventional Convolutional Neural Networks (CNNs) by mitigating their shortcomings in terms of the identification of object poses and deformations. Their research leveraged the unique capabilities of Capsule Networks to improve the accuracy and robustness of detection. Dang et al. [36] proposed using an attention mechanism to produce an improved feature map, which is then used for both fake detection and predicting associated manipulation regions. Wodajo and Atnafu [37] proposed using a CNN and Vision Transformer (ViT) hybrid model to learn both local and global features. The CNN acts as a learnable feature extractor. The features are input into the ViT and classified using the attention mechanism. Zhao et al. [38] used fine-grained classification, which gathers local discriminative features to differentiate between categories in order to solve the deepfake detection problem. This model uses a multi-attentional network that includes three key components: textural feature enhancement blocks, multiple spatial attention heads, and textural and semantic features aggregation. Luo et al. [39] suggested a model for solving the generalisation problem. They found that CNN-based detectors exhibit biases to fakery method-specific textures. Since high-frequency noises remove colour textures, they proposed using these types of noise to remove the colour textures, exposing statistical discrepancies between real and fake images. Wang et al. [40] proposed a hybrid model that combines both CNN and transformer architectures. This technique is designed to overcome the shortcomings of current deepfake detection methodologies, especially regarding their generalisability across diverse datasets. The proposed model demonstrates improved performance in detecting deepfakes compared to traditional CNN-based methods—particularly in cross-dataset evaluations—achieving an AUC of 0.98 on FF++, 0.74 on DFDC, and 0.72 on Celeb-DF.
Although multiple studies have tried to address the generalisation issue for deepfake detection, there is still significant room for improvement, and the effect of using an advanced deep learning model to solve this challenge should be explored.

2.3. CNN-Based Approaches

CNNs are a class of deep learning models that have revolutionised fields such as computer vision and natural language processing. CNNs are designed to automatically and adaptively learn spatial hierarchies of features from input images, rendering them highly effective for tasks such as image classification, object detection, and image denoising. This capability is achieved through the use of convolutional layers, pooling layers, and fully connected layers, which together form the architecture of a CNN. The convolutional layers apply a series of filters to the input data, capturing local patterns, while pooling layers reduce dimensionality, and fully connected layers integrate the learned features for classification or regression tasks [41].
Deep learning is the dominant deepfake detection approach, with CNNs being the most represented specific architecture [42]. CNNs employ a convolution filter that extracts important edges by filtering the surrounding pixel values, independent of their position [43]. There are two types of features in images that provide different information: local and global features. Local features describe small groups of pixels (also known as image “patches”), while global features describe the entire image [44]. Even though CNNs produce outstanding performance in learning local image information, their limited receptive fields prevent them from capturing the spatial interdependence of pixels; in other words, CNN models tend to concentrate only on the activated segment of the face and ignore other parts. As a result, CNNs cannot determine and leverage the relationships between the different parts of images; for instance, the model is unable to detect an unnatural relationship between the mouth and eyes. Additionally, CNNs present an overfitting problem and cannot carry out generalisation relative to unseen fake videos during training or diverse categories of deepfake generation techniques [45].

2.4. ViT-Based Approaches

A transformer (a form of neural network) learns context and meaning across sequential data. It harnesses the concepts of attention or self-attention to detect the relationships between elements, even if they are far away. Before the invention of transformers, users were required to train neural networks using large, labelled datasets. It is acknowledged that the production of such datasets is resource-intensive. Transformers eliminate this need by mathematically identifying patterns between elements. Additionally, the implementation of transformer theory lends itself to the use of parallel processing, allowing these models to run quickly [46]. Moreover, transformers discover the long-term dependency between video frames and are scalable to highly complex models on large-scale datasets [47].
Transformers have achieved considerable success in natural language processing (NLP) tasks. This has inspired their application to computer vision (CV) problems, including object detection [48], image recognition [49], video classification [50], image segmentation [51], image captioning [52], and visual question answering (where the developed model must respond to questions posed about an image) [53]. They have achieved state-of-the-art results.
The Vision Transformer (ViT) model was introduced in 2021 by Google [49]. Their model applies attention to small “patches” of the image, rather than individual pixels. As clarified in Figure 2, the ViT model divides an image into fixed-size (16 × 16 pixel) patches, flattens the patches, and includes positional embedding as an input to the transformer encoder. The encoder comprises Multi-Head Self-Attention (MSA) and Multi-Layer Perceptron (MLP) components. The model is then trained and fine-tuned for image classification. The features are linked by the Multi-Head Self-Attention Layer (MSL), which enables the information to be globally distributed across the overall image.
ViTs have two advantages over CNNs. Firstly, they have input-adaptive weighting. Unlike a convolution kernel, which is static and input-independent, their attention weights are dynamic and may change according to input. The second advantage is a global receptive field, which means that a ViT can observe the entire image in one glance (while a CNN usually does not, as mentioned above) [54].
However, transformers have some limitations. First, image attention networks often struggle with translational invariance, meaning that their performance can vary when objects in an image are shifted or repositioned. Thus, in order to outperform CNNs, ViTs had to be trained on large datasets comprising hundreds of millions of images [49]. If the ViT-based models are trained with insufficient data, they perform worse than CNNs and do not generalise well. Moreover, ViT-based models focus on global features and underperform CNNs in local features [55].

2.5. CoAtNet Model

Vision Transformers have received increasing interest in computer vision; however, they have some drawbacks. The same is true for CNNs. CoAtNet [49] seeks to combine the strengths of both. As clarified in Table 1, CoAtNet (pronounced “coat” net) is the abbreviation of convolution and self-attention, and it appeared at the end of 2021. It is a hybrid model built from ViTs and CNNs. It improves the generalisation ability, capacity, and efficiency of the model. Model generalisation refers to the ability of the model to maintain a level of performance relative to unseen data that is similar to that relative to training data. This requires the avoidance of “overfitting”. In comparison, model capacity refers to a model’s ability to accommodate large training datasets. When training data are numerous and overfitting is not a concern, the model with the higher capacity will achieve superior final performance results after an adequate training step. CoAtNets obtain state-of-the-art performance when applied to the ImageNet dataset under varying resource constraints [54]. Figure 3 presents the architecture of the CoAtNet model [54].
The architecture consists of five stages (S0, S1, S2, S3, and S4), starting with S0 and then C-C-T-T, where C represents convolution and T represents transformer. S0 is a simple two-layer convolutional stem, and it is used to lower dimensionality. S1 and S2 are convolution blocks. They contain Mobile Inverted Bottleneck Convolution (MBConv) blocks, which employ depth-wise convolution with Squeeze–Excitation (SE) to reduce the spatial size before being transferred to global attention mechanisms. S3 and S4 are transformer blocks, and they contain relative self-attention components followed by a Feed-Forward Network (FFN). Relative self-attention uses the position between patches instead of their absolute position. The latter approach is used by the standard ViT. Finally, the CoAtNet ends with global pooling and a fully connected layer.

3. Proposed Framework for Evaluating the Generalisation of the CoAtNet Model

The proposed framework is divided into two stages, each of which includes a set of experiments. In the first stage (Experiment Settings), the experiments explore various parameters (frame size, using a pre-trained model, data augmentation, and threshold strategies) and features (face alignment and Local Binary Pattern (LBP) features) to identify the most effective settings. These preliminary experiments serve as a foundation to determine which settings provide the best performance in the AUC of the CoAtNet model. The second stage (Performance Improvements) involves adopting these best-performing settings for the investigation of a variety of frame selection strategies: a single middle-of-video frame, fifteen random frames, fifteen optical flow frames (essentially consecutive frames allowing inter-frame relationships to be leveraged), or keyframes using cosine similarity and facial landmarks.
Figure 4 illustrates the details of the proposed framework for evaluating the generalisation ability of the CoAtNet model. The face images are extracted from video frames using Dlib [56] to remove non-facial (background) information that is useless for deepfake detection. In fact, tracking facial information rather than using the complete frame as input should improve performance. The cropped face images are resized to 224 × 224. A particular CoAtNet implementation has been selected [57], which was designed for multiclassification over the CIFAR10 dataset [58].

3.1. Stage 1: Experimental Settings

This section outlines the datasets used, the proposed model, the implementation details, and the initial experiments for deciding the best combination of settings. In this stage, all models are trained on a single middle frame from each video in the DeepFakes category of FF++ instead of using the complete dataset to see the effect of different features within a reasonable time frame.

3.1.1. Datasets

The FaceForensics++ [59], DFDC [60], Celeb-DF [61], and FaceShifter [59] datasets were used in this study to ensure robust evaluations across different manipulation techniques and real-world scenarios. They provide diverse datasets with varying difficulty levels, helping to assess model generalisation for both intra-dataset and cross-dataset performance. Using these datasets also enables direct comparisons with previous studies, ensuring fair benchmarking and highlighting improvements or limitations in generalisation across datasets.
  • FaceForensics++ [27]
FaceForensics++ includes four faking algorithms: DeepFakes (DF), Face2Face (F2F), FaceSwap (FS), and NeuralTextures (NT).
  • DF: This fakery approach uses two autoencoders with a shared encoder trained to reconstruct source and target face images. A face detector crops and aligns images, and the trained Encoder–Decoder of the source is applied to the target to generate a fake image. The final output is blended using Poisson image editing for seamless integration.
  • F2F: This fakery approach reconstructs a 3D face model; tracks expressions, poses, and lighting; and transfers 76 Blendshape coefficients from the source to the target. The approach automates keyframe selection and re-enactment manipulation for realistic facial synthesis.
  • FS: This fakery approach is a graphic-based method that transfers a face region from a source video to a target using detected facial landmarks. It fits a 3D template model with blended shapes, back-projects it onto the target, and blends the rendered model with the image, applying colour correction for a seamless result.
  • NT: This is a rendering approach that learns a neural texture of the target person from video data, incorporating a rendering network trained with photometric reconstruction and adversarial losses. It uses tracked geometry during training and testing, applying patch-based GAN loss for realistic facial re-enactment.
Figure 5 shows some examples from the FF++ dataset. The first two columns (“Original 1” and “Original 2”) contain unaltered images of individuals. The following four columns display manipulated versions of the original images using the four faking techniques. We have selected two examples from the FF++ dataset presented in the first and second rows of Figure 5. Each row demonstrates how the same person appears under different manipulation methods. Each method has different visual artefacts that highlight the challenges of detecting deepfakes, as some methods seem more realistic than others.
2.
DFDC [28]
The DFDC dataset contains instances created using eight faking methods: Deepfake Autoencoder (DFAE), which used two input/output resolutions of 128 × 128 (DF-128) and 256 × 256 (DF-256); Morphable Model/Neural Network (MM/NN); Neural Talking Head (NTH); FaceSwapGAN (FSGAN); StyleGAN; Refinement; and AudioSwap (Audio). The faking techniques can be integrated with one another:
  • Deepfake Autoencoder (DFAE): This is a convolutional autoencoder with a shared encoder and two separately trained decoders for each identity in a face swap. It extends the shared encoder beyond the bottleneck and uses PixelShuffle for upscaling. This design helps the encoder learn common features while the decoders capture identity-specific details, enabling realistic face swaps during inference. Two resolutions are used: 128 × 128 (DF-128) and 256 × 256 (DF-256).
  • Morphable Model/Neural Network (MM/NN): This method uses a frame-based morphable-mask model to perform face swaps. It aligns source and target facial landmarks, morphs source pixels to match the target, and blends the eyes and mouth from the original video. Spherical harmonics adjust illumination, and a nearest-neighbour approach selects the best source–target face pair based on expression similarity.
  • Neural Talking Head (NTH): This generates realistic talking heads using few-shot and one-shot learning. The process encompasses two phases: meta-learning, which facilitates the conversion of landmarks into authentic facial representations, and fine-tuning, wherein a pre-trained model rapidly adapts to new faces. It is fine-tuned on DFDC video pairs by extracting landmarks from a driving video and generating images with the target person’s appearance.
  • FaceSwapGAN (FSGAN): FSGAN uses GANs for face swapping and re-enactment, adapting to pose and expression changes. It employs adversarial loss for re-enactment and inpainting, with additional generators for face segmentation and Poisson blending.
  • StyleGAN: StyleGAN performs face swaps by projecting a fixed identity descriptor onto the latent face space for each video frame, ensuring consistent identity transfer throughout the video.
  • Refinement: In the final step of fake generation, a randomly selected set of videos underwent post-processing. Applying a basic sharpening filter to the blended faces significantly enhanced the visual quality of the final video, with almost no additional computational cost.
  • AudioSwap (Audio): Some video clips underwent audio swapping using the TTS Skins voice conversion method [62]. TTS Skins can perform multi-voice Text-to-Speech (TTS) by converting a TTS-generated voice into various target voices.
Figure 6 presents some examples from the DFDC dataset. The left column (“Original”) contains unaltered images of individuals, while the right one (“Fake”) contains deepfake-altered versions of the same individuals. Two examples from the DFDC dataset are presented in the first and second rows. As illustrated in Figure 6, deepfake manipulations are subtle. They may be difficult to detect with the naked eye, and automated detection may need to be more sophisticated.
3.
Celeb-DF [23]
Here, the standard DeepFake generation method is refined using multiple methods to address specific visual artefacts present in existing datasets. Figure 7 presents two examples of the Celeb-DF dataset. The “Original 1” and “Original 2” columns contain unaltered images of individuals, while the next one (“Fake”) contains a fake version generated by applying face swapping between the two original individuals. Two examples are shown. It is clear that some deepfakes appear highly realistic, making detection difficult without advanced models. In the Celeb-DF dataset, there is version 1 (V1), which includes 795 videos, and version 2 (V2), which includes 5639 videos. In this study, we used Celeb-DF V2.
4.
FaceShifter (FSh) [30]
FaceShifter includes a single faking method that consists of a two-stage face-swapping algorithm. Figure 8 shows two examples of the FaceShifter dataset in the first and second rows. It produces realistic deepfake data, which challenge the deepfake detection process.
Table 2 summarises the specifications of the well-known datasets. As clarified above, each dataset was created using different deepfake techniques, which implies a significant challenge in deepfake detection due to the diversity of manipulation methods and their unique artefacts. Each technique introduces distinct visual and temporal inconsistencies, rendering it difficult for a single model to carry out generalisation effectively across all types [31]. Table 3 shows the overlap between the datasets in the used faking algorithms. Although some common techniques exist, each dataset used an improved version to generate the fake data.
The datasets were split into training, validation, and testing sets with ratios of 70%, 15%, and 15% for the FF++, Celeb-DF, and FaceShifter datasets. The DFDC dataset comes pre-packaged with training, validation, and testing sets, and the research of others has respected this split. We calculated the adopted split as 93%, 3%, and 4%. The datasets are not balanced; thus, balancing was applied using oversampling for the training, validation, and testing sets.

3.1.2. Proposed Model: CoAtNet16A

In this study, we introduce CoAtNet16A, a hybrid architecture that combines the convolutional–transformer design of CoAtNet with transfer learning from VGG16 weights, further enhanced by a tailored augmentation strategy. This architecture is motivated by two observations: (1) CoAtNet effectively merges the strengths of CNNs and transformers, and (2) VGG16 pretraining, while traditionally used in CNNs, has not been systematically integrated into hybrid transformer-based models such as CoAtNet for deepfake detection tasks.
To evaluate the contribution of this innovative methodology, we performed a comparative analysis involving three model configurations: (i) a CoAtNet model trained from scratch, (ii) CoAtNet pre-trained on ImageNet [63], and (iii) CoAtNet pre-trained using VGG16 weights [64], referred to as CoAtNet16. To further improve CoAtNet16, we applied several data augmentation strategies [65,66,67]. The configuration with the augmentation method from [67] is denoted as CoAtNet16A. The details of the experiment result are found in Section 3.1.3.

3.1.3. Parameter Settings

In this section, we describe the parameter settings applied in our experiments, and we explain the different experiments used to explore and evaluate features and variations in deepfake video detection. The batch size that was used is 16, and the initial learning rate is 1 × 10−3. Moreover, the AdamW optimiser was used to train the model for 50 epochs. The code was run on an NVIDIA A100 Tensor Core GPU, which is supported by the Aziz Supercomputer operated by the Center of Excellence in High-Performance Computing [68]. The base model was trained with three frame sizes: 32 × 32, 128 × 128, and 224 × 224. According to the results in Table 4, using a frame size of 224 × 224 produced better AUC performances. In the following subsections, we present the results of three experiments for deepfake video detection. These experiments aim to determine the effectiveness of transfer learning, image transformation, and texture-based features in enhancing detection accuracy across various manipulation methods and datasets. We conducted extensive experiments; however, for the sake of clarity in the manuscript, we present the best experimental results for the following:
  • Training from scratch vs. the pre-trained model.
  • Image adjustment using face alignment.
  • Training on Local Binary Pattern (LBP) features.
  • Results of Training from Scratch vs. Pre-trained Model
The decision to use pre-trained models in this experiment stems from the potential benefits they offer in terms of generalisation ability and performance. Training a deep learning model from scratch often requires substantial amounts of data and computational resources. However, leveraging pre-trained models is a more efficient approach, allowing knowledge learned from large-scale datasets to be transferred to a new task. This process, known as transfer learning, significantly reduces the time and resources required for training while enhancing the model’s ability to generalise across unseen data. While such pre-trained models may not have been trained for the purposes of the machine learning task at hand (in our case, fake detection), they might still be expected to usefully capture important image features (e.g., edges, textures, and some elements of facial shapes). A comparative analysis was conducted, involving three configurations: a model without pretraining, one pre-trained on CoAtNet on ImageNet [63], and a model pre-trained on VGG16 [59] (referred to as CoAtNet16).
As presented in Table 5, the pre-trained CoAtNet using VGG16 weights yielded the highest average AUC across all datasets, demonstrating its superior performance relative to the other models. In addition, several augmentation methods [65,66,67] were tested to improve the performance of the selected model. As observed in Table 5, CoAtNet16, with the augmentation method provided in [67] (CoAtNet16A), exhibits the highest performance among the others.
A classification threshold represents a specific value that indicates the manner in which a predictive model allocates class labels according to the output probabilities. In the context of binary classification, models frequently yield a probability score that reflects the likelihood of an instance being categorised as part of the positive class. The classification threshold serves as the cutoff point beyond which the instance is considered positive; conversely, instances falling below this threshold are classified as negative [69]. In this experiment, the performance of using a fixed threshold (=0.5) when calculating the accuracy for CoAtNet16A is compared with using a dynamic threshold during the training epochs. As observed in Table 6, using a static threshold produces better AUC results by about 3%. In deepfake detection, where fake and real classes often have overlapping probability distributions, dynamically adjusting the threshold results in the more frequent misclassification of borderline cases.
2.
Results with and without Face Alignment
Facial alignment encompasses the identification of specified reference points on the face, including the centres of the eyes, the corners of the mouth, and the tip of the nose. A geometric transformation is calculated utilising these reference points to guarantee that the identified facial features are positioned consistently throughout all images within the dataset [70].
This experiment is applied to verify the importance of using facial alignment with CoATtNet16A. As clarified in Table 7, adding facial alignment decreases the performance slightly by about 4%. The alignment process typically includes resizing, warping, or pixel interpolation, which can smooth out key visual inconsistencies that the model could use for classification.
3.
Results using Local Binary Pattern (LBP) features.
As LBP is well known for capturing texture details, it was employed to observe its effect in deepfake detection. As depicted in Table 8, using LBP almost does not affect deepfake detection when also using CoAtNet16A.
Using LBP with CoAtNet16A results in approximately the same performance as without LBP because CoAtNet already learns strong local and global features; in this context, LBP becomes redundant, and there is no need to use LBP features.
As a result of Stage 1, all features and parameters with the best performance in all cases were selected and used in the next experiments (Stage 2). The selected features and parameters are as follows: frame size: 224 × 224; using a CoAtNet model pre-trained on VGG16 with augmentation (CoAtNet16A); using a static threshold for accuracy; without using face alignment; and without using LBP features. These features and parameters are summarised in Table 9.

3.2. Stage 2: Performance Improvements

In the second stage of the experiments, the best model settings obtained from Stage 1 (Table 9) were used to apply multiple variations to improve the model’s performance. As depicted in Figure 9, the variations related to the selected frames for training and testing are either a single frame for training and testing or a single frame for training and multiple frames for testing. Finally, voting is applied to the results of the best approaches.
Frame selection is a critical step in deepfake detection. The rationale for using frame selection instead of processing all frames in a video relates to computational efficiency and avoiding redundant information. Processing every frame in a video significantly increases the computational cost and storage requirements without necessarily providing additional benefits for classification performance. This approach strikes a balance between efficiency and effectiveness, enabling the evaluation of large datasets within reasonable resource constraints [71].
Five approaches were evaluated for frame selection: a single middle frame, fifteen random frames, fifteen optical flow frames, cosine similarity keyframes, and facial landmark keyframes. For the single middle frame, the middle frame in each video was selected regardless of the video’s length. If the middle frame did not contain a face, then a search of the adjacent frames was carried out until a face was found. For fifteen random frames, fifteen random points were selected to select the frames at these points, ensuring that the selected frames cover various positions. This method aims to capture a diverse and potentially representative subset of frames. The third approach involved generating the optical flow, which is used to estimate the movement of objects between consecutive frames using Gunnar Farneback’s algorithm [72]. It works by analysing image intensity patterns at the pixel level. This method provides dense flow computations for each pixel, capturing both the direction and magnitude of motion. For this study, sixteen consecutive frames were extracted with a randomly selected initial frame, and OpenCV was utilised to calculate fifteen optical flow frames [73]. The flow is displayed in RGB channels to indicate the direction and magnitude of motion [74]. The other method for frame selection comprises the use of cosine similarity, which aims to select representative frames that summarise a video’s content, ensuring temporal and semantic diversity while reducing redundancy. This technique exploits the mathematical characteristics inherent in cosine similarity to quantify the degree of similarity among frames, thus facilitating the identification of frames that are most distinctive of the video sequence and thereby using limited representative frames instead of the entire video [75]. Finally, the last method is the extraction of facial landmark keyframes. It encompasses the identification and selection of frames from a video that exhibit prominent facial characteristics. This methodology employs facial landmark extraction techniques to ascertain critical reference points on the human face, including the eyes, nose, and mouth, which are essential for the recognition of individuals [76].
Figure 10 shows the different frame selection approaches used in this study.
The extracted frames using different approaches are used in training and testing via different methods, as explained in the following subsections.

3.2.1. Using a Single Middle Frame for Training and Testing

In this experiment, the CoAtNet16A model was evaluated on multiple deepfake detection datasets using a single middle frame from each video. The goal was to assess the performance of CoAtNet16A when limited to only one representative frame per video, and to determine its effectiveness in detecting deepfakes under such constraints. According to Table 10, the model demonstrates varied performances across datasets, ranging from 0.63 to a maximum of 0.85 when trained on FF++.

3.2.2. Using a Single Frame for Training and Multiple Frames for Testing

This part of the experiment aims to verify the effect of using multiple frames with different types. The following cases were examined: fifteen random frames, fifteen optical flow frames, cosine similarity keyframes, and finally, facial landmark keyframes. In all cases, the model was trained on single frames and tested using the voting of multiple frames.
  • Fifteen Random Frames
This experiment involved extracting fifteen random frames from each video training process as single-frame inputs and testing the voting of the fifteen random frames. As demonstrated in Table 11, training on FF++ yields the highest average AUC (0.8605). Conversely, training on datasets such as FaceShifter results in the lowest average AUC (0.5187), highlighting challenges in generalising to other datasets.
2.
Fifteen Optical Flow Frames.
In this experiment, the model was trained on a single-frame optical flow extracted and then evaluated using the majority for fifteen optical flow frames. The highest average AUC is 0.8346, as illustrated in Table 12.
3.
Cosine Similarity Keyframes.
This experiment investigated deepfake detection using keyframes selected based on cosine similarity. The model was trained on FF++ categories and evaluated across all datasets. The highest average AUC was about 0.85 on the FF++ dataset, as observed in Table 13.
4.
Facial Landmark-Based Key Frame Selection.
Facial landmark-based key frame selection is a technique used in video analysis to identify the most informative frames in a video based on facial landmark movements [76]. This experiment explored the effect of using facial landmarks as keyframes. The model was trained on FF++ categories and evaluated across all datasets. The highest average AUC was about 0.84 on the FF++ dataset, as observed in Table 14.
Among the different types of multiple frames, the best average AUC result was obtained using fifteen random frames on the FF++ dataset, which equals approximately 0.86, as shown in Table 11.

3.2.3. Using Voting for the Best Results

Among the different experiments that were applied using various settings and frame selection, the best ones were selected in order to apply the average voting technique. Our voting strategy is applied by obtaining the prediction for each item in the dataset based on the result of a specific trained model. Then, the average for different model predictions for each item is calculated. If the average is greater than or equal to 0.5, then the item is considered label 1; otherwise, it is considered label 0.
For each voting process, three models were selected, which were trained on a single middle frame, fifteen random frames, and fifteen optical flow frames; these were trained either on FF++ or only on part of FF++, called DeepFakes. As we used three trained models, there are four different combinations of voting, which comprise voting on all three models; voting on fifteen random frames and fifteen optical flow frames; voting on a single middle frame and fifteen random frames; and finally, voting on a single middle frame and fifteen optical flow frames. As depicted in Table 15, the best performance for FF++ is obtained using voting for all (single frame, fifteen random frames, and fifteen optical flow frames), which equals 0.9996; in contrast, using voting of all when trained on the DeepFakes dataset provides the best AUC for DFDC, which equals 0.5705, and for Celeb-DF, the best AUC equals 0.7624, which was obtained when using the voting of fifteen random frames and fifteen optical flow frames.

4. Performance Evaluation Comparison

In the experimental evaluation, various state-of-the-art methods were compared with our proposed CoAtNet models on the selected datasets—including FF++, DFDC Preview, and Celeb-DF—to assess their effectiveness in deepfake detection. The DFDC Preview dataset, a publicly available subset of the larger DFDC dataset, comprises over 5000 labelled videos, including both real and fake videos [77]. As the results of the state-of-the-art studies reported in [40] were applied to DFDC Preview and not DFDC, we adopted the same dataset to ensure comparability and the fair evaluation of our model against previously published research. The comparative performance of baseline methods and proposed CoAtNet variations is summarised in Section 4.3.

4.1. Intra-Dataset Comparison (Performance on FF++ Dataset)

Our proposed model, leveraging voting of all (single frame, fifteen random frames, and fifteen optical flow frames), achieved an exceptional AUC of 0.9996, outperforming all baseline methods. For example, Xception [78] and the Deep Convolutional Pooling Transformer [40] produced AUC scores of 0.9651 and 0.9766, respectively. This improvement suggests that the use of voting for all is highly effective for intra-dataset training and testing scenarios.

4.2. Assessment of Generalisation Through Cross-Dataset Comparison

To evaluate the generalisability of our proposed models beyond the training dataset, we conducted cross-dataset evaluations by training on FF++ and testing on two unseen datasets: DFDC Preview and Celeb-DF. This process simulates real-world scenarios where a model may encounter manipulated videos that differ in synthesis techniques from the training data. The results are organised into three subsections: the first delineating performance metrics on the DFDC dataset, the second addressing the Celeb-DF dataset, and the third one measuring the generalisation gap, followed by a comprehensive analytical summary.

4.2.1. Generalisation to the DFDC Dataset

On the DFDC dataset, the performance of our model produced an AUC of 0.6781, which is the third-highest performance. In fact, this result aligns with other methods, which also exhibited performance drops when tested on DFDC. For instance, the authors of [35] achieved an AUC of 0.7368, while the authors of [78] achieved an AUC of 0.6695. These results indicate that generalisation to cross-dataset scenarios remains challenging, especially for the DFDC dataset, even for high-performing models on FF++. The reason for this is related to the different faking types, as explained in Section 3.1.1.

4.2.2. Generalisation on the Celeb-DF Dataset

Our model achieved the highest performance on Celeb-DF, with an AUC of 0.7624. Simultaneously, methods such as Capsule Networks [35] and the Deep Convolutional Pooling Transformer [40] achieved AUC values of 0.6586 and 0.7243, respectively.

4.2.3. Overall Generalisation

To assess the generalisation gap between the intra-dataset and cross-dataset, the following formula is used:
Generalisation Gap = (performance on intra-datasetperformance on cross-dataset) × 100
For comparison with prior studies, we selected only models that achieved over 80% in both accuracy and AUC. Table 16 presents a comparison of the generalisation gap between our model (CoAtNet16A, trained on DF and using voting from 15 random RGB frames and 15 optical flow frames) and the selected methods. As shown in Figure 11, our model achieves the lowest generalisation gap in accuracy for both the DFDC and Celeb-DF datasets, with gaps of 19% and 17.7%, respectively. For the AUC, our method yields a gap of 25.6% on the DFDC dataset and achieves the best performance on Celeb-DF, with a minimal gap of 7.3%—significantly outperforming the other methods.

4.3. Comparative Analysis

Among all methods, the performance of our model stands out on the FF++ dataset, with an AUC of 0.9996, as in Table 17. The performance of different methods in the cross-dataset evaluations reveals, in general, a drop in both accuracy and AUC values. This observation highlights the generalisation gap in existing methods. Meanwhile, our suggested method produced the best results for the Celeb-DF dataset and the third-best result for the DFDC dataset.
The performance comparison between the proposed CoAtNet16A model and existing baseline methods across multiple datasets is illustrated in Figure 12.

5. Discussion and Limitations

The performance evaluation results presented in this study demonstrate that our proposed model, CoAtNet16A—particularly the “Voting of All” variant—achieves superior accuracy and AUC scores on the FF++ dataset, outperforming state-of-the-art methods such as Xception [78] and the Deep Convolutional Pooling Transformer [40]. This performance underscores the strength of multi-frame and multi-modal voting strategies for intra-dataset detection tasks.
In cross-dataset evaluations, the performance of our models exhibited a noticeable decline, particularly relative to the DFDC dataset. Despite being among the top-performing models, our best AUC was 0.6781. This reflects a trend in deepfake detection research, where generalisation across datasets remains a significant challenge due to variations in deepfake generation techniques. Conversely, the model performed relatively well on the Celeb-DF dataset, achieving an AUC of 0.7624 and outperforming notable baselines such as Xception [78] and the Deep Convolutional Pooling Transformer [40].
These findings highlight two key contributions of this study: first, an ensemble voting strategy with CoAtNet16A that enhances performance in intra-dataset contexts and, second, the best performance on Celeb-DF, suggesting that our model captures subtle, generalisable deepfake characteristics better than some existing methods.
However, this study also has several limitations:
  • Cross-dataset generalisation: Despite improvements, our model, similarly to others, exhibits performance degradation when applied to datasets on which it was not trained. This underscores the need for models capable of learning more generalised deepfake features that are invariant across datasets.
  • Computational complexity: The ensemble approach, while effective, increases computational requirements due to the processing of multiple frames and modalities. This may hinder its applicability for real-time detection or on resource-constrained devices.

6. Conclusions

This study assessed the generalisation ability of the CoAtNet model in deepfake video detection using both intra-dataset and cross-dataset evaluations. Our strongest finding is that CoAtNet16A achieved an AUC of 0.9996 on the FaceForensics++ dataset, outperforming existing state-of-the-art models such as Xception (0.9651) and the Deep Convolutional Pooling Transformer (0.9766). For the cross-dataset scenario, our model attained the highest AUC of 0.7624 on the Celeb-DF dataset and a third-highest AUC of 0.6781 on the DFDC Preview dataset, demonstrating superior generalisation across different manipulation techniques and source distributions.
The media and content verification sector, especially in the context of journalism, law enforcement, and social media platforms, is poised to gain the most significant advantages from implementing advanced deepfake detection solutions. Automated and precise detection instruments, such as CoAtNet16A, can contribute to the preservation of public trust by identifying synthetic or manipulated content before it goes viral.
The most vulnerable sectors encompass politics, finance, and public safety, where deepfakes can be exploited for misinformation, impersonation, and fraudulent activities. Esteemed individuals, organisations, and platforms may encounter both reputational and legal risks. To mitigate the potential for such exploitation, it is imperative to implement a comprehensive defence strategy that incorporates advanced deepfake detection systems, user awareness initiatives, and transparent content origin verification.
Future research should explore the common features of various faking techniques that can be used to enhance the detection model. These developments are essential for creating trustworthy deepfake detection algorithms that can handle the quickly changing synthetic media ecosystem and guarantee their effectiveness in real-world applications.

Author Contributions

Conceptualisation, all; literature review, E.A.; investigation, E.A.; methodology, all; project administration, J.C. and S.K.J.; resources, E.A.; software, E.A.; supervision, J.C., S.K.J. and A.A.-A.; validation, J.C. and S.K.J.; visualisation, E.A.; writing—original draft, E.A.; writing—review and editing, J.C., S.K.J., and A.A.-A. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the General Authority for Defence Development (GADD) in Saudi Arabia through project number (GADD_2024_01_0204).

Institutional Review Board Statement

The study was conducted and approved by the University Research Ethics Committee of the University of Sheffield (Reference Numbers (056436, 056518) and date of approval (27 July 2023–2 August 2023).

Informed Consent Statement

Not applicable.

Data Availability Statement

Forensics++ and FaceShifter datasets are available under a non-commercial research licensing agreement [79]. Celeb-DF can also be accessed by agreement [80]. DFDC was released by Facebook and can be accessed freely [60] after creating an AWS account [81] and an IAM user account [82].

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
GANGenerative Adversarial Network
CVComputer Vision
DNNDeep Neural Network
AEAutoencoder
CNNConvolutional Neural Network
ViTVision Transformer
EDEncoder–Decoder
NLPNatural Language Processing
CVComputer Vision
MSAMulti-Headed Self-Attention
MLPMulti-Layer Perceptron
FFNFeed-Forward Network
LBPLocal Binary Pattern
FF++FaceForensics++
DFDeepFakes
F2FFace2Face
FSFaceSwap
NTNeuralTextures
DFDCDeepFake Detection Challenge
DFAEDeepfake Autoencoder
MM/NNMorphable Model/Neural Network
NTHNeural Talking Head
FSGANFaceSwapGAN
TTSText-to-Speech
FShFaceShifter
AUCArea Under the Curve

References

  1. Shelke, N.A.; Kasana, S.S. A comprehensive survey on passive techniques for digital video forgery detection. Multimed. Tools Appl. 2020, 80, 6247–6310. [Google Scholar] [CrossRef]
  2. Lyu, S. Deepfake Detection: Current Challenges and Next Steps. In Proceedings of the 2020 IEEE International Conference on Multimedia & Expo Workshops (ICMEW), London, UK, 6–10 July 2020. [Google Scholar]
  3. Rana, S.; Sung, A.H. DeepfakeStack: A Deep Ensemble-based Learning Technique for Deepfake Detection. In Proceedings of the 2020 7th IEEE International Conference on Cyber Security and Cloud Computing (CSCloud)/2020 6th IEEE International Conference on Edge Computing and Scalable Cloud (EdgeCom), New York, NY, USA, 1–3 August 2020; pp. 70–75. [Google Scholar] [CrossRef]
  4. Tolosana, R.; Vera-Rodriguez, R.; Fierrez, J.; Morales, A.; Ortega-Garcia, J. Deepfakes and beyond: A Survey of face manipulation and fake detection. Inf. Fusion 2020, 64, 131–148. [Google Scholar] [CrossRef]
  5. Malavida, FakeApp 2.2.0. 2022. Available online: https://www.malavida.com/en/soft/fakeapp/ (accessed on 1 March 2022).
  6. Dfaker, df. 2018. Available online: https://github.com/dfaker/df (accessed on 1 March 2022).
  7. Shaoanlu, Faceswap-GAN. GitHub. 2018. Available online: https://github.com/shaoanlu/faceswap-GAN (accessed on 1 March 2022).
  8. Deepfakes, Faceswap. 2021. Available online: https://github.com/deepfakes/faceswap (accessed on 1 March 2022).
  9. Iperov, DeepFaceLab. 2021. Available online: https://github.com/iperov/DeepFaceLab (accessed on 1 March 2022).
  10. BuzzFeedVideo, You Won’t Believe What Obama Says in This Video! ? YouTube. Available online: https://www.youtube.com/watch?v=cQ54GDm1eL0 (accessed on 1 March 2022).
  11. Vecanoi, Very Realistic Tom Cruise Deepfake|AI Tom Cruise. YouTube. 2021. Available online: https://www.youtube.com/watch?v=iyiOVUbsPcM (accessed on 1 March 2022).
  12. Mirsky, Y.; Lee, W. The Creation and Detection of Deepfakes. ACM Comput. Surv. 2021, 54, 1–41. [Google Scholar] [CrossRef]
  13. Sanjana, S.; Thushara, P.; Karthik, P.C.; Vijayan, M.P.A.; Wilson, A. Review of Deepfake Detection Techniques. Int. J. Eng. Res. Technol. 2021, 10, 813–816. [Google Scholar]
  14. Aduwala, S.A.; Arigala, M.; Desai, S.; Quan, H.J.; Eirinaki, M. Deepfake Detection using GAN Discriminators. In Proceedings of the 2021 IEEE Seventh International Conference on Big Data Computing Service and Applications (BigDataService), Oxford, UK, 23–26 August 2021; pp. 69–77. [Google Scholar] [CrossRef]
  15. ZAO, ZAO APP. Available online: https://zaodownload.com/ (accessed on 1 March 2022).
  16. Synthesia, Synthesia—AI Video Generation Platform. Available online: https://www.synthesia.io/ (accessed on 1 March 2022).
  17. Video Dialogue Replacement|Canny. Available online: https://www.cannyai.com/ (accessed on 1 March 2022).
  18. Danry, V.; Leong, J.; Pataranutaporn, P.; Tandon, P.; Liu, Y.; Shilkrot, R.; Punpongsanon, P.; Weissman, T.; Maes, P.; Sra, M. AI-Generated Characters: Puting Deepfakes to Good Use. In Proceedings of the CHI EA ‘22: CHI Conference on Human Factors in Computing Systems Extended Abstracts, New Orleans, LA, USA, 29 April –5 May 2022; pp. 2–6. [Google Scholar] [CrossRef]
  19. Agarwal, S.; Varshney, L.R. Limits of Deepfake Detection: A Robust Estimation Viewpoint. arXiv 2019, arXiv:1905.03493. [Google Scholar]
  20. Dayal, S.; Niekerk, B. Deepfake Video Detection. In Proceedings of the ECCWS 2021 20th European Conference on Cyber Warfare and Security. Academic Conferences Inter Ltd., Virtual Event, 24–25 June 2021. [Google Scholar]
  21. Y. D. [D-N.-9] Rep. Clarke, H.R.5586—118th Congress (2023–2024): DEEPFAKES Accountability Act. 2023. Available online: https://www.congress.gov/bill/118th-congress/house-bill/5586/text (accessed on 11 May 2025).
  22. Bill C-63: An Act to Enact the Online Harms Act. Available online: https://www.justice.gc.ca/eng/csj-sjc/pl/charter-charte/c63.html (accessed on 11 May 2025).
  23. China to Regulate Deep Synthesis (Deepfake) Technology from 2023. Available online: https://www.china-briefing.com/news/china-to-regulate-deep-synthesis-deep-fake-technology-starting-january-2023/ (accessed on 11 May 2025).
  24. Thakur, R.; Rohilla, R. Recent advances in digital image manipulation detection techniques: A brief review. Forensic Sci. Int. 2020, 312, 110311. [Google Scholar] [CrossRef]
  25. Zhang, X.; Karaman, S.; Chang, S.F. Detecting and Simulating Artifacts in GAN Fake Images. In Proceedings of the 2019 IEEE International Workshop on Information Forensics and Security (WIFS), Delft, The Netherlands, 9–12 December 2019; pp. 1–6. [Google Scholar] [CrossRef]
  26. Yu, P.; Xia, Z.; Fei, J.; Lu, Y. A Survey on Deepfake Video Detection. IET Biom. 2021, 10, 607–624. [Google Scholar] [CrossRef]
  27. Rossler, A.; Cozzolino, D.; Verdoliva, L.; Riess, C.; Thies, J.; Nießner, M. FaceForensics++: Learning to detect manipulated facial images. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 1–11. [Google Scholar] [CrossRef]
  28. Dolhansky, B.; Bitton, J.; Pflaum, B.; Lu, J.; Howes, R.; Wang, M.; Ferrer, C.C. The DeepFake Detection Challenge (DFDC) Dataset. arXiv 2020, arXiv:2006.07397. [Google Scholar]
  29. Li, Y.; Yang, X.; Sun, P.; Qi, H.; Lyu, S. Celeb-DF: A Large-Scale Challenging Dataset for DeepFake Forensics. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 3204–3213. [Google Scholar] [CrossRef]
  30. Li, L.; Bao, J.; Yang, H.; Chen, D.; Wen, F. Advancing high fidelity identity swapping for forgery detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 14–19 June 2020; pp. 5073–5082. [Google Scholar] [CrossRef]
  31. Masood, M.; Nawaz, M.; Malik, K.M.; Javed, A.; Irtaza, A. Deepfakes Generation and Detection: State-of-the-art, open challenges, countermeasures, and way forward. Appl. Intell. 2022, 53, 3974–4026. [Google Scholar] [CrossRef]
  32. Kana, M. Generative Adversarial Network (GAN) for Dummies—A Step by Step Tutorial|by Michel Kana, Ph.D|Towards Data Science. Available online: https://towardsdatascience.com/generative-adversarial-network-gan-for-dummies-a-step-by-step-tutorial-fdefff170391 (accessed on 13 March 2022).
  33. Saiwa, Deepfakes Detection—Visual Deepfakes: Part II. Medium. 2024. Available online: https://medium.com/@saiwadotai/deepfakes-detection-visual-deepfakes-part-ii-68989616e2cc (accessed on 18 February 2025).
  34. Afchar, D.; Nozick, V.; Yamagishi, J.; Echizen, I. MesoNet: A compact facial video forgery detection network. In Proceedings of the 2018 IEEE International Workshop on Information Forensics and Security (WIFS), Hong Kong, China, 11–13 December 2018; pp. 1–7. [Google Scholar]
  35. Nguyen, H.H.; Yamagishi, J.; Echizen, I. Use of a capsule network to detect fake images and videos. arXiv 2019, arXiv:1910.12467. [Google Scholar]
  36. Dang, H.; Liu, F.; Stehouwer, J.; Liu, X.; Jain, A.K. On the Detection of Digital Face Manipulation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 14–19 June 2020; pp. 5780–5789. [Google Scholar] [CrossRef]
  37. Wodajo, D.; Atnafu, S. Deepfake Video Detection Using Convolutional Vision Transformer. arXiv 2021, arXiv:2102.11126. [Google Scholar]
  38. Zhao, H.; Wei, T.; Zhou, W.; Zhang, W.; Chen, D.; Yu, N. Multi-attentional Deepfake Detection. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 2185–2194. [Google Scholar] [CrossRef]
  39. Luo, Y.; Zhang, Y.; Yan, J.; Liu, W. Generalizing Face Forgery Detection with High-frequency Features. In Proceedings of the 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021; pp. 16312–16321. [Google Scholar] [CrossRef]
  40. Wang, T.; Cheng, H.; Chow, K.P.; Nie, L. Deep Convolutional Pooling Transformer for Deepfake Detection. ACM Trans. Multimedia Comput. Commun. Appl. 2023, 19, 1–20. [Google Scholar] [CrossRef]
  41. Purwono, P.; Ma’Arif, A.; Rahmaniar, W.; Fathurrahman, H.I.K.; Frisky, A.Z.K.; Haq, Q.M.U. Understanding of Convolutional Neural Network (CNN): A Review. Int. J. Robot. Control. Syst. 2022, 2, 739–748. [Google Scholar] [CrossRef]
  42. Rana, S.; Nobi, M.N.; Murali, B.; Sung, A.H. Deepfake Detection: A Systematic Literature Review. IEEE Access 2022, 10, 25494–25513. [Google Scholar] [CrossRef]
  43. Khormali, A.; Yuan, J.S. DFDT: An End-to-End DeepFake Detection Framework Using Vision Transformer. Appl. Sci. 2022, 12, 2953. [Google Scholar] [CrossRef]
  44. Kabbai, L.; Abdellaoui, M.; Douik, A. Image classification by combining local and global features. Vis. Comput. 2019, 35, 679–693. [Google Scholar] [CrossRef]
  45. Neves, J.C.; Tolosana, R.; Vera-Rodriguez, R.; Lopes, V.; Proença, H.P.; Fierrez, J. GANprintR: Improved Fakes and Evaluation of the State of the Art in Face Manipulation Detection. IEEE J. Sel. Top. Signal Process. 2020, 14, 1038–1048. [Google Scholar] [CrossRef]
  46. Merritt, R. What Is a Transformer Model? NVIDIA Blog. 2022. Available online: https://blogs.nvidia.com/blog/2022/03/25/what-is-a-transformer-model/ (accessed on 14 July 2022).
  47. Khan, S.A.; Dai, H. Video Transformer for Deepfake Detection with Incremental Learning. In Proceedings of the MM 2021—MM ‘21: Proceedings of the 29th ACM International Conference on Multimedia, Virtual Event, 20–24 October 2021; pp. 1821–1828. [Google Scholar] [CrossRef]
  48. Zhu, X.; Su, W.; Lu, L.; Li, B.; Wang, X.; Dai, J. Deformable DETR: Deformable Transformers for End-to-End Object Detection. arXiv 2020, arXiv:2010.04159. [Google Scholar]
  49. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale. arXiv 2020, arXiv:2010.11929. [Google Scholar]
  50. Sun, C.; Myers, A.; Vondrick, C.; Murphy, K.; Schmid, C. VideoBERT: A joint model for video and language representation learning. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Republic of Korea, 27 October 2019–2 November 2019; pp. 7463–7472. [Google Scholar] [CrossRef]
  51. Chen, J.; Lu, Y.; Yu, Q.; Luo, X.; Adeli, E.; Wang, Y.; Lu, L.; Yuille, A.L.; Zhou, Y. TransUNet: Transformers Make Strong Encoders for Medical Image Segmentation. arXiv 2021, arXiv:2102.04306. [Google Scholar]
  52. Cornia, M.; Stefanini, M.; Baraldi, L.; Cucchiara, R. Meshed-memory transformer for image captioning. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 10575–10584. [Google Scholar] [CrossRef]
  53. Ren, F.; Zhou, Y. CGMVQA: A New Classification and Generative Model for Medical Visual Question Answering. IEEE Access 2020, 8, 50626–50636. [Google Scholar] [CrossRef]
  54. Dai, Z.; Liu, H.; Le, Q.V.; Tan, M. CoAtNet: Marrying Convolution and Attention for All Data Sizes. Adv. Neural Inf. Process. Syst. 2021, 5, 3965–3977. [Google Scholar]
  55. Kaddar, B.; Fezza, S.A.; Hamidouche, W.; Akhtar, Z.; Hadid, A. HCiT: Deepfake Video Detection Using a Hybrid Model of CNN features and Vision Transformer. In Proceedings of the 2021 International Conference on Visual Communications and Image Processing (VCIP), Munich, Germany, 5–8 December 2021; pp. 1–5. [Google Scholar] [CrossRef]
  56. Zhang, D.; Li, J.; Shan, Z. Implementation of Dlib deep learning face recognition technology. In Proceedings of the 2020 International Conference on Robots & Intelligent System (ICRIS), Sanya, China, 7–8 November 2020; pp. 88–91. [Google Scholar] [CrossRef]
  57. Ruseckas, J. Coatnet. Available online: https://juliusruseckas.github.io/ml/coatnet.html (accessed on 25 October 2023).
  58. Krizhevsky, A.; Nair, V.; Hinton, G. CIFAR-10 and CIFAR-100 Datasets. 2009. Available online: https://www.cs.toronto.edu/~kriz/cifar.html (accessed on 2 December 2024).
  59. Rossler, A. FaceForensics. Available online: https://github.com/ondyari/FaceForensics (accessed on 25 October 2023).
  60. DFDC. Available online: https://dfdc.ai/ (accessed on 24 October 2023).
  61. Li, Y. Celeb-Deepfakeforensics, GitHub. Available online: https://github.com/yuezunli/celeb-deepfakeforensics (accessed on 25 October 2023).
  62. Polyak, A.; Wolf, L.; Taigman, Y. TTS skins: Speaker conversion via ASR. arXiv 2019, arXiv:1904.08983. [Google Scholar] [CrossRef]
  63. timm/coatnet_3_rw_224.sw_in12k. Hugging Face. Available online: https://huggingface.co/timm/coatnet_3_rw_224.sw_in12k (accessed on 13 February 2025).
  64. Torchvision Main Documentation. Available online: https://pytorch.org/vision/main/models/generated/torchvision.models.vgg16.html (accessed on 13 February 2025).
  65. Chen, H. DFGC-2022-1st-Place. Available online: https://github.com/chenhanch/DFGC-2022-1st-place (accessed on 25 October 2023).
  66. Bondi, L.; Cannas, E.D.; Bestagini, P.; Tubaro, S. Training Strategies and Data Augmentations in CNN-based DeepFake Video Detection. In Proceedings of the 2020 IEEE International Workshop on Information Forensics and Security (WIFS), New York, NY, USA, 6–11 December 2020. [Google Scholar] [CrossRef]
  67. Transforming and Augmenting Images—Torchvision Main Documentation. Available online: https://pytorch.org/vision/main/transforms.html (accessed on 25 October 2023).
  68. Center of Excellence in High Performance Computing Center. Available online: https://hpc.kau.edu.sa/ (accessed on 11 November 2024).
  69. Ebner, J. Classification Threshold, Explained—R-Craft. 2023. Available online: https://r-craft.org/classification-threshold-explained/?utm_source=chatgpt.com (accessed on 23 December 2024).
  70. Onaran, E.; Sarıtaş, E.; Ekenel, H.K. Impact of Face Alignment on Face Image Quality. 2024, pp. 1–15. Available online: http://arxiv.org/abs/2412.11779 (accessed on 23 December 2024).
  71. Lalitha, S.; Sooda, K. DeepFake Detection Through Key Video Frame Extraction using GAN. In Proceedings of the 2022 International Conference on Automation, Computing and Renewable Systems (ICACRS), Pudukkottai, India, 13–15 December 2022; pp. 859–863. [Google Scholar] [CrossRef]
  72. Farnebäck, G. Two-Frame Motion Estimation Based on Polynomial Expansion. In Proceedings of the Scandinavian Conference on Image Analysis, Halmstad, Sweden, 29 June–2 July 2013; Volume 1716, pp. 363–370. [Google Scholar]
  73. OpenCV Object Tracking. 2018. Available online: https://docs.opencv.org/4.x/dc/d6b/group__video__track.html#gaffe9a09dd582f9264c31da4ee8d1f41a (accessed on 21 October 2024).
  74. Chintha, A.; Rao, A.; Sohrawardi, S.; Bhatt, K.; Wright, M.; Ptucha, R. Leveraging edges and optical flow on faces for deepfake detection. In Proceedings of the 2020 IEEE International Joint Conference on Biometrics (IJCB), Houston, TX, USA, 28 September 2020–1 October 2020. [Google Scholar] [CrossRef]
  75. Gornale, S.S.; Babaleshwar, A.K.; Yannawar, P.L. Extraction of Key Frame from Random Videos Based On Discrete Cosine Transformation. Commun. Comput. Inf. Sci. 2021, 1380, 270–278. [Google Scholar] [CrossRef]
  76. Gunawan, A.; Widyantoro, D.H. Key frame extraction with face biometric features in multi-shot human re-identification system. In Proceedings of the 2019 International Conference on Advanced Computer Science and Information Systems, ICACSIS 2019, Bali, Indonesia, 12–13 October 2019; pp. 139–144. [Google Scholar] [CrossRef]
  77. Dolhansky, B.; Howes, R.; Pflaum, B.; Baram, N.; Ferrer, C.C. The deepfake detection challenge (DFDC) preview dataset. arXiv 2019, arXiv:1910.08854. [Google Scholar]
  78. Chollet, F. Xception: Deep learning with depthwise separable convolutions. In Proceedings of the 30th IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 1251–1258. [Google Scholar]
  79. Term to Use FaceForensics, FaceForensics++, and DeepFakes Detection Dataset. Available online: https://docs.google.com/forms/d/e/1FAIpQLSdRRR3L5zAv6tQ_CKxmK4W96tAab_pfBu2EKAgQbeDVhmXagg/viewform?pli=1 (accessed on 11 May 2025).
  80. Terms to Use Celeb-DF. Available online: https://docs.google.com/forms/d/e/1FAIpQLScoXint8ndZXyJi2Rcy4MvDHkkZLyBFKN43lTeyiG88wrG0rA/viewform (accessed on 11 May 2025).
  81. AWS Console—Signup. Available online: https://signin.aws.amazon.com/signup?request_type=register (accessed on 12 May 2025).
  82. Create an IAM User in Your AWS Account—AWS Identity and Access Management. Available online: https://docs.aws.amazon.com/IAM/latest/UserGuide/id_users_create.html (accessed on 12 May 2025).
Figure 1. Deepfake manipulation types: (top) head puppetry, (middle) face swapping, and (bottom) lip syncing. Source: [2].
Figure 1. Deepfake manipulation types: (top) head puppetry, (middle) face swapping, and (bottom) lip syncing. Source: [2].
Jimaging 11 00194 g001
Figure 2. Vision Transformer (ViT) architecture. Source: [49].
Figure 2. Vision Transformer (ViT) architecture. Source: [49].
Jimaging 11 00194 g002
Figure 3. CoAtNet architecture. Source: [54].
Figure 3. CoAtNet architecture. Source: [54].
Jimaging 11 00194 g003
Figure 4. Proposed framework for evaluating the generalisation of the CoAtNet model.
Figure 4. Proposed framework for evaluating the generalisation of the CoAtNet model.
Jimaging 11 00194 g004
Figure 5. FaceForensics++ dataset examples. Source: [59].
Figure 5. FaceForensics++ dataset examples. Source: [59].
Jimaging 11 00194 g005
Figure 6. DFDC dataset examples. Source: [59].
Figure 6. DFDC dataset examples. Source: [59].
Jimaging 11 00194 g006
Figure 7. Celeb-DF dataset examples. Source: [61].
Figure 7. Celeb-DF dataset examples. Source: [61].
Jimaging 11 00194 g007
Figure 8. FaceShifter dataset examples. Source: [54].
Figure 8. FaceShifter dataset examples. Source: [54].
Jimaging 11 00194 g008
Figure 9. Frame selection approaches for performance improvements.
Figure 9. Frame selection approaches for performance improvements.
Jimaging 11 00194 g009
Figure 10. Used frame selection approaches. Frames are extracted from Celeb-DF dataset.
Figure 10. Used frame selection approaches. Frames are extracted from Celeb-DF dataset.
Jimaging 11 00194 g010
Figure 11. CoAtNet16A’s performance generalisation gap for cross-dataset evaluation.
Figure 11. CoAtNet16A’s performance generalisation gap for cross-dataset evaluation.
Jimaging 11 00194 g011
Figure 12. AUC scores of CoAtNet16A model variants compared to state-of-the-art deepfake detectors.
Figure 12. AUC scores of CoAtNet16A model variants compared to state-of-the-art deepfake detectors.
Jimaging 11 00194 g012
Table 1. Comparison between CNN, ViT, and CoAtNet based on source [49].
Table 1. Comparison between CNN, ViT, and CoAtNet based on source [49].
PropertiesCNNViTCoAtNet
Translation Equivariance
Local Features
Input-adaptive Weighting
Global Features
Table 2. Well-known public datasets.
Table 2. Well-known public datasets.
DatasetPublished DateImages/VideosReal VideosFake VideosNumber of Faking MethodsFaking Algorithm Type(s)Visual/Audio/Both
FF++ Jan-19Videos100040004DF, F2F, FS, NTVisual
Celeb-DF (v2)Nov-19Videos59056391Improved DeepFake synthesis algorithmVisual
FShJun-20Videos100010001Two-stage FaceShifter face-swapping Visual
DFDCJun-20Videos23,954104,5008DF-128, DF-256, MM/NN, NTH, FSGAN, StyleGAN, Refinement, and AudioSwapsBoth
Table 3. Overlapping of deepfake datasets.
Table 3. Overlapping of deepfake datasets.
Faking TypeFF++Celeb-DFFShDFDC
Face SwapYes (DeepFakes, FaceSwap)Yes (improved version)Yes (improved version)Yes (DFAE, MM/NN, StyleGAN)
Head PuppetryYes (Face2Face, NeuralTextures)NoNoYes (FSGAN, NTH)
Lip SyncingNoNoNoYes (AudioSwap)
Table 4. Model performance (AUC) with different frame sizes. The models were trained on DeepFakes.
Table 4. Model performance (AUC) with different frame sizes. The models were trained on DeepFakes.
Frame SizeDFF2FFSNTDFDCCeleb-DFFShAVG AUC
32 × 320.92810.61630.71380.62450.63990.6320.69870.6933
128 × 1280.98360.86390.51680.87870.59310.62430.72620.7409
224 × 2240.99770.98460.59520.98270.54230.49010.98240.7964
Table 5. Comparison of average AUC scores across different pre-trained models. The models were trained on DeepFakes.
Table 5. Comparison of average AUC scores across different pre-trained models. The models were trained on DeepFakes.
Type DF F2F FS NT DFDC Celeb-DF FSh AVG AUC
No Pretraining0.99400.94820.44900.88860.51370.45630.74680.7138
Pre-trained CoAtNet0.99440.95150.48470.91700.48580.45960.76580.7227
CoAtNet160.99740.94700.49170.90680.50100.48860.73250.7236
CoAtNet16A 0.99770.98460.59520.98270.54230.49010.98240.7964
Table 6. Model performance (AUC) with static threshold (STh) and dynamic threshold (DTh). The model is trained on DeepFakes.
Table 6. Model performance (AUC) with static threshold (STh) and dynamic threshold (DTh). The model is trained on DeepFakes.
Threshold TypeDFF2FFSNTDFDCCeleb-DF FShAVG AUC
CoAtNet16A with STh0.99770.98460.59520.98270.54230.49010.98240.7964
CoAtNet16A with DTh0.99730.96860.54440.96690.51650.48670.89970.7686
Table 7. Model performance (AUC) with and without facial alignment. The model is trained on DeepFakes.
Table 7. Model performance (AUC) with and without facial alignment. The model is trained on DeepFakes.
Face Alignment StatusDFF2FFSNTDFDCCeleb-DF FShAVG AUC
CoAtNet16A without Face Alignment0.99770.98460.59520.98270.54230.49010.98240.7964
CoAtNet16A with Face Alignment0.99860.92700.47930.93420.56910.59210.80930.7585
Table 8. Model performance (AUC) with and without Local Binary Pattern (LBP) features. The models were trained on DeepFakes.
Table 8. Model performance (AUC) with and without Local Binary Pattern (LBP) features. The models were trained on DeepFakes.
Using LBP Status DF F2F FS NT DFDC Celeb-DF FSh AVG AUC
CoAtNet16A without LBP0.99770.98460.59520.98270.54230.49010.98240.7964
CoAtNet16A with LBP0.99970.98050.73160.98190.50030.56200.85180.8011
Table 9. Selected features and parameters with CoAtNet16A.
Table 9. Selected features and parameters with CoAtNet16A.
Frame SizeTransfer LearningThresholdingImage TransformationTexture-Based Features
224 × 224CoAtNet16A (CoAtNet model pre-trained on VGG16 with augmentation)StaticNoNo
Table 10. Performance (AUC) of the CoAtNet16A model on various deepfake detection datasets using a single middle frame per video.
Table 10. Performance (AUC) of the CoAtNet16A model on various deepfake detection datasets using a single middle frame per video.
Trained on/Tested on DF F2F FS NT DFDC Celeb-DF FSh AVG AUC FF++ AVG AUC All DS
DF0.99770.98460.59520.98270.54230.49010.98240.89010.7964
F2F0.96530.99660.8770.94890.49750.54010.96950.94700.8278
FS0.55580.85030.98120.4610.50460.53390.54250.71210.6328
NT0.9950.9920.64930.99520.49120.50010.9480.90790.7958
DFDC0.94970.71010.790.70590.81350.74730.69530.78890.7731
Celeb-DF0.82440.65770.44630.66830.61070.94390.53950.64920.6701
FSh0.82270.78550.55640.79840.51870.52430.99630.74080.7146
FF++0.9980.99760.99710.99730.47570.49240.99480.99750.8504
Table 11. Model performance (AUC) with training on a single frame and testing for fifteen random frames.
Table 11. Model performance (AUC) with training on a single frame and testing for fifteen random frames.
Trained on/Tested on DF F2F FS NT DFDC Celeb-DF FSh AVG AUC FF++ AVG AUC All DS
DF0.99980.84010.33450.92610.5620.67350.77440.77510.7301
F2F0.70550.98640.58410.61650.50290.56390.55730.72310.6452
FS0.3954230.7988240.9962930.3202760.5033740.5617680.4093330.62770.5766
NT0.96850.85670.42130.96370.47230.48390.74880.80260.7022
DFDC 0.94010.67650.81280.67320.84920.84550.62070.77570.774
Celeb-DF 0.81260.62240.43960.63210.63620.97330.55710.62670.6676
FSh0.47290.44760.3030.42570.51090.4760.99480.41230.5187
FF++0.99880.9960.99550.99570.55470.51670.96640.99650.8605
Table 12. Model performance (AUC) with training on a single optical flow frame and testing for fifteen optical flow frames.
Table 12. Model performance (AUC) with training on a single optical flow frame and testing for fifteen optical flow frames.
Trained on/Tested on DF F2F FS NT DFDC Celeb-DF FSh AVG AUC FF++ AVG AUC All DS
DF0.99780.78380.58030.82720.54070.73520.69750.79730.7375
F2F0.80130.94220.43240.77250.46470.56010.41610.73710.6270
FS0.85130.64920.99120.67430.54600.65640.53390.79150.7003
NT0.90150.83720.52720.93630.50140.61830.70310.80060.7179
FF++0.98460.95370.97870.92920.54450.74880.70290.96150.8346
Table 13. Model performance (AUC) with cosine similarity keyframes.
Table 13. Model performance (AUC) with cosine similarity keyframes.
Trained on/Tested on DF F2F FS NT DFDC Celeb-DF FSh AVG AUC FF++AVG AUC All DS
DF0.99900.92390.35140.93580.52790.58500.68650.80250.7157
F2F0.92750.99970.47290.71060.47050.55410.55010.77770.6693
FS0.57430.69470.99930.58620.48100.52060.58830.71360.6349
NT0.98990.93480.46230.98120.51660.53380.81500.84200.7477
FF++0.99680.99300.98890.98530.55920.54860.87690.99100.8498
Table 14. Model performance (AUC) with facial landmark keyframes.
Table 14. Model performance (AUC) with facial landmark keyframes.
Trained on/Tested on DF F2F FS NT DFDC Celeb-DF FSh AVG AUC FF++ AVG AUC All DS
DF0.99910.89030.37790.93820.51370.49410.84620.80140.7228
F2F0.96010.99830.54670.85450.52170.51400.62420.83990.7171
FS0.60110.57600.99790.57140.51490.48130.52870.68660.6102
NT0.99760.96270.45180.99440.50960.40580.95250.85160.7535
FF++0.99940.99870.99880.99600.50830.42770.93310.99820.8374
Table 15. Voting performance on the CoAtNet16A model (ACC: accuracy; AUC: area under the curve).
Table 15. Voting performance on the CoAtNet16A model (ACC: accuracy; AUC: area under the curve).
Frames TypeTrained OnTesting Datasets
FF++DFDCCeleb-DF
ACCAUCACCAUCACCAUC
Single Frame FF++ 98.16180.997549.53860.475749.46810.4924
15 Random Frames97.52620.996553.34040.554770.58220.5167
15 Optical Flow Frames90.43320.961551.63370.544533.06040.7488
Voting of All99.05840.999650.90550.537352.96210.6531
Voting of 15 Random Frames and 15 Optical Flow98.12270.994352.12010.565561.84830.6664
Voting of Single Frame and 15 Random Frames98.55810.9996 *50.50800.522948.57820.5032
Voting of Single Frame and 15 Optical Flow Frames98.62090.999050.37540.512351.00710.7233
Single Frame DF 83.57840.890152.64260.542349.52720.4901
15 Random Frames75.46710.775150.30950.562017.87540.6735
15 Optical Flow Frames69.52740.797349.96670.540715.66020.7352
Voting of All71.84100.8867 50.4196 0.5705 *50.53320.7033
Voting of 15 Random Frames and 15 Optical Flow68.46620.8353 50.2650 0.552950.77010.7624 *
Voting of Single Frame and 15 Random Frames78.10670.8658 50.9717 0.567451.18480.6005
Voting of Single Frame and 15 Optical Flow Frames79.36350.896550.53000.564350.29620.6782
* Bold indicates highest AUC result.
Table 16. Comparison of generalisation gap between CoAtNet16A and other studies.
Table 16. Comparison of generalisation gap between CoAtNet16A and other studies.
MethodAccuracyAUC
DFDCCeleb-DFDFDCCeleb-DF
FFD [36]22.8536.1023.0126.62
CViT [37]22.2929.7923.6527.48
MAT [38]24.3442.7225.2937.65
SRM [39]28.2435.2230.1334.03
Xception [78]31.3135.8429.5630.65
Deep Convolutional Pooling Transformer [40]26.3528.8423.9825.23
CoAtNet16A-Voting of 15 Random Frames-15 Optical Flow19.0017.7025.567.29
Table 17. ACC and AUC performance comparisons on each testing set after training on the FF++ dataset and the results of previous studies reported in [35].
Table 17. ACC and AUC performance comparisons on each testing set after training on the FF++ dataset and the results of previous studies reported in [35].
MethodTrained on
FF++DFDC PreviewCeleb-DF
ACCAUCACCAUCACCAUC
MesoNet [34]FF++61.030.581350.020.501636.730.5001
Capsule [35]FF++76.40.834451.30.561661.960.5993
FFD [36]FF++82.290.824859.440.594746.190.5586
CViT [37]FF++83.050.910860.760.674353.260.636
MAT [38]FF++87.50.948563.160.695644.780.572
SRM [39]FF++88.170.949359.930.64852.950.609
Xception [78]FF++90.080.965158.770.669554.240.6586
Deep Convolutional Pooling Transformer [40]FF++92.110.976665.760.7368 *63.270.7243
CoAtNet16A—Voting of AllFF++99.05840.9996 *51.61290.678152.96210.6531
CoAtNet16A—Voting of 15 Random Frames and 15 Optical Flow FramesFF++98.12270.994353.49460.651561.84830.6664
CoAtNet16A—Voting of AllDF71.84100.8867 49.7312 0.553050.53320.7033
CoAtNet16A—Voting of 15 Random Frames and 15 Optical FlowDF68.46620.8353 49.4624 0.579750.77010.7624 *
* Bold indicates highest AUC result.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Alattas, E.; Clark, J.; Al-Aama, A.; Jarraya, S.K. Evaluating Features and Variations in Deepfake Videos Using the CoAtNet Model. J. Imaging 2025, 11, 194. https://doi.org/10.3390/jimaging11060194

AMA Style

Alattas E, Clark J, Al-Aama A, Jarraya SK. Evaluating Features and Variations in Deepfake Videos Using the CoAtNet Model. Journal of Imaging. 2025; 11(6):194. https://doi.org/10.3390/jimaging11060194

Chicago/Turabian Style

Alattas, Eman, John Clark, Arwa Al-Aama, and Salma Kammoun Jarraya. 2025. "Evaluating Features and Variations in Deepfake Videos Using the CoAtNet Model" Journal of Imaging 11, no. 6: 194. https://doi.org/10.3390/jimaging11060194

APA Style

Alattas, E., Clark, J., Al-Aama, A., & Jarraya, S. K. (2025). Evaluating Features and Variations in Deepfake Videos Using the CoAtNet Model. Journal of Imaging, 11(6), 194. https://doi.org/10.3390/jimaging11060194

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop