Next Article in Journal
Simulation Analysis of Capacity Evaluation of Bus Stops under Connected and Automated Vehicles Environment
Previous Article in Journal
The Quality Control System of Planks Using Machine Vision
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Review

Detecting Deceptive Behaviours through Facial Cues from Videos: A Systematic Review

Consiglio Nazionale delle Ricerche (CNR)—IRPPS, 00185 Rome, Italy
*
Author to whom correspondence should be addressed.
Appl. Sci. 2023, 13(16), 9188; https://doi.org/10.3390/app13169188
Submission received: 10 June 2023 / Revised: 7 August 2023 / Accepted: 10 August 2023 / Published: 12 August 2023

Abstract

:
Interest in detecting deceptive behaviours by various application fields, such as security systems, political debates, advanced intelligent user interfaces, etc., makes automatic deception detection an active research topic. This interest has stimulated the development of many deception-detection methods in the literature in recent years. This work systematically reviews the literature focused on facial cues of deception. The most relevant methods applied in the literature of the last decade have been surveyed and classified according to the main steps of the facial-deception-detection process (video pre-processing, facial feature extraction, and decision making). Moreover, datasets used for the evaluation and future research directions have also been analysed.

1. Introduction

Lying is a complex social activity by someone (i.e., the deceiver) aimed at causing a specific behaviour in another person (i.e., the deceived) by making their view of the situation more congruent with the view and behaviour of the deceived that usually occurs, according to the deceiver’s knowledge [1]. Deception may destroy the relationship and hamper communication, leading to negative consequences [2]. Therefore, deception detection has been an investigated research topic for decades, and the development of tools and methods for detecting deceptive behaviours has become an urgent need for society. However, despite a rich corpus of deception research, detecting deceptive behaviours is still a very challenging task, mainly because human beings do not have good lie-detection abilities. A study by Bond and DePaulo [3] quantifies the human accuracy of discriminating between lies and truths at 54% on average, in other words just slightly above a random guess. This percentage is also confirmed by the Hartwig’s [4] more recent study. In response to the necessity to improve this accuracy rate, researchers have long been trying to decode human behaviour in an attempt to discover deceptive cues. In this study, what we consider “deceptive” is human behaviour in which the deceiver intentionally acts to make the deceived believe in something the deceiver considers false [5]. This is a conscious and deliberated act, opposed to the unconscious (non-deceptive) behaviour in which a person provides false information believed to be true.
To discover deceptive cues, different methods have been proposed by scientists from different disciplines, ranging from psychologists and physiologists to technologists. For instance, methods based on the psycho-physiological detection of deception that uses mainly psychological tests and physiological monitoring have been applied [6,7]. A positive contribution is also given by technologists that developed methods for deceptive detection based on technological tools. This study focuses on the technological field, in which a considerable number of methods for detecting verbal (explicit) and non-verbal (implicit) cues of deception have been developed [8]. Explicit cues are categorised according to their nature (i.e., visual, verbal, or paralinguistic) [9], while implicit cues involve facial expressions, body movements/posture, eye contact, and hand movements [10].
Detecting deceptive behaviours from explicit cues makes deception harder because explicit cues are easier to manipulate through carefully chosen language and wording, while implicit cues, even if also subject to manipulation, are more spontaneous and challenging to control consciously [11]. On the other hand, the interpretation of implicit cues tends to be more subjective and context-dependent, requiring a deeper understanding of individual differences and cultural norms, while explicit cues are typically more straightforward and easier to analyse in isolation. Finally, detection deception using implicit cues requires collecting data through video recordings or specialised sensors to capture non-verbal behaviours, which is more challenging compared to collecting text data for explicit cues.
Considering the above-mentioned differences, the solutions provided in the literature for detecting deceptive behaviours have different characteristics if relying on explicit or implicit cues. The systematic literature review provided in this paper deals with implicit cues and, in particular, focuses on a specific type of implicit cues, i.e., facial cues. Specifically, the most relevant deception-detection systems focusing on facial cues have been described and classified considering the methods used in the main steps of the facial-deception-detection process (i.e., video pre-processing, facial feature extraction, and decision making). The choice to focus on facial cues is justified by the results of several existing studies [12,13] arguing that facial cues have an important role in detecting deceptive behaviours due to their involuntary aspects. As stated by Darwin in his “inhibition hypothesis” [13], some actions of the facial muscles are the hardest to be inhibited due to the difficulty to control them voluntarily.
The main contributions of this review are as follows:
  • to collect and systematise the scientific knowledge related to automated face deception detection from videos;
  • to summarise (i) the methods used for video pre-processing, (ii) the facial extracted features, (iii) the decision-making algorithms, and (iv) the datasets used for the evaluation;
  • to point out future research directions that emerged from the analysis of the existing studies.
The paper is structured as follows. Section 2 introduces the related work. Section 3 illustrates the methodology followed for the systematic literature review, while in Section 4, the results are discussed. Section 5 describes how the surveyed studies answered the review questions defined in the paper. Section 6 summarises future research directions that emerged from the literature review. Finally, Section 7 concludes the paper.

2. Related Work

Automated deception detection from videos is a challenging task that can find applications in many real-world scenarios, including airport security screening, job interviews, court trials, and personal credit risk assessment [14].
Various surveys of automated deception-detection methods from videos have been developed in the literature, classifying the methods according to different dimensions. In particular, recent surveys (see Table 1) have provided a very broad overview of deception-detection methods considering monomodal and multimodal features. The survey developed by Constâncio et al. [15] provides a systematic literature review of deception detection based on machine-learning methods, underlying the techniques and approaches applied, the difficulties, the kind of data, and the performance levels achieved. In this survey, verbal and non-verbal cues, such as facial expressions, gestures, and prosodic and linguistic features, have been considered. The 81 surveyed papers are classified according to the type of extracted features (emotional, psychological, and facial) and the kinds of machine-learning algorithms. Analogously, the survey in [16] gives an overview of automated deception detection through machine-intelligence-based techniques by providing a critical analysis of the existing tools and available datasets. The authors focused on deception detection through text, speech, and video data analysis by classifying the 100 surveyed papers according to both the research domain (i.e., psychological, professional, and computational) and the type of extracted features (verbal, non-verbal, and multimodal). A further survey focused on monomodal features (speech cues) is proposed by Herchonvicz and de Santiago [17], in which deep-learning-based techniques, available datasets, and metrics extracted from the surveyed papers are discussed.
Table 1. Comparative analysis of surveys.
Table 1. Comparative analysis of surveys.
ReferenceCuesTechniquesDatasetsSteps of the Deception Detection Process (Figure 1)
Constâncio et al., 2023 [15]Verbal and non-verbal cuesMachine-learning methodsNot surveyedStep 2 and Step 3
Alaskar et al., 2023 [16]Verbal and non-verbal cuesMachine-intelligence-based techniquesSurveyedStep 2 and Step 3
Herchonvicz and de Santiago, 2021 [17]Verbal cuesDeep-learning-based techniquesSurveyedStep 2 and Step 3
Our surveyNon-verbal cuesMachine-learning methodsSurveyedStep 1, Step 2, and Step 3
Figure 1. The workflow of the deception detection process from videos followed in the review.
Figure 1. The workflow of the deception detection process from videos followed in the review.
Applsci 13 09188 g001
Different from existing surveys, this work focuses on deception detection methods specifically based on facial cues and proposes a classification of these methods based on the main steps of the facial deception detection process (video pre-processing, facial feature extraction, and decision making), as proposed by Thannoon et al. [18]. On the contrary, the existing surveys considered only the second and third steps of this process.
A general workflow of the followed deception detection process that used video datasets is provided in Figure 1.
During the video pre-processing step, the video is analysed by a face detector, which bounds the box containing the face, and by a facial-landmark-tracking tool, which locates and tracks the facial landmarks. In the facial-feature-extraction step, the extraction of a set of facial features for recognizing the facial cues meaningful to deception is carried out. Finally, the classification of the extracted features into truthful or deceptive behaviour is performed in the decision-making step. According to these steps, the studies included in the review are analysed to extract (i) the methods used for video pre-processing, (ii) the facial extracted features, (iii) the decision-making algorithms, and (iv) the datasets used for the evaluation.

3. Materials and Methods

This section illustrates the methodology for systematically searching and analysing the existing literature on automated face deception detection from videos. The followed methodology was the systematic literature review (SLR) described in the PRISMA recommendations [19] composed of the following steps: (1) identifying the review focus; (2) specifying the review question(s); (3) identifying studies to include in the review; (4) data extraction and study quality appraisal; (5) synthesizing the findings; and (6) reporting the results.
The review focus consists of the collection and systematization of the scientific knowledge related to automated deception detection from videos focusing on facial signals. Specifically, we aim to study the methods for video pre-processing, the facial extracted features, the decision-making algorithms, and the available datasets.
As required by the second step of the SLR process, the following review questions (RQ) were defined:
RQ 1: Which methods for video pre-processing are used?;
RQ 2: Which facial features for automated deception detection are extracted?;
RQ 3: Which decision-making algorithms for automated deception detection are used?;
RQ 4: Which datasets for automated deception detection are used?
In the third step of the SLR process, the identification of the studies to include in the review was carried out following the four phases of the PRISMA recommendations [19], as shown in Figure 2.
For the identification of an initial set of scientific works, two indexed scientific databases (Web of Science (WoS) and Scopus) containing formally published literature (e.g., journal articles, books, and conference papers) were used. Specifically, WoS and Scopus were used due to their proven usefulness in conducting bibliometric analyses of the existing literature in various research domains to establish trends [21]. Grey literature has not been included in the systematic review because there is no gold standard method for searching grey literature, nor is there a defined methodology for the same, which makes it all the more difficult [22].
The search strings defined to search the scientific papers on the databases are as follows: (“deception detection” OR “lie detection” OR “deceptive behaviour*” OR “lie behaviour*” OR “detect* deception” AND “video*” AND “facial expression*”). Note that this study does not disambiguate among the different forms of deception (e.g., bluff, mystification, propaganda, etc.) because it aims to find studies focused on deceptive behaviours in general without restricting the search to specific forms of deception.
During the screening phase, the obtained results were filtered according to the inclusion and exclusion criteria shown in Table 2.
The full text of the eligible studies was analysed by two reviewers using a quality evaluation checklist composed of four questions and related scores (see Table 3). The “disagreed” studies were evaluated by a moderator who provided the final scores.
Studies that obtained a score < 1.5 were excluded from the qualitative analysis, while studies that scored 1.5 or more were included in the review.
Finally, analysis of the full texts of the included studies was performed, extracting the following information (if any):
  • Pre-processing techniques of facial cues;
  • Facial features, facial action coding system, and action units;
  • Facial deception classification approaches;
  • Evaluation datasets
In the following sections, the findings resulting from the application of the SLR process are described.

4. Results of the SLR

As depicted in Figure 2, during the identification phase (performed in May 2023), a total of 230 articles were retrieved using the two search engines: 203 from Scopus and 27 from WoS.
As required by the duplication criterion, removing duplicate records resulted in 120 studies. By applying the understandability criterion, 6 studies that were not written in English were excluded, and therefore, a total of 114 articles were screened for the inclusion criteria. Moreover, the studies that did not satisfy the document type criterion (2 articles) and those that did not contain in the titles the terms used for the search (70 studies) were removed, and therefore, a total of 42 studies resulted at the end of the screening phase.
Four studies that were not accessible in full text (availability criterion) and six studies that were considered not relevant for the focus of the review (relevance criterion) were removed. The relevance criterion to the abstract was also applied, and therefore, a total of 32 articles were retained for a full evaluation of eligibility, evaluated by two reviewers according to the quality evaluation checklist shown in Table 3. Seven studies with a score less than 1.5 were excluded, while the remaining twenty-four studies were included in the qualitative synthesis. Table 4 provides an overview of the selected studies.
The selected studies were published mainly in conference proceedings (15 studies; 62.5%), followed by journals (9 studies; 37.5%). Figure 3 shows the temporal distribution of the selected studies, underlying the growing interest of the scientific community in the facial-deception-detection topic, which reached a peak in 2020.

5. Discussion

In this section, an analysis of the answers to the four review questions (introduced in Section 3) extracted from the 24 surveyed studies is carried out. Specifically, to deal with RQ1, we analyse the methods for video pre-processing. Addressing RQ2, the facial features are classified. Concerning RQ3, the decision-making algorithms used for automated deception detection are analysed. Finally, as part of RQ4, the used datasets are analysed.

5.1. RQ1: Which Methods for Video Pre-Processing Are Used?

Pre-processing consists of those operations that prepare data for subsequent analysis in an attempt to compensate for systematic errors. The videos are subjected to several corrections, such as geometric, radiometric, and atmospheric, although all these corrections might not be necessarily applied in all cases. These errors are systematic and can be removed before they reach the user. According to the nature of the information to be extracted from the videos, the most relevant pre-processing techniques are chosen.
Facial region detection, landmark localisation, and tracking are necessary for pre-processing when capturing videos in unconstrained recording conditions. This step has the aim of localizing and extracting the face region from the background and identifying the locations of the facial key landmark points. These points have been defined as “either the dominant points describing the unique location of a facial component (e.g., eye corner) or an interpolated point connecting those dominant points around the facial components and facial contour” [44]. Regardless of the application domain, face detection and landmark tracking from input video streams are preliminary steps necessary to process any kind of information extractable from the face.
Over the past few decades, several face-detection and landmark-tracking algorithms have been developed [33,43,45,46], as well as academic or commercial software that performs joint face detection and landmark localization [47,48,49,50,51,52,53,54].
The vast majority of these methods and tools are based on 2D image processing. The main drawback of 2D face-detection and landmark-tracking methods is that their performances depend strictly on the following four factors: the pose of the head that can generate self-occlusions and deformations; occlusions caused by masks, sunglasses, or other faces; illumination variations caused by internal camera control and skin reflectance properties; time delay caused by face changes over time. However, these methods have the advantages of being generally easier and less expensive in terms of computational load compared to 3D face-detection methods. This last class of methods uses a 3D dataset that contains both face and head shapes as range data or polygonal meshes [55]. The 3D methods allow the fundamental limitations of 2D methods to be overcome since they are insensitive to pose and illumination variations.
A widely accepted classification of face-detection techniques [56] distinguishes them into two categories: feature-based and image-based approaches. On the one hand, feature-based techniques make explicit use of face knowledge by extracting local or global features (e.g., eyes, nose, mouth) and use this knowledge to classify between facial and non-facial regions. The active shape model (ASM) [38] and the Viola–Jones method [33] are two of the most applied and successful feature-based methods so far. On the other hand, image-based approaches learn to recognise a face pattern from examples following a training step for classifying examples into face and non-face prototype classes. The only image-based approach, applied by the surveyed studies, is the Constrained Local Neural Field (CLNF) [46].
There is also academic and commercial software that performs joint face-detection and landmark localization, such as OpenFace [47], FaceSDK [49], IntraFace [48], ‘dlib’ library [50], OpenCV [51], PittPatt SDK [52], InsightFace [53], and FAN [54].
A total of 18 papers among those surveyed (24) use the face-detection and landmark-tracking methods represented in Table 5.
From the table, we can observe that during the video pre-processing step, mainly 2D methods are used (17 studies; 94%), while only two works apply 3D methods. Moreover, the majority of the studies (14 studies; 78%) apply academic and commercial software that allows joint face-detection and landmark localization tasks to be performed. OpenFace [45] is the open-source tool most applied by surveyed facial-deception-detection works (7 studies; 39%). The remaining three surveyed works use specific face-detection and face-tracking algorithms, mainly belonging to the feature-based category. The most applied algorithms are the Viola–Jones face detector [33] and the Constrained Local Neural Field (CLNF) [46] used for face detection and face tracking, respectively.

5.2. RQ2: Which Facial Features for Automated Deception Detection Are Extracted?

The feature-extraction process consists of transforming raw data into a set of features to be used as reliable indicators for detecting deceptive behaviours.
A multitude of facial cues has been correlated with deception in the literature, including lip pressing, eye blinking, facial rigidity, smiling, and head movement. A total of 22 papers among those surveyed (24 studies) extracted the following 10 facial features, represented in Table 6 with the respective number: (I) facial expression; (II) facial rigidity; (III) head pose; (IV) eye blink; (V) eyebrow motion; (VI) wrinkle occurrence; (VII) mouth motion; (VIII) gaze motion (IX); lip movement (X).
More than half of the studies (15 studies; 68%) rely on the Facial Action Coding System (FACS), an anatomically based system for representing all visually discernible facial movements proposed by Ekman et al. [42]. The FACS provides a list of action units (AUs) that are the individual components of muscle movement extracted from facial expressions. Ekman et al. [57] provided a list of AUs that are the most difficult to produce deliberately and the hardest to inhibit. Therefore, the majority of the analysed studies detect the movements of these AUs as potential indicators for distinguishing liars from truth-tellers in high-stakes situations.
A total of 18 studies (82%) rely on extracting facial expression features for detecting deceptive behaviours. Facial expressions, particularly micro-expressions (i.e., facial expressions of emotion that are expressed for 0.5 s or less), are often considered strong indicators of deception. Six basic facial expressions of emotion are defined in the literature: disgust, anger, enjoyment, fear, sadness (or distress), and surprise. In particular, the surveyed studies have detected deceptive behaviours by analysing the presence of both extreme facial movements, which can indicate high-intensity emotions that are not really being experienced, and facial rigidity (i.e., little or no facial movements) that deceivers could purposefully inhibit to appear truthful. Moreover, half of the studies (11) rely on head pose, followed by gaze (9 studies; 41%).

5.3. RQ3: Which Decision-Making Algorithms for Automated Deception Detection Are Used?

During the decision-making step, the extracted features are used to train a classification approach to automatically distinguish between deceptive and truthful videos.
Specifically, 20 studies among the surveyed ones apply a deception classification approach, as represented in Table 7. Most of them (19 studies; 95%) use supervised learning techniques that require labelled training data (i.e., a set of feature vectors with their corresponding labels (deceptive vs. truthful)). The most applied supervised technique is the support vector machine (SVM), which is applied in seven studies (35%), followed by six studies (30%) applying Random Forest, five studies applying neural networks (25%), and four studies (20%) applying K-nearest neighbourhood. Various types of regression models have been experimented by surveyed papers on facial deception detection, including multinomial logistic and multivariate regression. Only one study relies on unsupervised methods that do not need manually annotated facial attributes to train the model. Specifically, it applies hidden Markov models [42].
To evaluate the deception classification approaches shown in Table 7, different metrics have been proposed in the reviewed literature, as listed below:
-
Area under the curve (AUC): it measures the probability that the model ranks a random positive example more highly than a random negative example;
-
Accuracy (ACC): it measures the probability of correspondence between a positive decision and true condition (i.e., the proportion of correct classification decisions);
-
Precision: it is defined as the number of true positives divided by the number of true positives plus the number of true negatives;
-
Recall: it is defined as the number of true positives divided by the number of true positives plus the number of false negatives;
-
F1 score: It is a weighted average of precision and recall.
Table 8 provides a summary of the extracted metrics. Not all the surveyed studies are included in the table because only some of them evaluated their methods. Note that this table does not provide a comparative analysis because the studies rely on different databases to validate the approaches.

5.4. RQ4: Which Datasets for Automated Deception Detection Are Used?

An important step for a systematic evaluation of the performance of the deception-detection method is the preparation and use of well-defined evaluation datasets [58].
Table 9 summarises the datasets used by the selected studies and their main characteristics, i.e., the size, the source used to acquire the data, and the availability. The most popular dataset (seven studies; 33%) is the real-life trial dataset [59], which is available upon request. This dataset consists of 121 videos (average length of 28.0 s) collected from public court trials, including 61 deceptive and 60 truthful trial clips. The remaining studies (14 studies; 67%) developed their own dataset to evaluate the deception-detection method mainly by recording videos from game/TV shows (5 studies; 24%), followed by the use of spontaneous (4 studies; 19%) or controlled (4 studies; 19%) interviews, and by the use of storytelling (2 studies; 10%) or the download of videos from YouTube (2 studies; 10%). The larger dataset is Abd et al.’s database [18] containing 888 video clips from 102 participants. Considering the availability, the majority of the studies collected datasets without making them available (11 studies; 52%), while 7 studies (33%) made datasets available on request. Only three studies (14%) published the collected dataset.

6. Future Research Directions

The development of facial-deceptive-detection methods brings many opportunities for future research. Table 10 summarises the future research directions extracted from the surveyed studies and opportunely classifies them according to the aim of the future research. Note that only the studies that provided a future work discussion have been included in the table (14 studies; 58%).
The majority of the surveyed studies underline the need to extend the current dataset for the generalization of the results (five studies; 36%) and use of a multimodal approach (five studies; 36%) that combines different types of features (speech, video, text, hand gestures, head movements, etc.). In more detail, evaluating the accuracy of the detection algorithms over a larger database yields a generalization of results, as well as an increase in the performance of the algorithms, while the integration of several modalities yields higher accuracy.
Relevance is also given to the complexity reduction (four studies; 29%) by exploring innovative ways of combining and classifying the features for real-time deception detection.
Furthermore, a look at the development of further methods and the use of larger samples (three studies; 21% each) is also suggested by the surveyed studies. In more detail, the need to develop more accurate methods of action unit detection, more complex classifiers to model conversational context and time-dependent features, and affect-aware systems resulted from the analysis. On considering the use of larger samples, particular attention is given to an equal representation by gender and race and generalization to other scenarios.
Finally, it is also important to mention the future research directions looking for further psychological analysis, transferability, and increase in the extracted features (two studies; 14% each), as well as the explainability of the algorithms, model validation, and use of 3D deformable models (one study; 7% each).

7. Conclusions

This study provided a systematic literature analysis on facial deception detection in videos. The main contribution was a discussion of the methods applied in the literature of the last decade classified according to the main steps of the facial-deception-detection process (video pre-processing, facial feature extraction, and decision making). Moreover, datasets used for the evaluation and future research directions have also been analysed.
The initial search on WoS and Scopus databases returned 230 studies, of which 38 studies were retained after the application of the inclusion and exclusion criteria defined in the methodology. Finally, 24 studies were included in the qualitative synthesis according to the scores obtained in the quality evaluation checklist. The analysis revealed that in the last ten years, the interest of the scientific community in the topic of facial deception detection began growing in 2014 and reached a peak in 2020.
Considering the methods for video pre-processing, we observed that the majority of the surveyed studies use academic and commercial software for face detection and landmark tracking. Moreover, the most applied and successful feature-based methods are the active shape model [43] and the Viola–Jones method [33], while the Constrained Local Neural Field (CLNF) [46] is the only used image-based approach.
Concerning the extracted facial features, the Facial Action Coding System (FACS) is the most used system to represent all visually discernible facial movements, applied by 68% of the studies. Moreover, 82% of the studies rely on extracting facial expression features, followed by head pose (50% of the studies) and gaze motion (41%).
Concerning the decision-making algorithms, 95% of the studies use supervised learning techniques, mainly the support vector machine (SVM) (35% of the studies), followed by random forest (30%), neural networks (25%), and K-nearest neighbourhood (20%).
Furthermore, the real-life trial dataset [59] is the most popular dataset used for the evaluation of the performance of the decision-making algorithms (33% of the studies), while the remaining studies (67%) developed their own dataset mainly by recording videos from game/TV shows (24%), using spontaneous (19%) or controlled (19%) interviews, and using storytelling (10%).
Finally, the surveyed studies highlight the following future research directions: (i) extension of the current datasets for evaluating the detection algorithms; (ii) use of a multimodal approach that combines different types of features (facial, speech, video, text, hand gestures, etc.); (iii) reduction in the complexity of the detection process for real-time use; (iv) use of larger samples for an equal representation by gender and race; and (v) development of further more accurate and affect-aware methods.
The study has some limitations mainly related to the focus on facial cues from videos. Future studies could also include analysis of explicit cues (verbal communication) from face-to-face interactions. Indeed, multimodal analysis that combines both verbal and non-verbal cues is the most valuable approach for detecting deception [15].

Author Contributions

Conceptualization, A.D. (Alessia D’Andrea), A.D. (Arianna D’Ulizia), P.G. and F.F.; Methodology, A.D. (Alessia D’Andrea), A.D. (Arianna D’Ulizia), and P.G.; Validation, A.D. (Alessia D’Andrea) and A.D. (Arianna D’Ulizia); Formal analysis, A.D. (Alessia D’Andrea) and A.D. (Arianna D’Ulizia); Investigation, A.D. (Alessia D’Andrea), A.D. (Arianna D’Ulizia), and P.G.; Data curation, A.D. (Alessia D’Andrea), A.D. (Arianna D’Ulizia), and F.F.; Writing—original draft preparation, A.D. (Alessia D’Andrea) and A.D. (Arianna D’Ulizia); Writing—review and editing, A.D. (Alessia D’Andrea) and A.D. (Arianna D’Ulizia); Supervision, A.D. (Arianna D’Ulizia). All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are available upon request from the corresponding author.

Conflicts of Interest

The authors declare no conflict of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript; or in the decision to publish the results.

References

  1. Jakubowska, J.; Białecka-Pikul, M. A new model of the development of deception: Disentangling the role of false-belief understanding in deceptive ability. Soc. Dev. 2020, 29, 21–40. [Google Scholar] [CrossRef]
  2. Feng, Y.; Hung, S.; Hsieh, P. Detecting spontaneous deception in the brain. Hum. Brain Mapp. 2022, 43, 3257–3269. [Google Scholar] [CrossRef] [PubMed]
  3. Bond, C.F., Jr.; DePaulo, B.M. Accuracy of deception judgments. Personal. Soc. Psychol. Rev. 2006, 10, 214–234. [Google Scholar] [CrossRef] [PubMed]
  4. Hartwig, M.; Voss, J.A.; Brimbal, L.; Wallace, D.B. Investment Professionals’ Ability to Detect Deception: Accuracy, Bias and Metacognitive Realism. J. Behav. Financ. 2017, 18, 1–13. [Google Scholar] [CrossRef]
  5. Zuckerman, M.; DePaulo, B.M.; Rosenthal, R. Verbal and Nonverbal Communication of Deception. In Advances in Experimental Social Psychology; Academic Press Inc.: Cambridge, MA, USA, 1981. [Google Scholar]
  6. Viglione, D.J.; Giromini, L.; Landis, P. The Development of the Inventory of Problems: A Brief Self-Administered Measure for Discriminating Bona Fide From Feigned Psychiatric and Cognitive Complaints. J. Pers. Assess. 2016, 99, 534–544. [Google Scholar] [CrossRef] [Green Version]
  7. Vance, N.; Speth, J.; Khan, S.; Czajka, A.; Bowyer, K.W.; Wright, D.; Flynn, P. Deception Detection and Remote Physiological Monitoring: A Dataset and Baseline Experimental Results. IEEE Trans. Biom. Behav. Identity Sci. 2022, 4, 522–532. [Google Scholar] [CrossRef]
  8. Randhavane, T.; Bhattacharya, U.; Kapsaskis, K.; Gray, K.; Bera, A.; Manocha, D. The liar’s walk: Detecting deception with gait and gesture. arXiv 2019, arXiv:1912.06874. [Google Scholar]
  9. Malik, J.S.; Pang, G.; Hengel, A.V.D. Deep learning for hate speech detection: A comparative study. arXiv 2022, arXiv:2202.09517. [Google Scholar]
  10. Ogale, N.A. A Survey of Techniques for Human Detection from Video; University of Maryland: College Park, MD, USA, 2006; Volume 125, p. 19. [Google Scholar]
  11. Ekman, P. Darwin, deception, and facial expression. Ann. N. Y. Acad. Sci. 2003, 1000, 205–221. [Google Scholar] [PubMed] [Green Version]
  12. Matsumoto, D.; Hwang, H.C. Microexpressions Differentiate Truths from Lies About Future Malicious Intent. Front. Psychol. 2018, 9, 2545. [Google Scholar] [CrossRef]
  13. Darwin, C. The Expression of the Emotions in Man and Animals; J. Murray: London, UK, 1872. [Google Scholar]
  14. Ding, M.; Zhao, A.; Lu, Z.; Xiang, T.; Wen, J.R. Face-focused cross-stream network for deception detection in videos. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019. [Google Scholar]
  15. Constâncio, A.S.; Tsunoda, D.F.; Silva, H.d.F.N.; da Silveira, J.M.; Carvalho, D.R. Deception detection with machine learning: A systematic review and statistical analysis. PLoS ONE 2023, 18, e0281323. [Google Scholar] [CrossRef]
  16. Alaskar, H.; Sbaï, Z.; Khan, W.; Hussain, A.; Alrawais, A. Intelligent techniques for deception detection: A survey and critical study. Soft Comput. 2023, 27, 3581–3600. [Google Scholar] [CrossRef]
  17. Herchonvicz, A.L.; de Santiago, R. Deep Neural Network Architectures for Speech Deception Detection: A Brief Survey. In Progress in Artificial Intelligence, Proceedings of the 20th EPIA Conference on Artificial Intelligence, EPIA 2021, Virtual Event, 7–9 September 2021; Springer International Publishing: Cham, Switzerland, 2021; pp. 301–312. [Google Scholar]
  18. Thannoon, H.H.; Ali, W.H.; Hashim, I.A. Detection of Deception Using Facial Expressions Based on Different Classification Algorithms. In Proceedings of the Third Scientific Conference of Electrical Engineering (SCEE), IEEE, Baghdad, Iraq, 19–20 December 2019; pp. 51–56. [Google Scholar]
  19. Liberati, M.; Tetzlaff, J.; Altman, D.G.; PRISMA Group. Preferred reporting items for systematic reviews and meta-analyses: The PRISMA statement. PLoS Med. 2009, 6, e1000097. [Google Scholar] [CrossRef] [Green Version]
  20. Page, M.J.; McKenzie, J.E.; Bossuyt, P.M.; Boutron, I.; Hoffmann, T.C.; Mulrow, C.D. The PRISMA 2020 statement: An updated guideline for reporting systematic reviews. BMJ 2021, 372, n71. [Google Scholar] [CrossRef] [PubMed]
  21. Pranckutė, R. Web of Science (WoS) and Scopus: The Titans of Bibliographic Information in Today’s Academic World. Publications 2021, 9, 12. [Google Scholar] [CrossRef]
  22. Paez, A. Grey literature: An important resource in systematic reviews. J. Evid. Based Med. 2017, 10, 233–240. [Google Scholar] [CrossRef] [PubMed]
  23. Chebbi, S.; Jebara, S.B. An Audio-Visual based Feature Level Fusion Approach Applied to Deception Detection. In Proceedings of the 15th International Conference on Computer Vision Theory and Applications (4: VISAPP), Valetta, Malta, 27–29 February 2020; pp. 197–205. [Google Scholar]
  24. Abbas, Z.K.; Al-Ani, A.A. An adaptive algorithm based on principal component analysis-deep learning for anomalous events detection. Indones. J. Electr. Eng. Comput. Sci. 2023, 29, 421–430. [Google Scholar] [CrossRef]
  25. Shen, X.; Fan, G.; Niu, C.; Chen, Z. Catching a Liar Through Facial Expression of Fear. Front. Psychol. 2021, 12, 675097. [Google Scholar] [CrossRef]
  26. Gadea, M.; Aliño, M.; Espert, R.; Salvador, A. Deceit and facial expression in children: The enabling role of the “poker face” child and the dependent personality of the detector. Front. Psychol. 2015, 6, 1089. [Google Scholar] [CrossRef] [Green Version]
  27. Monaro, M.; Maldera, S.; Scarpazza, C.; Sartori, G.; Navarin, N. Detecting deception through facial expressions in a dataset of videotaped interviews: A comparison between human judges and machine learning models. Comput. Hum. Behav. 2022, 127, 107063. [Google Scholar] [CrossRef]
  28. Su, L.; Levine, M. Does “lie to me” lie to you? An evaluation of facial clues to high-stakes deception. Comput. Vis. Image Underst. 2016, 147, 52–68. [Google Scholar]
  29. Abd, S.H.; Hashim, I.A.; Jalal, A.S.A. Hardware implementation of deception detection system classifier. Period. Eng. Nat. Sci. PEN 2021, 10, 151–163. [Google Scholar] [CrossRef]
  30. Su, L.; Levine, M.D. High-stakes deception detection based on facial expressions. In Proceedings of the 2014 22nd International Conference on Pattern Recognition, IEEE, Stockholm, Sweden, 24–28 August 2014; pp. 2519–2524. [Google Scholar]
  31. Ngo, L.M.; Wang, W.; Mandira, B.; Karaoglu, S.; Bouma, H.; Dibeklioglu, H.; Gevers, T. Identity Unbiased Deception Detection by 2D-to-3D Face Reconstruction. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Virtual, 5–9 January 2021; pp. 145–154. [Google Scholar]
  32. Mathur, L.; Matarić, M.J. Introducing Representations of Facial Affect in Automated Multimodal Deception Detection. In Proceedings of the 2020 International Conference on Multimodal Interaction, Virtual, 25–29 October 2020. [Google Scholar] [CrossRef]
  33. Yu, X.; Zhang, S.; Yan, Z.; Yang, F.; Huang, J.; Dunbar, N.E.; Jensen, M.L.; Burgoon, J.K.; Metaxas, D.N. Is Interactional Dissynchrony a Clue to Deception? Insights from Automated Analysis of Nonverbal Visual Cues. IEEE Trans. Cybern. 2014, 45, 492–506. [Google Scholar] [CrossRef]
  34. Zhang, J.; Levitan, S.I.; Hirschberg, J. Multimodal Deception Detection Using Automatically Extracted Acoustic, Visual, and Lexical Features. In Proceedings of the INterspeech, Shanghai, China, 25–29 October 2020; pp. 359–363. [Google Scholar] [CrossRef]
  35. Belavadi, V.; Zhou, Y.; Bakdash, J.Z.; Kantarcioglu, M.; Krawczyk, D.C.; Nguyen, L.; Rakic, J.; Thuriasingham, B. MultiModal Deception Detection: Accuracy, Applicability and Generalizability. In Proceedings of the 2020 Second IEEE International Conference on Trust, Privacy and Security in Intelligent Systems and Applications (TPS-ISA), IEEE, Atlanta, GA, USA, 28–31 October 2020; pp. 99–106. [Google Scholar]
  36. Feinland, J.; Barkovitch, J.; Lee, D.; Kaforey, A.; Ciftci, U.A. Poker Bluff Detection Dataset Based on Facial Analysis. In Image Analysis and Processing–ICIAP 2022: 21st International Conference, Lecce, Italy, May 23–27, 2022, Proceedings, Part III; Springer International Publishing: Cham, Switzerland, 2022; pp. 400–410. [Google Scholar]
  37. Sen, T.K. Temporal patterns of facial expression in deceptive and honest communication. In Proceedings of the 2017 Seventh International Conference on Affective Computing and Intelligent Interaction (ACII), IEEE, San Antonio, TX, USA, 23–26 October 2017; pp. 616–620. [Google Scholar]
  38. Yildirim, S.; Chimeumanu, M.S.; Rana, Z.A. The influence of micro-expressions on deception detection. Multimed. Tools Appl. 2023, 82, 29115–29133. [Google Scholar] [CrossRef]
  39. Mathur, L.; Mataric, M.J. Unsupervised audio-visual subspace alignment for high-stakes deception detection. In Proceedings of the ICASSP 2021–2021 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), IEEE, Toronto, ON, Canada, 6–11 June 2021; pp. 2255–2259. [Google Scholar]
  40. Pérez-Rosas, V.; Abouelenien, M.; Mihalcea, R.; Xiao, Y.; Linton, C.J.; Burzo, M. Verbal and nonverbal clues for real-life deception detection. In Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing, Lisbon, Portugal, 17–21 September 2015; pp. 2336–2346. [Google Scholar]
  41. Venkatesh, S.; Ramachandra, R.; Bours, P. Video based deception detection using deep recurrent convolutional neural network. In Computer Vision and Image Processing, Proceedings of the 4th International Conference, CVIP 2019, Jaipur, India, 27–29 September 2019; Revised Selected Papers, Part II 4; Springer: Singapore, 2020; pp. 163–169. [Google Scholar]
  42. Demyanov, S.; Bailey, J.; Ramamohanarao, K.; Leckie, C. Detection of deception in the mafia party game. In Proceedings of the 2015 ACM on International Conference on Multimodal Interaction, ACM, Seattle, WA, USA, 9–13 November 2015; pp. 335–342. [Google Scholar]
  43. Pentland, S.J.; Twyman, N.W.; Burgoon, J.K.; Nunamaker, J.F., Jr.; Diller, C.B. A video-based screening system for automated risk assessment using nuanced facial features. J. Manag. Inf. Syst. 2017, 34, 970–993. [Google Scholar] [CrossRef]
  44. Carissimi, N.; Beyan, C.; Murino, V. A multi-view learning approach to deception detection. In Proceedings of the 2018 13th IEEE International Conference on Automatic Face & Gesture Recognition, IEEE, Xi’an, China, 15–19 May 2018; pp. 599–606. [Google Scholar]
  45. Kazemi, V.; Sullivan, J. One millisecond face alignment with an ensemble of regression trees. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 1867–1874. [Google Scholar]
  46. Lu, S.; Tsechpenakis, G.; Metaxas, D.; Jensen, M.L.; Kruse, J. Blob analysis of the head and hands: A method for deception detection and emotional state identification. In Proceedings of the 38th Annual Hawaii International Conference on System Sciences, Big Island, HI, USA, 6 January 2005. [Google Scholar]
  47. Michael, N.; Dilsizian, M.; Metaxas, D.; Burgoon, J.K. Motion profiles for deception detection using visual cues. In European Conference on Computer Vision; Springer: Berlin/Heidelberg, Germany, 2010; pp. 462–475. [Google Scholar]
  48. Ekman, P.; Friesen, W.V.; Hager, J.C. Facial Action Coding System: The Manual on CD-ROM. Instructor’s Guide; Network Information Research Co.: Salt Lake City, UT, USA, 2002. [Google Scholar]
  49. Kanaujia, A.; Huang, Y.; Metaxas, D. Tracking facial features using mixture of point distribution models. In Computer Vision, Graphics and Image Processing; Springer: Berlin/Heidelberg, Germany, 2006; pp. 492–503. [Google Scholar]
  50. Wu, Y.; Ji, Q. Facial Landmark Detection: A Literature Survey. Int. J. Comput. Vis. 2019, 127, 115–142. [Google Scholar] [CrossRef] [Green Version]
  51. Viola, P.; Jones, M. Robust real-time object detection. Int. J. Comput. Vis. 2001, 4, 4. [Google Scholar]
  52. Nechyba, M.C.; Brandy, L.; Schneiderman, H. PittPatt Face Detection and Tracking for the CLEAR 2007 Evaluation. In Multimodal Technologies for Perception of Humans. RT CLEAR 2007; Lecture Notes in Computer Science; Stiefelhagen, R., Bowers, R., Fiscus, J., Eds.; Springer: Berlin/Heidelberg, Germany, 2007; Volume 4625. [Google Scholar] [CrossRef]
  53. Guo, J.; Deng, J.; An, X.; Yu, J. Insightface. Available online: https://github.com/deepinsight/insightface (accessed on 21 August 2020).
  54. Bulat, A.; Tzimiropoulos, G. How far are we from solving the 2D & 3D face alignment problem? (and a dataset of 230,000 3d facial landmarks). In Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017. [Google Scholar]
  55. Amos, B.; Ludwiczuk, B.; Satyanarayanan, M. Openface: A general-purpose face recognition library with mobile applications. CMU Sch. Comput. Sci. 2016, 6, 20. [Google Scholar]
  56. De la Torre, F.; Chu, W.S.; Xiong, X.; Vicente, F.; Ding, X.; Cohn, J. IntraFace. In Proceedings of the IEEE International Conference on Automatic Face & Gesture Recognition and Workshops, Ljubljana, Slovenia, 4–8 May 2015; Volume 1. [Google Scholar]
  57. Ekman, P.; Friesen, W.V.; Hager, J.C. The Facial Action Coding System: A Technique for the Measurement of Facial Movement; Consulting Psychologists Press: San Francisco, CA, USA, 2002. [Google Scholar]
  58. D’Ulizia, A.; Caschera, M.C.; Ferri, F.; Grifoni, P. Fake news detection: A survey of evaluation datasets. PeerJ Comput. Sci. 2021, 7, e518. [Google Scholar] [CrossRef]
  59. Perez-Rosas, V.; Abouelenien, M.; Mihalcea, R.; Burzo, M. Deception detection using real-life trial data. In Proceedings of the 2015 ACM on International Conference on Multimodal Interaction, ACM, Seattle, WA, USA, 9–13 November 2015; pp. 59–66. [Google Scholar]
Figure 2. The PRISMA flow diagram, adapted from [20]. For more information, visit: http://www.prisma-statement.org/ (accessed on 9 August 2023).
Figure 2. The PRISMA flow diagram, adapted from [20]. For more information, visit: http://www.prisma-statement.org/ (accessed on 9 August 2023).
Applsci 13 09188 g002
Figure 3. Temporal distribution of the surveyed studies.
Figure 3. Temporal distribution of the surveyed studies.
Applsci 13 09188 g003
Table 2. Exclusion and inclusion criteria formulated for the study.
Table 2. Exclusion and inclusion criteria formulated for the study.
PhaseExclusion CriteriaInclusion Criteria
Screening phasee1. Understandability criterion:
-
articles that are written not in English.
i1. Relevance criterion on the title:
-
studies relevant to the review focus, i.e., describing automated deception detection from videos focusing on facial expressions;
-
studies answering the defined review questions: (i) the methods for video pre-processing, (ii) the facial extracted features, (iii) the decision-making algorithms, or (iv) the available datasets.
e2. Duplication criterion:
-
same articles retrieved from two different databases;
-
articles having the same title and authors retrieved from the same database but published in different sources.
i2. Temporal criterion:
-
studies published in the period 2013–2023.
i3. Document type criterion:
-
studies belonging to the following document types: article, review, book chapter, conference/proceedings paper.
Eligibility phasee3. Availability criterion:
-
studies that are not accessible in full text.
i4. Relevance criterion on the abstract:
-
studies relevant to the review focus, i.e., describing automated deception detection from videos focusing on facial expressions;
-
studies answering the review questions: (i) the methods for video pre-processing, (ii) the facial extracted features, (iii) the decision-making algorithms, or (iv) the available datasets.
Table 3. Quality evaluation checklist and scores.
Table 3. Quality evaluation checklist and scores.
Quality Assessment QuestionsScores
QA1 Does the study describe the methods for video pre-processing?1—yes, the methods for video pre-processing are described in detail.
0.5—partially, the methods for video pre-processing are summarised without providing details on the different steps.
0—no, the methods for video pre-processing are only cited, without providing a description.
QA2Does the study describe the facial extracted features?1—yes, the facial extracted features are described in detail.
0.5—partially, the facial extracted features are summarised without providing a detailed description.
0—no, the facial extracted features are not described.
QA3Does the study describe the decision-making algorithm?1—yes, the decision-making algorithm is fully described.
0.5—partially, the decision-making algorithm is summarised without providing a detailed description.
0—no, the decision-making algorithm is not described.
QA4 Does the study use some datasets for the evaluation of the method? 1—yes, one or more datasets are used.
0—no, datasets are not used.
Table 4. Studies selected for the analysis.
Table 4. Studies selected for the analysis.
ReferenceAuthorsYearDocument Type
[23]Chebbi and Jebara 2020Conference paper
[24]Abbas and Al-Ani2023Article
[25]Shen et al.2021Article
[26]Gadea et al.2015Article
[27]Monaro et al.2022Article
[28]Su and Levine2016Article
[14]Ding et al.2019Conference Paper
[29]Abd et al.2021Article
[30]Su and Levine2014Conference Paper
[31]Ngo et al.2021Conference Paper
[32]Mathur and Matarić2020Conference paper
[33]Yu et al.2014Article
[34]Zhang et al.2020Conference Paper
[35]Belavadi et al.2020Conference paper
[36]Feinland et al.2022Conference Paper
[37]Sen2017Conference paper
[38]Yildirim et al.2023Article
[39]Mathur and Matarić2021Conference Paper
[40]Pérez-Rosas et al.2015Conference paper
[41]Venkates et al.2020Conference paper
[18]Thannoon et al.2019Conference paper
[42]Demyanov et al.2015Conference paper
[43]Pentland et al.2017Article
[44]Carissimi et al.2018Conference paper
Table 5. Algorithms and software tools applied for face detection and landmark tracking by surveyed papers.
Table 5. Algorithms and software tools applied for face detection and landmark tracking by surveyed papers.
Reference2D3DWhich Face-Detection Technique?Which Face Tracking Technique?
AlgorithmSoftware ToolAlgorithmSoftware Tool
Feature-BasedImage-BasedFeature-BasedImage-Based
Abbas and Al-Ani, 2023 [24]XXViola–Jones face detector [33] Constrained Local Neural Field (CLNF) [46]
Shen et al., 2021 [25]
Monaro et al., 2022 [27]
Mathur and Matarić, 2020 [32]
Belavadi et al., 2020 [35]
Sen, 2017 [37]
Mathur and Matarić, 2021 [39]
Carissimi et al., 2018 [44]
X OpenFace [47] OpenFace [47]
Su and Levine, 2016 [28]
Su and Levine, 2014 [30]
X PittPatt SDK [52] PittPatt SDK [52]
Abd et al., 2021 [29]
Thannoon et al., 2019 [18]
X Viola–Jones face detector [33] Constrained Local Neural Field (CLNF) [46]
Ngo et al., 2021 [31]X ‘dlib’ library [50] FAN [54]
Yu et al., 2014 [33] XActive Shape Model (ASM) tracker [43] Active Shape Model (ASM) tracker [43]
Zhang et al., 2020 [34]X ‘dlib’ library [50] ‘dlib’ library [50]
Pérez-Rosas et al., 2015 [40]X InsightFace [53] InsightFace [53]
Demyanov et al., 2015 [42]X OpenCV lib [51] FaceSDK [49]
Pentland et al., 2017 [43]X IntraFace [48] IntraFace [48]
Table 6. Facial features extracted by surveyed studies.
Table 6. Facial features extracted by surveyed studies.
ReferenceFacial FeaturesFacial Action Coding SystemWhich Action Units?
IIIIIIIVVVIVIIVIIIIXX
Chebbi and Jebara, 2020 [23]X X X XXX
Abbas and Al-Ani, 2023 [24]X X X XAU 6, 7, 10, 12, 14, 28
Shen et al., 2021 [25]X XAU 1, 2, 4, 5, 20, 26
Gadea et al., 2015 [26]X
Monaro et al., 2022 [27]X XAU 1, 2, 4, 5, 6, 7, 9, 10, 12, 14, 15, 17, 20, 23, 25, 26, 28, 45
Su and Levine, 2016 [28]X XXXX XAU1, 2, 4, 12, 15, 45
Ding et al., 2019 [14]X X
Abd et al., 2021 [29]X X X X
Su and Levine, 2014 [30]X XXXX XAU1, 2, 4, 12, 15, 45
Ngo et al., 2021 [31]X X
Mathur and Matarić, 2020 [32]X X X X
Yu et al., 2014 [33] X X XAU 1, 2, 5, 15, 17, 26, 45
Zhang et al., 2020 [34]X X
Belavadi et al., 2020 [35]X XXX XXX
Feinland et al., 2022 [36] X X XX
Yildirim et al., 2023 [38]X XAU 6, 12
Pérez-Rosas et al., 2015 [40] X X X
Venkates et al., 2020 [41]X XXX XXX
Thannoon et al., 2019 [18] XAU 5, 6, 7, 10, 12, 14, 23, and 28
Demyanov et al., 2015 [42] XAU 6 and 12
Pentland et al., 2017 [43] XAU 1, 2, 4, 5, 6, 7, 9, 12, 15, 16, 17, 20, 23, and 24
Carissimi et al., 2018 [44]XX
Table 7. Deception classification approaches applied by surveyed papers on facial deception detection.
Table 7. Deception classification approaches applied by surveyed papers on facial deception detection.
ReferenceSupervised (S) vs. Unsupervised (U)Regression ModelsNeural NetworksK-Nearest NeighborhoodSupport Vector MachineRandom ForestHidden Markov ModelLong Short-Term MemoryDecision Tree
Multinomial Logistic RegressionMultivariate RegressionMultilayer Perceptron Neural NetworkVirtual Generalizing Random Access Memory Weightless Neural NetworkCNN3D CNN ModelR-CNN
Chebbi and Jebara, 2020 [23]S X
Abbas and Al-Ani, 2023 [24]SX
Shen et al., 2021 [25]S X X
Monaro et al., 2022 [27]S X X X
Su and Levine, 2016 [28]S X
Ding et al., 2019 [14]S X
Abd et al., 2021 [29]S X
Su and Levine, 2014 [30]S X
Ngo et al., 2021 [31]S X
Mathur and Matarić, 2020 [32]S X
Yu et al., 2014 [33]S X
Zhang et al., 2020 [34]S X
Belavadi et al., 2020 [35]S X
Feinland et al., 2022 [36]S X
Pérez-Rosas et al., 2015 [40]S X
Venkates et al., 2020 [41]S XX X
Thannoon et al., 2019 [18]S XX XX
Demyanov et al., 2015 [42]U X
Pentland et al., 2017 [43]SX
Carissimi et al., 2018 [44]S X
Table 8. Performance measures obtained by deception classification approaches applied by surveyed papers.
Table 8. Performance measures obtained by deception classification approaches applied by surveyed papers.
ReferenceAUC (%)ACC (%)Precision (%)Recall (%)F1 Score (%)
Chebbi and Jebara, 2020 [23]-948810094
Abbas and Al-Ani, 2023 [24]94.2188.46---
Shen et al., 2021 [25]-----
Monaro et al., 2022 [27]78----
Su and Levine, 2016 [28]-76.92-72.41-
Ding et al., 2019 [14]96.7193.16---
Abd et al., 2021 [29]-89.64---
Su and Levine, 2014 [30]-76.92---
Ngo et al., 2021 [31]-686672-
Mathur and Matarić, 2020 [32]9184--84
Yu et al., 2014 [33]-66.866.865.1-
Zhang et al., 2020 [34]7363.6---
Belavadi et al., 2020 [35]-48.15---
Yildirim et al., 2023 [38]-74.17---
Mathur and Matarić, 2021 [39]7574--73
Pérez-Rosas et al., 2015 [40]-82.14---
Venkates et al., 2020 [41]-100---
Thannoon et al., 2019 [18]-90---
Demyanov et al., 2015 [42]61----
Carissimi et al., 2018 [44]-99---
Table 9. Datasets used by the selected studies.
Table 9. Datasets used by the selected studies.
ReferenceDatasetAvailabilitySizeSource
Chebbi and Jebara, 2020 [23]Real-life trialAvailable upon request121 videos (61 deceptive and 60 truthful)“The Innocence Project” website
Abbas and Al-Ani, 2023 [24]Abd et al.’s databaseNot available888 video clips from 102 participantsSpontaneous interview
Shen et al., 2021 [25]Shen et al.’s databaseAvailable on request32 video clips of 16 individualsThe game show “the Moment of Truth.”
Gadea et al., 2015 [26]Deceit-detection testNot available12 video-recordingsStorytelling of life moments
Monaro et al., 2022 [27]Monaro et al.’s datasetNot available62 videos of Italian volunteer participantsSpontaneous interview
Su and Levine, 2016 [28]Su and Levin’s datasetNot available324 video clipsYouTube
Ding et al., 2019 [14]Real-life trialAvailable upon request121 videos (61 deceptive and 60 truthful).“The Innocence Project” website
Abd et al., 2021 [29]Abd et al.’s databaseNot available888 video clips from 102 participantsSpontaneous interview
Su and Levine, 2014 [30]Su and Levin’s datasetNot available324 video clipsYouTube
Ngo et al., 2021 [31]Low-stakes deceit (LSD) datasetNot available624 recordings of 312 participants Interviewees under a controlled environment
Real-life trialAvailable upon request121 videos (61 deceptive and 60 truthful).“The Innocence Project” website
Mathur and Matarić, 2020 [32]Real-life trialAvailable upon request121 videos (61 deceptive and 60 truthful).“The Innocence Project” website
Yu et al., 2014 [33]Yu et al.’s datasetNot available242 videotaped interviews of 121 interviewer–interviewee pairsSpontaneous interview
Zhang et al., 2020 [34]Box of Lies corpusPublicly available68 recorded clips (39 truths and 29 lies) “Box of Lies” game
Belavadi et al., 2020 [35]Real-life trialAvailable upon request121 videos (61 deceptive and 60 truthful).“The Innocence Project” website
Opinion datasetNot availableNot availableStorytelling of movies
Crime datasetNot availableNot availableInterviewees under a controlled environment
Feinland et al., 2022 [36] Poker Bluff Detection DatasetNot available65 recordings of bluff events (15 unique faces with 2320 face images)Poker game
Pérez-Rosas et al., 2015 [40]Perez-Rosas et al.’s datasetPublicly available118 deceptive and truthful video clips TV shows
Venkates et al., 2020 [41]Real-life Trial Available upon request121 videos (61 deceptive and 60 truthful).“The Innocence Project” website
Thannoon et al., 2019 [18] Not available400 video clips (50% lie response, 50% truth response)Interviewees under a controlled environment
Demyanov et al., 2015 [42]Mafia game database Publicly available(60 deceptive and 210 truthful players)Mafia TV show
Pentland et al., 2017 [43]Pentland et al.’s datasetNot available95 videos of interviewsInterviewees under a controlled environment
Carissimi et al., 2018 [44]Real-life Trial Available upon request121 videos (61 deceptive and 60 truthful)“The Innocence Project” website
Table 10. Future research directions envisaged by the selected studies.
Table 10. Future research directions envisaged by the selected studies.
ReferenceFuture Research DirectionsAim of the Future Research
Chebbi and Jebara, 2020 [23]
-
Reducing the complexity for real-time use through other approaches for combining speech and video modalities.
-
Complexity reduction
Shen et al., 2021 [25]
-
Extending the number of participants in the current dataset for the generalization of the results.
-
Dataset extension
Gadea et al., 2015 [26]
-
Performing experiments with larger gender-matched samples to check if women are better lie detectors than men; increasing the number of videos in the dataset.
-
Larger gender/race-matched samples
-
Dataset extension
Monaro et al., 2022 [27]
-
Analysis of the personality and morality of the subjects to understand how they react in a high-stakes setting.
-
Improving the explainability of the machine-learning models to obtain reliable hints on which specific features are useful for detecting deception.
-
Using a multimodal approach to include the audio features and the text transcripts in the machine-learning analysis.
-
Exploring transfer learning techniques for exploiting other lie-detection datasets.
-
Further psychological analysis
-
Explainability of the algorithm
-
Multimodal approach
-
Transferability
Su and Levine, 2016 [28]
-
Integrating verbal, non-verbal, and facial cues for higher accuracy.
-
Extending the size of the database
-
Multimodal approach
-
Dataset extension
Mathur and Matarić, 2020 [32]
-
Developing affect-aware systems for automatically detecting deception and other social behaviours, particularly those occurring in unconstrained situations in the wild.
-
Development of further methods
Yu et al., 2014 [33]
-
Generalization to the greater sample population and to other scenarios.
-
Incorporating 3D deformable models.
-
Larger samples
-
Transferability
-
3D deformable models
Zhang et al., 2020 [34]
-
Evaluating the models on real-world multimodal deception data.
-
Exploring more complex classifiers to model conversational context and time-dependent features.
-
Model validation
-
Development of further methods
Feinland et al., 2022 [36]
-
Extending the dataset with equal representation by gender and race and with a large number of videos.
-
Dataset extension
-
Larger gender/race-matched samples
Sen, 2017 [37]
-
Utilizing an individual’s expression during specific baseline mental states.
-
Further psychological analysis
Yildirim et al., 2023 [38]
-
Developing a fully autonomous deception detection system by combining other modalities, such as vocal and text.
-
Increasing video features using hand gestures, head movements, and body posture.
-
Reducing the complexity for real-time deception detection.
-
Multimodal approach
-
Increasing the features
-
Complexity reduction
Pérez-Rosas et al., 2015 [40]
-
Using automatic facial and gesture identification and automated speech transcription for real-time deception detection.
-
Multimodal approach
Demyanov et al., 2015 [42]
-
Developing more accurate methods of action unit detection.
-
Extraction of more meaningful features.
-
Using a multimodal approach combining all types of features.
-
Using less expensive features or more powerful hardware for real-time use.
-
Increasing the number of videos on the database.
-
Development of further methods
-
Increasing the features
-
Multimodal approach
-
Complexity reduction
-
Dataset extension
Carissimi et al., 2018 [44]
-
Exploring ways to automatically learn the best views combination for the classification task for reducing the complexity.
-
Complexity reduction
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

D’Ulizia, A.; D’Andrea, A.; Grifoni, P.; Ferri, F. Detecting Deceptive Behaviours through Facial Cues from Videos: A Systematic Review. Appl. Sci. 2023, 13, 9188. https://doi.org/10.3390/app13169188

AMA Style

D’Ulizia A, D’Andrea A, Grifoni P, Ferri F. Detecting Deceptive Behaviours through Facial Cues from Videos: A Systematic Review. Applied Sciences. 2023; 13(16):9188. https://doi.org/10.3390/app13169188

Chicago/Turabian Style

D’Ulizia, Arianna, Alessia D’Andrea, Patrizia Grifoni, and Fernando Ferri. 2023. "Detecting Deceptive Behaviours through Facial Cues from Videos: A Systematic Review" Applied Sciences 13, no. 16: 9188. https://doi.org/10.3390/app13169188

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop