Next Article in Journal
Experimental Investigation of Inductive Sensor Characteristic for Blade Tip Clearance Measurement at High Temperature
Next Article in Special Issue
An Empirical Study Comparing Unobtrusive Physiological Sensors for Stress Detection in Computer Work
Previous Article in Journal
Reliable Indoor Pseudolite Positioning Based on a Robust Estimation and Partial Ambiguity Resolution Method
Previous Article in Special Issue
Game-Calibrated and User-Tailored Remote Detection of Stress and Boredom in Games
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Predicting Depression, Anxiety, and Stress Levels from Videos Using the Facial Action Coding System

by
Mihai Gavrilescu
* and
Nicolae Vizireanu
Department of Telecommunications, Faculty of Electronics, Telecommunications and Information Technology, University “Politehnica”, Bucharest 061071, Romania
*
Author to whom correspondence should be addressed.
Sensors 2019, 19(17), 3693; https://doi.org/10.3390/s19173693
Submission received: 28 June 2019 / Revised: 11 August 2019 / Accepted: 20 August 2019 / Published: 25 August 2019
(This article belongs to the Special Issue Sensors for Affective Computing and Sentiment Analysis)

Abstract

:
We present the first study in the literature that has aimed to determine Depression Anxiety Stress Scale (DASS) levels by analyzing facial expressions using Facial Action Coding System (FACS) by means of a unique noninvasive architecture on three layers designed to offer high accuracy and fast convergence: in the first layer, Active Appearance Models (AAM) and a set of multiclass Support Vector Machines (SVM) are used for Action Unit (AU) classification; in the second layer, a matrix is built containing the AUs’ intensity levels; and in the third layer, an optimal feedforward neural network (FFNN) analyzes the matrix from the second layer in a pattern recognition task, predicting the DASS levels. We obtained 87.2% accuracy for depression, 77.9% for anxiety, and 90.2% for stress. The average prediction time was 64 s, and the architecture could be used in real time, allowing health practitioners to evaluate the evolution of DASS levels over time. The architecture could discriminate with 93% accuracy between healthy subjects and those affected by Major Depressive Disorder (MDD) or Post-traumatic Stress Disorder (PTSD), and 85% for Generalized Anxiety Disorder (GAD). For the first time in the literature, we determined a set of correlations between DASS, induced emotions, and FACS, which led to an increase in accuracy of 5%. When tested on AVEC 2014 and ANUStressDB, the method offered 5% higher accuracy, sensitivity, and specificity compared to other state-of-the-art methods.

Graphical Abstract

1. Introduction

Stress, depression, and anxiety are emotional states that affect our mental and physical health. Research has shown that these three emotional states are strongly connected: chronic stress can activate defensive mechanisms in the brain that lead to anxiety; constant anxious feelings can cause depression; and, subsequently, depressive states intensify mental and physical stress [1]. This connection indicates that evaluating the three emotional states together would be more suitable than analyzing them separately. Knowing the severity of stress, depression, or anxiety associated with a subject as well as monitoring patterns over time can help in acting proactively to stop chronic mental problems (such as major depressive disorder (MDD), generalized anxiety disorder (GAD), or post-traumatic stress disorder (PTSD)) from developing and assist affected individuals in their day-to-day activities.
In this research, we propose a nonintrusive architecture for determining depression, anxiety, and stress levels by analyzing the facial dynamics of individuals using the Facial Action Coding System (FACS). We, therefore, aimed to build a low-cost nonintrusive system that provides high accuracy and that can be used by health practitioners for assessing and monitoring the severity of the three emotional states, and it also has the advantage of functioning in real time (and allowing experts to analyze severity patterns over time) as well as being less subjective than questionnaire-based methods (which are known to be prone to bias, such that different psychologists could provide different results [2]). We also describe the first research in the literature that studied the relationship between FACS and the Depression Anxiety and Stress Scale (DASS), a commonly used tool for evaluating the three emotional states. We show the results obtained from testing the proposed architecture using intrasubject and intersubject methodologies, and we compared our system in terms of accuracy, sensitivity, and specificity to other state-of-the-art methods used for similar tasks. Apart from being the first study in the literature that has analyzed the relationship between FACS and DASS, this study also proposes a novel architecture that is based on three layers, with the aim of providing DASS levels with high accuracy and with a low prediction time: in the first layer, a method based on Active Appearance Models (AAM) and set of multiclass Support Vector Machine (SVM) classifiers are used for fast Action Unit (AU) classification; in the second layer, a matrix is built containing the intensity levels of the classified AUs; and in the third layer, an optimal feedforward neural network (FFNN) analyzes the matrix from the second layer as a pattern recognition task, predicting the stress, anxiety, and depression levels. The proposed architecture can not only act as a tool for health practitioners, allowing them to evaluate the severity of the three emotional states, but can also be used for a wide range of other applications, including virtual psychology [3], computer-driven therapy [3], personalized health assistance [3], career counseling [3], and even diagnoses of physical diseases that have different stress, anxiety, or depression patterns as symptoms [4,5,6].
There are multiple ways of assessing levels of stress, depression, and anxiety, which can be divided into intrusive or contact-based methods and nonintrusive noncontact approaches. The intrusive or contact-based approaches refer to using specific sensors for measuring different parameters such as heart rate variability (HRV) [7], electrodermal activity [8], or an electroencephalogram (EEG) [9,10,11]. They have the advantage of being more precise, but are considerably more costly and need more effort from health practitioners and patients. On the other hand, nonintrusive noncontact approaches such as facial or speech analysis or questionnaire-based methods are easier to use and are less expensive, but are usually less accurate. In the next paragraph, we focus on detailing the current nonintrusive noncontact methods based on facial analysis used for predicting the three emotional states.
For evaluating stress, a significant study was conducted by N. Sharma et al. [12] in which a video database for a temporal thermal spectrum (TS) and visible spectrum (VS) called ANUStressDB was proposed. Researchers used a model based on local binary patterns on three orthogonal planes (LBP-TOPs) for stress detection, which was designed to evaluate histogram dynamic thermal patterns (HDTPs) and exploit the spatiotemporal features from videos by using an SVM-based classifier that classifies the subject as stressed or not. The proposed system offers an accuracy of 86%. This research was extended by R. Iran et al. [13] by extracting facial features from superpixels and providing them as input to different types of classifiers that discriminate between stressed and nonstressed subjects. The method was tested on the same ANUStressDB [12] and offers 3% higher accuracy compared to the method based on LBP-TOPs [12]. Another method was presented in Reference [14], with researchers assuming that mouth activity contains enough relevant information to evaluate stress. They proposed a semiautomated algorithm based on Eigen features and used template matching to classify mouth activity related to mouth openings and mouth deformations. Over 89% accuracy is obtained for discriminating between stressed and nonstressed subjects. Prasetio et al. [15] proposed a method for stress classification (neutral, low stress, high stress) by analyzing images with frontal faces divided into three parts: eyes, nose, and mouth. The facial features are extracted using difference of Gaussians (DoG), histogram of oriented gradients (HOGs), and discrete wavelet transform (DWT) methods. The histogram features determined are processed by a convolutional neural network (CNN) used to model the facial stress expressions. Tested on a facial recognition technology (FERET) [16] database using five-fold cross-validation, the system offers over 82% accuracy for stress classification.
For evaluating depression, local curvelet patterns in three orthogonal planes (LCBP-TOPs) have been successfully used for analyzing facial features and determining whether a subject is affected by depression or not, with over 73% accuracy [17]. Similarly, LBP-TOPs have been used [18] to analyze a set of visual cues in temporal segments using Fisher vectors (FVs). The method was tested on the Audio-Visual Emotion Challenge (AVEC2014) German database [19] and provided 75% accuracy in distinguishing between depressed and nondepressed subjects. LBP-TOPs have also been used for computing a temporal visual words dictionary [20] that was fetched to an SVM-based classifier to determine whether the subject was affected by depression or not, reaching an accuracy of 77%. The multiscale entropy (MSE) method has also been used [21] for the quantization of the temporal variability of facial expressions extracted from unsupervised features in video recordings collected from patients affected by MDD. Researchers have used a dynamic latent variable model (DLVM) for learning, and the results showed that unsupervised learned features could be used to distinguish between different levels of depression, with an accuracy of over 70%. Yang et al. [22] proposed an audio-visual multimodal depression recognition framework based on deep convolutional neural networks (DCNNs) and deep neural networks (DNNs). Each modality is provided as input to a DCNN that learns high-level global features. These features are then provided to a DNN to predict the Patient Health Questionnaire (PHQ)-8 depression scores. The PHQ-8 scores are then processed by a DNN for final prediction. Tested on the AVEC2014 database [19], the system offers an accuracy of over 85% for depression prediction from speech and face modalities. Similarly, De Melo et al. [23] fused multiple 3D CNNs (C3D) with the purpose of improving the accuracy of standard CNN-based methods for predicting depressive behaviors. The proposed method proved to be fast and reduced the number of model parameters and the risk of overfitting the networks, offering accuracy of over 80% for depression prediction when tested on the AVEC2014 database [19]. Zhu et al. [24] presented a method that uses joint tuning layers to integrate facial appearances and dynamics through a set of DNNs. Tested on AVEC2014 [19], the method could predict with over 80% accuracy depressive behavior. Yang et al. [25] proposed a system for hybridizing deep and shallow models for depression prediction from audio, video, and text descriptors. Researchers employed a DCNN–DNN model for audio-visual multimodal depression recognition using the PHQ-8 framework, and a paragraph vector (PV) analyzed the interview transcript using an SVM-based model in order to infer the physical and mental conditions of the subject. A random forest (RF) model has been used for depression classification from estimated PHQ-8 scores and inferred conditions of the subject. Experimental results showed that the proposed hybrid method improved the accuracy for depression prediction to over 74% when tested on the AVEC2014 database [19]. Deep regression networks (DepressNets) have also been used for predicting depressive disorders from facial features [26]. A DCNN with a global average pooling layer was first trained on faces affected by depression, which allowed the identification of salient regions of the input image related to different depression severity scores. Further, a multiregion DepressNet was used that contained multiple local deep regression models for different face regions that were jointly learned, and their responses were fused. The method was tested on the AVEC2014 database [19], offering over 82% depression prediction accuracy.
Evaluating anxiety is done separately or together with stress assessment. A way to automatically predict anxiety scores using the different facial expression classification scores was proposed in Reference [27]. Researchers determined a set of features that were analyzed in both raw data and principal component analysis (PCA)-transformed space and looked for correlations between facial expressions and anxiety levels. The results showed that the approach offers reasonable scores for anxiety prediction and confirmed the relationship between facial expressions and anxiety. Active shape models (ASMs) have also been used [28] to extract 68 feature points from five facial areas that were fetched to an SVM-based classifier in order to assess if 18 postgraduates suffered from public speaking anxiety (PSA) in their master’s thesis defense. An accuracy of 99% was obtained when comparing the results provided by the ASM-based method to those from self- and audience assessments. Pediaditis et al. [29] started from a similar premise as the current paper, mentioning the shortcomings of questionnaire-based methods that are commonly used to evaluate stress and anxiety, knowing that these are subjective and prone to bias. They studied the facial features that could be used to predict stress or anxiety efficiently. By combining different classification methods and analyzing sets of 9 to 10 features from different facial areas targeted separately, the methods offered prediction accuracy for both stress and anxiety of up to 73%.
In terms of FACS, limited research has been conducted that used this model for predicting depression, anxiety, or stress. McIntyre et al. [30] used action unit (AU) groups called region units (RUs) determined based on active appearance models (AAMs) and employed Multiboost to classify the facial expressions. The results showed that using this method, depression could be predicted with up to 75% accuracy. Researchers in Reference [31] presented another FACS-based method to investigate the relationship between symptoms of depression and facial expressions over time. Several subjects under depression treatment were recorded during a series of clinical interviews, and their facial expressions were analyzed using FACS. The study showed that when depression symptoms were high, the subjects displayed more facial expressions associated with emotions such as contempt and sadness, confirming the social risk hypothesis of depression [31], which states that people suffering from severe depression withdraw from any social activity and enter a self-protective state. As the symptoms become less severe, subjects show more openness to socializing, and emotions such as happiness or surprise become more conspicuous. Of note is that there have been several studies that have used Red Green Blue—Depth (RGB-D) databases for facial expression recognition, which typically can be collected by means of a Microsoft Kinect sensor with low computational costs, allowing for fast convergence. Szwoch [32] proposed a Facial Expression and Emotion Database (FEEDB) by taking advantage of the fast acquisition, indexing, and storing of the KinectRecorder tool, significantly reducing the time usually spent on data acquisition, indexing, and validation when creating a fully indexed database. In Reference [33], the database was used to predict nine emotions through an algorithm that used local movement detection within the area of the face in order to recognize actual facial expressions. The system offered an average recognition accuracy of above 50%, with the main advantage that it was highly independent of lighting conditions and could provide high accuracy when the sensor was very close to the user. Similarly, Ballihi et al. [34] used upper body movements and facial expressions analyzed from a dataset containing RGB-D movies, extracting a set of positive and negative cues from 2D (RGB) information. A geometric method was used to model depth flows and determine human body dynamics from a depth dataset. Given the temporal changes in pixel and depth intensity, depth features were employed to describe the relationship between the changes in the upper body movements and the affect. The experiments, conducted on a Cam3D dataset [35], showed promising results in detecting people’s moods using upper body postures and monitoring their evolution in time. An interesting study using RGB-D sensors (color and depth) was one conducted by Jaiswal et al. [36], who proposed a framework for helping with the diagnosis of attention deficit hyperactivity disorder (ADHD) and autism spectrum disorder (ASD), neurodevelopmental conditions that impact a significant population of children and adults. Researchers focused on detecting several behavioral markers, providing a fully automatic end-to-end system to predict ADHD and ASD in adults. They performed facial expression analysis based on dynamic deep learning (DDL) and 3D analysis, obtaining a classification rate of 96% for controls versus condition (ADHD/ASD) groups and 94% for the comorbid (ADHD and ASD) versus ASD-only group.
With all this research previously conducted as a premise, this paper introduces for the first time in the literature a neural network-based architecture for predicting levels of depression, anxiety, and stress based on FACS in a nonintrusive and real-time manner. We have also conducted the first study in the literature that has analyzed the correlations between FACS AUs and DASS levels.

2. Materials and Methods

2.1. Psychological Grounds

This research relied on two popular tools, which we describe in the next paragraphs: the Depression Anxiety Stress Scale (DASS) [37] and the Facial Action Coding System (FACS) [38].
DASS is a tool broadly used by psychiatrists to determine the severity of three emotional states: depression, anxiety, and stress. In this study, we employed a DASS version with 42 items, which uses a specific self-analysis questionnaire (SAQ) in which subjects are asked to self-report on 14 items associated with emotional states, as follows [37]:
  • Depression. Dysphoria, anhedonia, inertia, hopelessness, devaluation of life, lack of interest, self-deprecation;
  • Anxiety. Autonomic arousal, situational anxiety, skeletal muscle effects, subjective experience of anxious affect;
  • Stress. Difficulty relaxing, nervous arousal, easily agitated/upset.
For each of these items, the subjects answer particular questions by self-rating, using a score on a scale from 0 to 3. The sum of all scores obtained from each item determines the severity of the emotional state as follows: Normal (0–9 score for depression, 0–7 score for anxiety, 0–14 score for stress), Mild (10–13 score for depression, 8–9 score for anxiety, 15–18 score for stress), Moderate (14–20 score for depression, 10–14 score for anxiety, 19–25 score for stress), Severe (21–27 score for depression, 15–19 score for anxiety, 26–33 score for stress), and Extremely Severe (28+ score for depression, 20+ score for anxiety, 34+ score for stress).
FACS is a system developed by Eckman and Friesen [38] for analyzing the microexpressions on human faces and determining the emotion of the analyzed subjects. It is based on the idea that each emotion can be associated with different facial muscle patterns and that by analyzing facial regions where these muscles are activated, we can determine the emotion of the individual. The main advantage of FACS compared to other face analysis methods is that it can determine hidden emotions, even if the subject is trying to mimic different ones. FACS divides the face into 46 action units (AUs), which can be either nonadditive (the activity of an AU is independent of the activity of another AU) or additive (when an AU is activated, it triggers the activation of another AU or group of AUs). Based on FACS, the AUs’ presence and intensities are measured using the following levels: Level O, AU is not active; Level A, trace; Level B, slight; Level C, marked or pronounced; Level D, severe or extreme; and Level E, maximum.
FACS is typically used to determine the emotions of individuals. In this paper, we took this challenge further and tried to determine the possibility of using FACS to evaluate the severity of depression, anxiety, and stress. Although limited, there is previous research that has suggested possible connections between emotions and emotional states (depression, anxiety, or stress). In Reference [39], researchers proposed an emotions recognition system for classifying four basic emotions (happiness, fury, sadness, and neutral) using a speech modality, and they showed that there were links between these four emotions and depression or anxiety. Similarly, in Reference [40], researchers studied emotion-related variables in correlation with anxiety and depression, and they showed that there was a close link between depression and a high frequency of negative emotions and a low frequency of positive emotions, as well as between anxiety and specific patterns in the frequency of the emotional experience. Similarly, stress was linked to increased negative emotions, such as anger [41], while positive emotions (such as happiness) were linked to an increased physiological ability to recover from chronic stress [42].
In the following section, we present how these two tools are used in our proposed architecture as well as detail its structure, the neural network configuration, and the training and testing phases.

2.2. Proposed Architecture

We propose a novel nonintrusive multilayer neural network-based architecture for predicting DASS levels by analyzing facial expressions using FACS. The architecture has a unique structure on three layers: in the first layer, the video frames are normalized, and the analyzed AUs are classified using an AAM-based method and a set of multiclass SVMs; in the second layer, a matrix is built based on the intensity levels of the selected AUs; and in the third layer, a neural network-based structure is developed that analyzes the patterns from the matrix built in the second layer and predicts the DASS levels (Normal, Mild, Moderate, Severe, or Extremely Severe) for each of the three emotional states. The reason for building a system on three layers instead of an end-to-end learning model was that we wanted to study the possible correlations between FACS and the analyzed AUs (classified in the first layer) and the predicted DASS levels (from the top layer). The study is also part of a broader research project aiming to build a system for diagnosing mental disorders and physical diseases for applications in the healthcare industry in English-speaking countries by analyzing information collected from different sources in a noninvasive manner. Hence, we opted for a modular approach that would allow us to easily add other modalities as input, other psychological conditions as output, as well as weight and fuse different modalities in the second layer matrix, reducing these problems to pattern recognition tasks.
The architecture was built using a Java virtual machine (JVM) with Eclipse as an integrated development environment (IDE) and Scala as a programming language using Spark Library. The testbed has an Intel i7 processor with 8 GB of random access memory (RAM) running Linux Solaris 11.3 as an operating system (OS). In the following subsections, we detail the three constituting layers.

2.2.1. The First Layer

As mentioned previously, the last FACS revision proposed 46 primary AUs [43], but only 30 of them were anatomically correlated with facial muscle contraction: 12 in the upper part of the face and 18 in the lower part [43]. Out of these 30 AUs, we analyzed only the following:
  • Upper face: AU2 (outer brow raiser), AU4 (brow lowerer), AU7 (lid tightener), AU43 (eyes closed), AU6 (cheek raiser), AU1 (inner brow raiser), AU5 (upper lid raiser), and AU45 (blink);
  • Lower face: AU8 (lips toward each other), AU25 (lips part), AU16 (lower lip depressor), AU11 (nasolabial deepener), AU13 (sharp lip puller), AU15 (lip corner depressor), AU17 (chin raiser), AU10 (upper lip raiser), AU18 (lip pucker), AU19 (tongue show), AU20 (lip stretcher), AU22 (lip funneler), AU23 (lip tightener), AU14 (dimpler), AU12 (lip corner puller), AU24 (lip pressor), AU26 (jaw drop), AU27 (mouth stretch), AU9 (nose wrinkler), and AU28 (lip suck).
All of these selected AUs are part of the standard set used in most facial expression recognition (FER) systems based on FACS [44,45]. We excluded AU41 (lid droop), AU42 (slit), and AU46 (wink) from the upper part of the face because these AUs were not coded with intensity levels in the last FACS revision [43], and a binary classification (AU present/absent) in a context where all other AUs were evaluated based on a level from O to E could negatively impact the accuracy of the system. Apart from the 28 AUs mentioned above, we also analyzed AU33 (cheek blow), AU34(cheek puff), and AU35 (cheek suck) to enrich the facial features of the cheek areas, as these 3 AUs were coded with intensity levels in the latest FACS revision [43]. Therefore, we used 31 AUs that could be classified fast and with high accuracy: For the remaining AUs, more complex methods would have been required, which would also have required more processing time, so they were not evaluated in this study, since we aimed to build a system able to predict the DASS levels in real time. Another reason for which we selected this set of AUs was that they are present in several large databases employed for testing FER systems [44,45], such as Cohn-Kanade extended (CK+) and MMI, which we used in this study for evaluating the accuracy of the AU classification task. All AUs analyzed were treated individually, as nonadditive.
The first step is face detection, for which we created a statistical model of the skin, and we classified pixels from the analyzed frame in facial and nonfacial pixels, obtaining an image containing only the areas classified as related to the face [46]. We also used a noise reduction filter based on Gabor filters and non-negative matrix factorization [47]. After these filters were applied, we obtained a candidate face for the current video frame. We continued with determining the position of the eyes, as they are known to have the lowest variation of all face elements [48]. Therefore, after face detection, we evaluated the image horizontally, and we calculated the number of white pixels in each row. The row that contained the highest number of white pixels from the upper part of the candidate face was considered the median of the block containing the eyes, while the one from the lower part of the candidate face was the median of the mouth block. Because we know that the eyes are located between 2/20 and 9/20 of face height and that the mouth is located between 11/20 and 15/20 of face height [49], we performed additional validations based on these ratios to ensure that the blocks delimited as pertaining to the eyes and mouth were accurate. This method proved to be up to 80% faster than other state-of-the-art methods based on facial landmarks when analyzing frontal faces. Therefore, we opted for this method, because in our study, we wanted to exclusively analyze frontal face video recordings and because we aimed for a real-time prediction of the three emotional states, which requires fast convergent methods.
After the face and the eyes and mouth blocks were detected, we extracted facial features using AAM [50], as this method is fast for similar problems [51,52] and offers good results for AU detection [53]. We employed a Principal Component Analysis (PCA) to produce a parameterized model describing the faces used for training and to estimate new faces [50,51]. Matching the model’s parameters with the candidate faces was done in the first instance by matching the eye and mouth areas, then using a steepest descent algorithm and Jacobian matrices to match the entire parametrized model [51]. The parameters of the model were evaluated iteratively until they provided the most accurate description of the candidate face. We used a vector of coordinates from a set of facial landmarks called shape [51]. All of the shapes were aligned with the standard model, and we studied the shape variation using generalized Procrustes analysis (GPA) [54]. We determined a set of nonrigid parameters for 72 facial landmarks (as seen in Figure 1), and testing the algorithm on 128 different subjects using leave-one-out cross-validation, we determined, through trial and error, that 1500 iterations were required to determine these parameters with over 90% accuracy.
The 72 nonrigid parameters are provided as input to a set of 31 multiclass SVMs used for the classification of the 31 AUs’ intensity levels, with one multiclass SVM for each analyzed AU. We used SVMs for classification because they have proven to be accurate and fast for similar problems [55]. The SVM classifiers search for the hyperplanes that maximize the margins of a class and use a vector to separate hyperplanes and a bias estimated to minimize the error of the training set. We used LIBSVM [56] for training and testing the multiclass SVM classifiers. Each such classifier receives as input the 72 nonrigid parameters computed for a frame from the video sample and is trained to associate the analyzed frame with 1 of 6 classes (for the 6 AU intensity levels). The output of the SVMs will, therefore, provide the intensity level of the analyzed AU (O, A, B, C, D, or E). The multiclass SVMs were evaluated in cross-database tests on the CK+ [57], MMI [58], Japanese Female Facial Expression (JAFFE) [59], Belfast [60], Acted Facial Expressions in the Wild (AFEW) [61], and RU-FACS [62] databases. The results obtained are presented in Figure 2, where we can observe the classification accuracy for each AU in each of these cross-database tests. The results were averaged such that, for example, for MMI/CK+, we trained the system on MMI and tested it on each sample from CK+, and then we trained on CK+ and tested each sample from MMI, averaging the accuracy of all tests conducted.
As can be observed, for most of the cross-database tests, all AUs were classified with over 90% accuracy, which offered a strong baseline for the more complex task of predicting the DASS levels. We also observed a lower accuracy obtained when the JAFFE [59] database was used for training or testing, mainly because this database contains Mongoloid subjects, while the other databases have Caucasian subjects. Given the fact that in our database, we used Caucasian subjects exclusively, this did not pose a concern for further tests conducted to evaluate DASS levels. We compared the average accuracy obtained by using this multiclass SVM-based method to other state-of-the-art CNN-based methods for AU classification (CNN-based classifiers [63], multilabel CNN classifiers [64], and a combination of CNN-based and SVM-based classifiers [65]) that are known to offer high classification accuracy, and the results are displayed in Figure 3. As can be observed, the proposed multiclass SVM-based method offered an accuracy comparable to the one obtained using the method with CNN-based and SVM-based classifiers proposed in Reference [65], but it was significantly faster (hence why we used it in this study, as we aimed to build a system able to predict DASS levels in real time).

2.2.2. The Second Layer

The output of the first layer consists of an intensity level from O to E estimated for each of the 31 AUs. This level is normalized in [0,1] intervals as follows: level A = 0.2, level B = 0.4, level C = 0.6, level D = 0.8, level E = 0.9, AU absence = 0. This normalization is required to allow the neural network-based pattern analysis in the third layer. With the normalized values, in this second layer, a facial matrix (FM) is built. The FM contains a column for each frame of the analyzed recording, each column consisting of the intensity levels of all the AUs present in that given frame. Figure 4 shows the structure of the first and second layers. When 30 new columns are added to the FM (corresponding to 30 frames), they are sent to the third layer, where a neural structure analyzes the patterns and provides the predicted DASS levels of the analyzed subject.

2.2.3. The Third Layer

As mentioned in previous sections, although there are multiple methods in the literature for predicting the three emotional states, in this paper, we evaluate only neural networks: hence, we aimed to determine the best neural network-based structure and configuration for predicting DASS levels using facial features, with high accuracy. There are typically 4 types of neural networks commonly used: feed-forward neural networks (FFNNs), radial basis function neural networks (RBFNNs), recurrent neural networks (RNNs), and convolutional neural networks (CNNs, a particular type of FFNN). RBFNNs [66] have limited applications in the affective computing (AC) domain and are mainly used for energy restoration systems, while RNNs [67], although designed to process sequential data, have been studied and have proven to be more efficient in speech and handwriting recognition systems, while for facial analysis tasks they have been understudied. On the other side, FFNNs have been successfully used for image processing and voice recognition, offering efficient classification in cases where the input is nonlinear or affected by noise [68]. Based on these aspects, in this paper, we focused exclusively on analyzing how FFNNs can be used to predict DASS levels with high accuracy.
Because for most pattern recognition problems, two hidden layers are enough [69], we only evaluated FFNNs with one or two hidden layers. As activation functions, we evaluated tanh, sigmoid, Rectified Linear Unit (ReLU), and softmax. Because ReLU outputs are in (0, +∞) intervals and given the fact that they are affected by the exceeding gradient problem [70], we only used them in hidden layers. Softmax, on the other hand, has been shown to be inefficient in hidden layers [71], and hence we used it only for the output layer. We also analyzed the use of only one FFNN to predict all three emotional states as well as dedicated FFNNs for each emotional state. Therefore, we evaluated whether a single FFNN can be used to predict depression, anxiety, and stress or whether three FFNNs used for each of the three emotional states provide better accuracy, in each case determining the optimal hyperparameters of the FFNNs. The results obtained after testing all possible combinations are displayed in Figure 5.
As can be observed, the highest accuracy for the prediction of all three emotional states was obtained when we used a single FFNN to model all three emotional states. This finding shows that depression, anxiety, and stress are strongly connected from a facial point of view, validating the argument mentioned in Reference [1]. The optimal FFNN had 3 output nodes in [0,1] intervals representing the DASS levels predicted for each of the three emotional states, as follows: [0,0.15] for Normal, [0.15,0.3] for Mild, [0.3,0.6] for Moderate, [0.6,0.8] for Severe, and over 0.8 for Extremely Severe. The thresholds for each of the 5 levels were determined as optimal through trial and error.
We denoted this FFNN: Face Depression Anxiety Stress Scale Neural Network (FDASSNN). The structure determined as optimal had two hidden layers, with tanh as the activation function for the first hidden layer and sigmoid for the second hidden layer and the output layer. We considered N i n , the number of training vectors provided, as input, with a set of N-dimensional input vectors ( X ( F D A S S N N ) = { x ( n ) ( F D A S S N N ) } ,   n = 1 , 2 N i n , such as x ( FDASSNN ) = [ x ( 1 ) ( F D A S S N N ) , x ( 2 ) ( F D A S S N N ) , , x ( N ) ( F D A S S N N ) ] T ); and K o u t , the number of output vectors, with a set of K-dimensional output vectors ( Y ( F D A S S N N ) = { y ( k ) ( FDASSNN ) } ,   k = 1 , 2 K o u t ); and we denoted the weight matrix between the input layer and the first hidden layer W ( 2 L , H 1 ) ( F D A S S N N ) , the weight matrix between the first hidden layer and the second hidden layer W ( 2 L , H 2 ) ( F D A S S N N ) , the weight matrix between the second hidden layer and the output layer W ( 2 L , O ) ( F D A S S N N ) , L ( 1 ) ( F D A S S N N ) (the number of hidden nodes from the first hidden layer), and L ( 2 ) ( F D A S S N N ) (the number of hidden nodes from the second hidden layer). This structure can be observed in Figure 6.
By splitting the training data into a training and validation dataset with an 80:20 dataset split ratio and fine-tuning the FDASSNN hyperparameters, we determined that the optimal number of neurons required in the first hidden layer was 75, with 65 neurons for the second layer (Figure 7). During training, the bias for the hidden layers was set to 0. The optimal learning rate determined was 0.2, the optimal momentum was 0.4, and we required 7000 training epochs to train FDASSNN, with the process lasting 5.5 h.
As a training method, we used backpropagation because it is largely used for training FFNNs and in pattern recognition tasks, as it is fast and accurate [72]. We used gradient descent (GD) to optimize the weights and biases and minimize the final error, with the purpose of reducing it to below 0.02. We also used the Nguyen–Widrow initialization method [73] to spread the weights in the initial state evenly. The error was calculated using the average absolute relative error (AARE) formula, with y p as desired output and y e as estimated output:
AARE = 1 N i n p = 1 N i n | ( y p y e y e ) | .

2.3. Operating Modes

The proposed architecture operates in two phases: training and testing.
In the training phase, videos from the training dataset are provided as input to the first layer, frame by frame. In the first layer, the face and the eye and mouth facial blocks are detected, and the 72 nonrigid parameters are determined and provided as input to the set of 31 multiclass SVMs, previously trained on the CK+ database, which output the intensity levels of each of the 31 AUs. The intensity levels are sent to the second layer, which normalizes them in [0;1] intervals and adds the normalized results in a new column to the FM. When 30 new columns are present in the FM, they are sent to the third layer, where the FDASSNN is trained through backpropagation, comparing the predicted output to the expected one (the results obtained from SAQ) and correcting the weights in order to minimize the AARE. We opted for 30 columns because they corresponded to 30 frames, and given the fact that we had a frame rate of 30 fps, they linked to 1 s of facial recording, high enough to catch microexpressions (lasting maximum 500 milliseconds [20]) and low enough to avoid overfitting of the FDASSNN. The training process stops when the AARE is minimized (i.e., has a value lower than 0.02) or when 7000 training epochs are reached.
In the testing phase, the steps are similar. A video sequence that is not used in the training phase is provided as input to the first layer, frame by frame. The first layer detects the face and corresponding eye and mouth blocks and determines the 72 nonrigid facial parameters, based on which the 31 multiclass SVMs estimate the intensity levels of the 31 AUs. The intensity levels are sent to the second layer, normalized, and added in a new column to the FM. When we have 30 new columns in the FM, they are sent to the third layer, where the FDASSNN outputs the DASS levels. One final rule-based classifier (RBC) is used, which collects the predictions from the FDASSNN and provides a final result, as follows:
  • If the same prediction is provided for five consecutive seconds, that prediction is the final output of the system;
  • Otherwise, the RBC marks the output as Undefined, which means the information is insufficient for a prediction.
The user interface (UI) of the application can be observed in Figure 8.

3. Results

In this section, we present the database created for evaluating the proposed architecture, as well as the results obtained from intrasubject and intersubject methodologies. We conclude this section with a comparison between the proposed architecture and other state-of-the-art methods for evaluating depression, anxiety, and stress based on facial analysis.

3.1. FACS–DASS Database

Because of a lack of a database that links FACS with DASS levels, we created our own database by collecting samples from 128 Caucasian subjects (64 males and 64 females) aged between 18 and 35 years old. The subjects participated in conformity with the Helsinki Ethical Declaration [74] and were chosen in order to display different severities of stress, depression, and anxiety, 20 of them suffering from various stages of MDD, 19 of various stages of GAD, and 17 of PTSD. No subject displayed comorbid conditions (e.g., subjects suffering from MDD and also affected by GAD). The remaining healthy subjects were evaluated by trained psychologists, and they displayed different anxiety, stress, and depression levels, but were not suffering from any specific disorders. The distribution of the DASS levels across the 128 subjects is presented in Figure 9, which shows that we had intersubject variability in order to use the database for testing the proposed architecture with the intersubject methodology.
During each session with each of the 128 subjects, the following samples were collected:
  • Frontal face recordings when subjects were asked to watch emotion-inducing videos (inducing the six basic emotions: sadness, fear, happiness, anger, surprise, and disgust). In total, six frontal face one-minute video recordings were collected, one for each emotion;
  • Six discrete emotion questionnaire (DEQ) [75] results after the subjects watched each of the six emotion-inducing videos;
  • SAQ results after the subjects watched the emotion-inducing videos;
  • Five frontal face one-minute video recordings when the subjects were watching emotionally neutral videos;
  • SAQ results after the subjects watched emotionally neutral videos.
The subjects were naive participants and were not medicated at the time the samples were collected. Both the SAQ and DEQ results were evaluated by five trained psychologists to ensure they were consistent, and the questionnaire results and associated facial recordings that were not considered proper after reviewing the video recordings and interviewing each subject were excluded (2.5% of the samples were excluded). The DEQ results were used to validate that the emotion induced by the videos was indeed induced in the subject, and recordings where the DEQ provided an emotion other than the one induced were also excluded (2% of recordings were excluded). To induce the six emotions, we used videos from the LIRIS-ACCEDE database [76]. Because these videos were between 8 and 12 seconds, but analyzing facial expressions needed more time, we combined multiple videos for the same emotion in a one-minute compilation. We chose to induce emotions not only to generate emotionally rich samples but also to increase the number of AUs activated in a short time, which was useful when analyzing how the induced emotion influenced predictions as well as the system’s ability to be accurate in naturalistic conditions. For recording, we used a Sony PMW-300 One camera with 30 fps and 1024 × 768 resolution.
The samples were collected for each subject six times, with a break of two weeks between sessions in order to include the possible dynamics of emotional states over time. The average intrasubject DASS variability was 0.8 levels from one session to another, which allowed for further tests using intrasubject methodology. At the end of the three months, we had, for each subject, 36 facial recordings where emotion was induced (one for each emotion from each recording session), 30 facial recordings where no emotion was induced, 36 DEQ results, and 12 SAQ results. Based on how the information was collected, we divided the database into two datasets:
  • Dataset (controlled) (DSC), containing samples collected when emotion was induced; and
  • Dataset (random) (DSR), containing samples collected when no emotion was induced.
Note that the average time taken to fill in the SAQ was 25.4 min.
To evaluate the proposed system, we used the following statistical measures (where TP refers to the number of true positives, TN to the number of true negatives, FP to the number of false positives, and FN to the number of false negatives) [77].
Accuracy (AC):
AC = ( T P + T N ) / ( T P + T N + F P + F N ) .
Sensitivity or true positive rate (SE/TPR):
TPR = T P / ( T P + F N ) .
Specificity or true negative rate (SP/TNR):
TNR = T N / ( T N + F P ) .
Note that, in both the intrasubject and intersubject methodologies, we evaluated the prediction of each DASS level. For example, when predicting the Normal level, we considered TP to be associated with cases where the system correctly predicted the Normal level, TN to be associated with cases where the system correctly predicted that it was not a Normal level, FP where the system incorrectly predicted the Normal level, and FN when the system incorrectly predicted that it is not a Normal level.

3.2. Intrasubject Methodology

For intrasubject methodology, we trained and tested the architecture using samples pertaining to the same subject, varying the testing and training datasets (DSC, DSR, or both) and using leave-one-out cross-validation. The results can be observed in Table 1.
Since the output of the architecture is represented by the five DASS levels, a successful prediction was considered to be when the predicted DASS level was the same as the one that resulted from a subject’s completed SAQ during the same session the sample used for testing was collected. As we can observe, the highest prediction accuracy for all three emotional states was obtained when the DSC was used for both training and testing. In that case, we obtained 89.3% accuracy for depression, 79.2% accuracy for anxiety, and 92.1% for stress. If we kept the DSC for training and we used nonemotion-elicited samples from the DSR for testing the architecture, we observed the accuracy, sensitivity, and specificity decrease at less than 2%, which shows that, when training and testing using samples from the same subject, inducing emotions was essential only for training the system, while for testing we could use video recordings collected in naturalistic conditions. This observation makes this architecture practical, as it only needs to be pretrained with emotion-induced samples from the subject, while subsequent assessments of a subject’s DASS levels can be done in real conditions, without inducing emotions, and hence it can be used in real time to monitor the facial dynamics of a patient and evaluate the level of stress, anxiety, or depression. We observe in Table 1 and Figure 10 that the highest accuracy for all three emotional states was obtained for Normal (90.4% for depression in the most practical case when we trained using samples from the DSC and tested using samples from the DSR, 84.1% for anxiety and 94.2% for stress) and Extremely Severe levels (93.2% for depression, 82.9% for anxiety, and 93.5% for stress). For the other intermediary levels, the accuracy was lower, and these were most often mistaken for neighboring levels (e.g., the Mild level was mostly mistaken for the Moderate level for depression and stress, and for the Normal level for anxiety). We conclude that the proposed architecture could successfully discriminate between depressive/nondepressive, anxious/nonanxious, and stressed/nonstressed subjects, but for discriminating between DASS levels, the results were lower, and other methods need to be researched to enhance this further, such as including more AUs in the facial analysis, fusing the FFNN-based method with other state-of-the-art methods for facial analysis, or using the face modality alongside other modalities (such as voice) in a multimodal approach.
In the most practical case, when we trained the system using samples from the DSC and we tested it using samples from the DSR, the average time needed to predict the three emotional states was 14 s (as seen in Figure 11). As processing one second of video recording lasts, on average, three seconds, the time needed for predicting the DASS levels in the intrasubject methodology was 56 s, significantly faster than the time needed to complete the SAQ (25.4 min), which makes this architecture fast enough to be used in real time and attractive for health practitioners to evaluate and monitor the DASS levels of their patients.
As we observed that the architecture provided high accuracy for predicting Normal and Extremely Severe levels, we evaluated the same architecture on the task of determining if the analyzed subjects were affected by MDD, GAD, or PTSD. We trained the architecture using the DSC samples from the subjects affected by these conditions and we tested the architecture on the samples from the DSR (since this was the most practical case). We also determined the optimal FFNN hyperparameters in this case, keeping the same 80:20 split ratio, and we obtained as optimal an FFNN with two hidden layers, tanh as the activation function for the first hidden layer and sigmoid for the second hidden layer and the output layer. The structure was similar to the one obtained for predicting DASS levels, the difference being in the number of neurons for the two hidden layers: in this case, we needed 30 neurons for the first hidden layer and 25 for the second hidden layer. During training, the bias for the hidden layers was also set to 0. The FFNN had three output neurons providing a binary result for each of the three disorders (MDD, GAD, and PTSD), such that if the subject was suffering from the disorder, the output was 1, and if he/she was healthy, the output was 0. The results are displayed in Figure 12, and we can see that we obtained over 90% prediction accuracy and over 89% sensitivity and specificity for predicting MDD and PTSD, which shows that our architecture can be successfully used for diagnosing these disorders.

3.3. Intersubject Methodology

In the intersubject methodology, we trained the architecture using samples pertaining to a set of subjects and we tested it on samples associated with a different subject that was not used for training, varying the testing and training datasets (DSC, DSR, or both). We used a leave-one-out approach such that we trained on DSC, DSR, or both DSC and DSR samples from 127 subjects, and we tested the samples from the remaining subject. We repeated the tests until all samples were used for testing the system and all subjects had been involved in the testing phase, and we averaged the results, which can be observed in Table 2.
The results were consistent with those obtained in the intrasubject methodology, with the highest prediction accuracy being obtained when we trained and tested the architecture using samples from the DSC, in this case reaching 79.8% accuracy for depression, 68.7% accuracy for anxiety, and 81.5% accuracy for stress. We observed that when we used samples from the DSR for both testing and training, the accuracy was 14% lower, while when we trained the system using samples from the DSC from 127 subjects and tested it with samples from the DSR from the remaining subject, the accuracy was only 1.5% lower.
We reached a conclusion similar to that from the intrasubject methodology, with the emotion-inducing facial activity adding more value only in the training phase, while for testing we could use recordings collected in nonemotion eliciting scenarios without any significant influence on the accuracy, sensitivity, and specificity of the architecture. For the intrasubject methodology, this observation was even more valuable, as the system could be pretrained using video recordings of subjects whose emotions were induced, while the end-user could assess the DASS levels in naturalistic scenarios, without the need for training the system with his/her own samples.
We also observed (Figure 13) that the architecture offered an accuracy of over 80% for stress and depression and over 71% for anxiety when predicting Normal and Extremely Severe levels, confirming, in this case as well, the fact that it can successfully discriminate between healthy subjects and those affected by depression, anxiety, and stress, but cannot discriminate with the same accuracy between intermediary levels (Mild, Moderate, and Severe), with these being often mistaken for their neighboring levels (e.g., the Mild level mistaken for the Normal level for depression and for the Moderate level for anxiety and stress), similar to the intrasubject methodology results. The fact that the results obtained with the two methodologies were similar shows that the proposed architecture is robust.
The average duration of the video sample needed for predicting all three emotional states was 16 s (as seen in Figure 14), and therefore the average DASS level prediction time was 64 s, a lot faster than the time needed for completing the SAQ, and hence the architecture can be used for monitoring the three emotional states in real time and can be an attractive tool for health practitioners to assess DASS levels, if the prediction of intermediary DASS levels is improved.
We performed a test similar to the one conducted for the intrasubject methodology, and we evaluated the architecture on the task of determining if the analyzed subjects were affected by MDD, GAD, or PTSD. We used the same FFNN and hyperparameters detailed for the intrasubject methodology. We trained the architecture using DSC samples from 127 subjects and we tested it using samples from the DSR for the remaining subject, repeating the tests until all samples and all subjects were involved in the testing phase. The results are displayed in Figure 15, and we observed over 93% accuracy for MDD and PTSD and over 85% accuracy for GAD, which shows that the architecture can be used as a tool for diagnosing these psychological disorders. Moreover, the accuracy obtained in the intersubject methodology was higher than the one obtained in the intrasubject methodology, which is explained by the fact that the training data variability was higher in the intersubject tests.
We also tested the accuracy of the architecture based on the number of subjects we used for training in order to determine the lowest number of subjects required to get satisfying results. For this, we trained the system on DSC samples from a set of subjects and tested it on DSR samples from subjects that were not involved in training, repeating the tests until all DSR samples and all users were used for testing and averaging the results. The results are displayed in Figure 16, and we show that we needed a minimum of 116 subjects for stress and depression for the system to provide reasonable results, while for anxiety the 127 subjects used in this study did not seem to be sufficient. This also shows that, for stress and depression, increasing the number of subjects in the database might not improve significantly the prediction accuracy, and hence other state-of-the-art methods can be fused with the FFNN-based method presented in this study to improve the accuracy more. For anxiety, increasing the number of subjects used for training could definitely lead to an increase in the prediction accuracy of this emotional state.

3.4. Correlations between Emotions, FACS, and DASS Levels

As we observed that the samples collected when emotion was induced led to higher accuracy when used in the training phase, we built a background application that searched for AUs of intensity D or E in the FM whenever an Extremely Severe level was correctly predicted. The top four AUs determined for each of the three emotional states are displayed in Table 3. We see that these AU combinations were unique for each of the three emotional states, and we modified the architecture such that when the AUs were present at intensity level D or E, we provided the Extremely Severe score without passing through the neural network structure. This enhancement led to increasing the accuracy for predicting the Extremely Severe and Severe levels by 8% (as the Severe levels were less often mistaken for Extremely Severe thanks to this enhancement) and the overall accuracy of the architecture by 5%.
We also performed intersubject tests in which we trained the FDASSNN with samples from the DSC pertaining to one emotion only, from 127 subjects, and we tested it on samples from the DSR pertaining to the remaining subject, repeating the tests until all samples from the DSR and all subjects were involved in the testing phase, as well as until all emotions were used for the training phase. We identified that when using the samples where happiness or sadness were induced for training, all depression levels could be predicted with over 85% accuracy; when using samples where surprise or disgust were induced for training, the anxiety levels were predicted with over 80% accuracy; and when using the samples where sadness and disgust were induced for training, all of the stress levels were predicted with over 90% accuracy. The overall accuracy obtained in these cases and the associated emotions are presented in Table 4. Therefore, we show that by selectively training the architecture with samples where specific emotions were induced, not only did the accuracy improve, but intermediary levels were predicted with higher accuracy, too.

3.5. Comparison to State-of-the-Art Methods

In order to compare the performance of the proposed architecture to that of the state-of-the-art methods, we used the database proposed in this study as well as ANUStressDB [12] for evaluating stress and AVEC2014 [19] for evaluating depression. We compared our proposed method to the state-of-the-art systems presented in the introductory section as well as to several classic backbone networks, such as VGG (VGG-11, VGG-13, VGG-16, VGG19), ResNet (ResNet-18, ResNet34, ResNet-50, ResNet-101, ResNet-152), Inception (v3 and v4), GoogLeNet, BN-Inception, and AlexNet [78].
For stress, we kept the multiclass SVM classifiers trained using CK+ and we trained the FFNN based on ANUStressDB [12], using leave-one-out cross-validation and evaluating only the results provided by the output neuron assessing stress levels (for Normal and Mild, we considered the subject to be nonstressed, while for other levels, we considered the subject to be stressed). The results are displayed in Figure 17. As we can observe, our architecture surpassed all other state-of-the-art methods, offering 88% accuracy with 85% sensitivity and 86% specificity in the task of discriminating between stressed and nonstressed subjects, with a lowest prediction time of 70 s.
For depression, we used AVEC2014 [19], and the results are detailed in Figure 18. As can be observed, our architecture offered accuracy 1.5% higher than the FV LBP-TOPs method [18], and it also offered one of the lowest prediction times (85 s) (AlexNet offered the lowest prediction time of 76 s in this case, but with significantly lower accuracy).
We also tested these state-of-the-art methods on our own database using the intersubject methodology with a leave-one-out approach (training the system on DSC samples from 127 subjects and testing it on DSR samples from the remaining subject, repeating the tests until all DSR samples and all subjects were used for testing and averaging the results), and the accuracy, sensitivity, and specificity obtained can be observed in Figure 19 (depression), Figure 20 (anxiety), and Figure 21 (stress).
For depression, the proposed architecture offered accuracy up to 5% higher compared to other methods in the literature, surpassing these methods also in terms of sensitivity and specificity. For anxiety, we obtained 5% higher accuracy than the method employing ASM and SVM classification [28], as well as sensitivity and specificity up to 4% higher. For stress, our architecture offered 3% higher accuracy and 2% higher sensitivity and specificity compared to the most accurate method in the literature (using Eigen features and template matching) [14]. In all these tests, the method proposed in this study offered the lowest prediction time, which shows that it is robust and can be confidently used for real-time monitoring. Because our architecture offered an accuracy higher than other systems that evaluate these three emotional states separately, we can state that the higher results were not only as a result of the use of FACS and an optimal FFNN, but also because we treated the three emotional states together, incorporating their possible correlations [1].

4. Discussion

We proposed a novel multilayer neural network-based architecture for determining DASS levels by analyzing the facial expression dynamics of a subject using FACS. This is the first study of this kind in the literature that has proposed a unique architecture built on three layers designed to offer high accuracy and low prediction time: in the first layer, the video frames are normalized, and the analyzed AUs are classified using an AAM-based method and a set of multiclass SVMs; in the second layer, a matrix is built containing the intensity levels of the selected AUs for each video frame; in the third layer, we determine the optimal structure of an FFNN trained to analyze the patterns from the matrix computed in the second layer and predict the DASS levels (Normal, Mild, Moderate, Severe, or Extremely Severe) for each of the three emotional states. To test the proposed architecture, we built a new database containing frontal face recordings of subjects as well as SAQ results in both controlled scenarios (when emotion was induced in the subjects) and random scenarios (when no emotion was induced), ensuring it displayed both intrasubject and intersubject variability. We obtained an accuracy of 87.2% for depression, 77.9% for anxiety, and 90.2% for stress with the intrasubject methodology and 78.6% for depression, 67.3% for anxiety, and 80.4% for stress with the intersubject methodology. In both methodologies, we showed that the intermediary DASS levels (Mild, Moderate, and Severe) were predicted with an accuracy significantly lower, as these levels were most often mistaken for neighboring levels, and therefore the proposed architecture can successfully predict whether a subject is affected by depression, anxiety, or stress or is healthy, but does not offer the same accurate results for distinguishing between the severity levels of the three emotional states. Other methods can be fused with the current FFNN-based one or a multimodal approach can be evaluated to increase the accuracy of the intermediary level prediction. The average prediction time was 56 s (intrasubject methodology) and 64 s (intersubject methodology), faster than the time needed to complete the SAQ (25.4 min), and therefore the architecture is fast and can be used in real time to monitor the three emotional states, an attractive alternative to questionnaire-based methods and a tool to help health practitioners if the accuracy for predicting the intermediary DASS levels is improved.
We used the proposed architecture to discriminate between healthy subjects and those affected by major depressive disorder (MDD), post-traumatic stress disorder (PTSD), or generalized anxiety disorder (GAD), and we obtained over 93% accuracy in predicting MDD and PTSD and over 85% accuracy in predicting GAD: hence, the architecture can be used by health practitioners as a tool for diagnosing these psychological disorders. We determined a set of connections between DASS levels and analyzed AUs, with this being the first study in the literature that has identified these connections. By modifying the architecture to include these connections, the overall accuracy increased up to 5%. We also identified correlations between the induced emotions and DASS levels, and we showed that by inducing only specific emotions for each of the three emotional states, the accuracy improved up to 8%, and the intermediary levels were also predicted with high accuracy. This shows that inducing specific emotions triggered a set of facial features that could be more relevant for each emotional state analyzed.
We compared our proposed architecture to other state-of-the-art methods, and we showed that it provided up to 5% more accuracy for depression and anxiety prediction and 3% higher accuracy for stress prediction.
The architecture can be further enhanced by analyzing other AUs or fusing at the feature or score level other methods with the FFNN-based one proposed in this study. We also showed that increasing the number of subjects and samples in the database could lead to increasing the prediction accuracy for anxiety. Since this architecture was designed to analyze exclusively frontal face video recordings, other methods (such as using facial landmarks) to increase accuracy in the case of head pose changes or partial face occlusions can be explored to enhance the proposed architecture. The architecture can be used as a diagnostic tool for health practitioners, but can also be integrated in applications such as virtual psychology [3], computer-driven therapy [3], career counseling [3], or even diagnoses of physical diseases based on patterns of stress, anxiety, or depression levels [4,5,6].

Author Contributions

Conceptualization, M.G.; methodology, M.G. and N.V.; software, M.G.; validation, M.G. and N.V.; formal analysis, M.G.; investigation, M.G. and N.V.; resources, M.G. and N.V.; data curation, M.G.; writing—original draft preparation, M.G.; writing—review and editing, M.G. and N.V.; visualization, M.G.; project administration, M.G. and N.V; supervision, M.G. and N.V.

Funding

This research received no external funding.

Acknowledgments

No further acknowledgments are made.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Friedman, E.S.; Clark, D.B.; Gershon, S. Stress, anxiety, and depression: Review of biological, diagnostic, and nosologic issues. J. Anxiety Disord. 1992, 6, 337–363. [Google Scholar] [CrossRef]
  2. Bertrand, M.; Mullainathan, S. Do People Mean What They Say? Implications for Subjective Survey Data. Am. Econ. Rev. 2001, 91, 67–72. [Google Scholar] [CrossRef]
  3. Kolakowska, A.; Ladowska, A.; Szwoch, M.; Szwock, W.; Wrobel, M. Emotion Recognition and Its Applications. In Human-Computer Systems Interaction: Backgrounds and Applications; Springer: New York, NY, USA, 2014. [Google Scholar]
  4. Covic, T.; Cumming, S.R.; Pallant, J.F.; Manolios, N.; Emery, P.; Conaghan, P.G.; Tennant, A. Depression and anxiety in patients with rheumatoid arthritis: Prevalence rates based on a comparison of the Depression, Anxiety and Stress Scale (DASS) and the hospital, Anxiety and Depression Scale (HADS). BMC Psychiatry 2012, 12, 6. [Google Scholar] [CrossRef] [PubMed]
  5. Gomaa, M.A.M.; Elmagd, M.H.A.; Elbadry, M.M.; Kader, R.M.A. Depression, Anxiety and Stress Scale in patients with tinnitus and hearing loss. Eur. Arch. Oto Rhino Laryngol. 2014, 271, 2177–2184. [Google Scholar] [CrossRef] [PubMed]
  6. Grailet, J.; Deckers, S.; Fraipont, V.; Etienne, A.; Damas, F. Stress, anxiety and depression in cardial infarct and bypass surgery patients. Crit. Care 2005, 9, 242. [Google Scholar] [CrossRef]
  7. Li, M.; Xie, L.; Wang, Z. A Transductive Model-based Stress Recognition Method Using Peripheral Physiological Signals. Sensors 2019, 19, 429. [Google Scholar] [CrossRef] [PubMed]
  8. Bitkina, O.VI.; Kim, J.; Park, J.; Park, J.; Kim, H.K. Identifying Traffic context using Driving Stress: A Longitudinal Preliminary Case Study. Sensors 2019, 19, 2152. [Google Scholar] [CrossRef] [PubMed]
  9. Liao, S.C.; Wu, C.T.; Huang, H.C.; Cheng, W.T.; Liu, Y.H. Major Depression Detection from EEG Signals Using Kernel Eigen-Filter-Bank Common Spatial Patterns. Sensors 2017, 17, 1385. [Google Scholar] [CrossRef]
  10. Ahn, J.W.; Ku, Y.; Kim, H.C. A Novel Wearable EEG and ECG Recording System for Stress Assessment. Sensors 2019, 19, 1991. [Google Scholar] [CrossRef]
  11. Blanco, J.A.; Vanleer, A.C.; Calibo, T.K.; Firebaugh, S.L. Single-trial cognitive stress classification using portable wireless electroencephalography. Sensors 2019, 19, 499. [Google Scholar] [CrossRef]
  12. Sharma, N. Thermal spatio-temporal data for stress recognition. EURASIP J. Image Video Process. 2014, 28, 1–12. [Google Scholar] [CrossRef]
  13. Iran, R.; Nasrollahi, K.; Dhall, A.; Moeslund, T.B.; Gedeon, T. Thermal super-pixels for bimodal stress recognition. In Proceedings of the 2016 Sixth International Conference on Image Processing Theory, Tools and Applications (IPTA), Oulu, Finland, 12–15 December 2016. [Google Scholar]
  14. Pampouchidou, A. Automated characterization of mouth activity for stress and anxiety assessment. In Proceedings of the 2016 IEEE International Conference on Imaging Systems and Techniques (IST), Chania, Greece, 4–6 October 2016. [Google Scholar]
  15. Prasetio, B.H.; Tamura, H.; Tanno, K. The Facial Stress Recognition Based on Multi-histogram Features and Convolutional Neural Network. In Proceedings of the IEEE International Conference on Systems, Man, and Cybernetics, Miyazaki, Japan, 7–10 October 2018. [Google Scholar]
  16. Phillips, P.J.; Wechsler, H.; Huang, J.; Rauss, P.J. The FERET database and evaluation procedure for face-recognition algorithms. Image Vis. Comput. 1998, 16, 295–306. [Google Scholar] [CrossRef]
  17. Pampouchidou, A.; Marias, K.; Tsiknakis, M.; Simos, P.; Yang, F.; Meriaudeau, F. Designing a framework for assisting depression severity assessment from facial image analysis. In Proceedings of the 2015 IEEE International Conference on Signal and Image Processing Applications (ICSIPA), Kuala Lumpur, Malaysia, 19–21 October 2015. [Google Scholar]
  18. Dhall, A.; Goecke, R. A temporally piece-wise fisher vector approach for depression analysis. In Proceedings of the 2015 International Conference on Affective Computing and Intelligent Interaction (ACII), Xi’an, China, 21–21 September 2015. [Google Scholar]
  19. Valstar, M.; Schuller, B.; Smith, K.; Almaev, T.; Eyben, F.; Krajewski, J.; Cowie, R.; Pantic, M. AVEC 2014: 3D dimensional affect and depression recognition challenge. In Proceedings of the 4th International Workshop on Audio/Visual Emotion Challenge (AVEC), Orlando, FL, USA, 7 November 2014. [Google Scholar]
  20. Joshi, J.; Dhall, A.; Goecke, R.; Breakspear, M.; Parker, G. Neural-net classification for spatio-temporal descriptor based depression analysis. In Proceedings of the 21st International Conference on Pattern Recognition (ICPR2012), Tsukuba, Japan, 11–15 November 2012. [Google Scholar]
  21. Harati, S.; Crowell, A.; Mayberg, H.; Kong, J.; Nemati, S. Discriminating clinical phases of recovery from major depressive disorder using the dynamics of facial expression. In Proceedings of the 38th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), Orlando, FL, USA, 17–20 August 2016. [Google Scholar]
  22. Yang, L.; Jiang, D.; Han, W.; Sahli, H. DCNN and DNN based multi-modal depression recognition. In Proceedings of the 7th International Conference on Affective Computing and Intelligent Interaction (ACII), San Antonio, TX, USA, 23–26 October 2017. [Google Scholar]
  23. De Melo, W.C.; Granger, E.; Hadid, A. Combining Global and Local Convolutional 3D Networks for Detecting Depression from Facial Expressions. In Proceedings of the 14th International Conference on Automatic Face & Gesture Recognition, Lile, France, 14–18 May 2019. [Google Scholar]
  24. Zhu, Y.; Shang, Y.; Shao, Z.; Guo, G. Automated Depression diagnosis based on Deep Networks to Encode Facial Appearance and Dynamics. IEEE Trans. Affect. Comput. 2018, 9, 578–584. [Google Scholar] [CrossRef]
  25. Yang, L.; Jiang, D.; Sahli, H. Integrating Deep and Shallow Models for Multi-Modal Depression Analysis—Hybrid Architectures. IEEE Trans. Affect. Comput. 2018. [Google Scholar] [CrossRef]
  26. Zhou, X.; Jin, K.; Shang, Y.; Guo, G. Visually Interpretable Representation Learning for Depression Recognition from Facial Images. IEEE Trans. Affect. Comput. 2018. [Google Scholar] [CrossRef]
  27. Huang, X.; Chen, D.; Huang, Y.; Han, X.; Chen, Y.W. Automatic prediction of trait anxiety degree using recognition rates of facial emotions. In Proceedings of the Sixth International Conference on Advanced Computational Intelligence (ICACI), Hangzhou, China, 19–21 October 2013. [Google Scholar]
  28. Huang, F.; Wen, W.; Liu, G. Facial Expression Recognition of Public Speaking Anxiety. In Proceedings of the 9th International Symposium on Computational Intelligence and Design (ISCID), Hangzhou, China, 10–11 December 2016. [Google Scholar]
  29. Pediaditis, M.; Giannakakis, G.; Chiarugi, F.; Manousos, D.; Pampouchidou, A.; Christanaki, E.; Iatraki, G.; Kazantzaki, E.; Simos, P.G.; Marias, K.; et al. Extraction of Facial features as indicators of stress and anxiety. In Proceedings of the Annual International Conference of the IEEE Engineering in Medicine and Biology Society, Milan, Italy, 25–29 August 2015. [Google Scholar]
  30. McIntyre, G.; Göcke, R.; Hyett, M.; Green, M.; Breakspear, M. An approach for automatically measuring facial activity in depressed subjects. In Proceedings of the 2009 3rd International Conference on Affective Computing and Intelligent Interaction and Workshops, Amsterdam, The Netherlands, 10–12 September 2009. [Google Scholar]
  31. Girard, J.M.; Cohn, J.F.; Mahoor, M.H.; Mavadati, S.; Rosenwald, D.P. Social risk and depression: Evidence from manual and automatic facial expression analysis. In Proceedings of the 10th IEEE International Conference and Workshops on Automatic Face and Gesture Recognition (FG), Shanghai, China, 22–26 April 2013. [Google Scholar]
  32. Szwoch, M. Aquisition and indexing of RGB-D recordings for facial expressions recognition. Stud. Inform. 2015, 36, 7–18. [Google Scholar]
  33. Szwock, M.; Pieniazek, P. Facial emotion recognition using depth data. In Proceedings of the 8th International Conference on Human System Interaction (HSI), Sopot, Poland, 6–8 June 2015. [Google Scholar]
  34. Ballihi, L.; Lablack, A.; Ben Amor, B.; Bilasco, I.M.; Daoudi, M. Positive/Negative emotion detection from RGB-D Upper Body Images. In Proceedings of the International Workshop on Face and Facial Expression Recognition from Real World Videos, Stockholm, Sweden, 24 August 2014. [Google Scholar]
  35. Mahmoud, M.; Baltrusaitis, T.; Robinson, P.; Riek, L.D. 3D Corpus of Spontaneous Complex Mental States. In Proceedings of the International Conference on Affective Computing and Intelligent Interaction, Memphis, TN, USA, 9–12 October 2011. [Google Scholar]
  36. Jaiswal, S.; Valstar, M.F.; Gillot, A.; Daley, D. Automatic detection of ADHD and ASD from expressive behaviour in RGBD data. In Proceedings of the 12th IEEE International Conference on Face and Gesture Recognition, Washington, DC, USA, 30 May–3 June 2017. [Google Scholar]
  37. Lovibond, S.H.; Lovibond, P.F. Manual for the Depression Anxiety Stress Scales; Psychology Foundation of Australia: Sydney, Australia, 1995. [Google Scholar]
  38. Eckman, P.; Friesen, W.V. Facial Action Coding System: Investigator’s Guide; Consulting Psychologists Press: Palo Alto, CA, USA, 1978. [Google Scholar]
  39. Zbancioc, M.D.; Feraru, S.M. A study about the automatic recognition of the anxiety emotional state using Emo-DB. In Proceedings of the 2015 E-Health and Bioengineering Conference (EHB), Iasi, Romania, 19–21 November 2015. [Google Scholar]
  40. Suveg, C.; Hoffman, B.; Janice, L.Z.; Thomassin, K. Common and specific emotion-related predictors of anxious and depressive symptoms in youth. Child Psychiatry Hum. Dev. 2009, 40, 223. [Google Scholar] [CrossRef]
  41. Feldman, P.J.; Cohen, S.; Lepore, S.J.; Matthews, K.A.; Kamarck, T.W.; Marsland, A.L. Negative emotions and acute physiological responses to stress. Ann. Behav. Med. 1999, 21, 216–222. [Google Scholar] [CrossRef]
  42. Cavanagh, C.E.; Larkin, K.T. A critical review of the “Undoing Hypothesis”: Do Positive Emotions Undo the Effects of Stress? Appl. Psychophysiol. Biofeedback 2018, 43, 259–273. [Google Scholar] [CrossRef]
  43. Ekman, P.; Friesen, W.V.; Hager, J.C. Facial Action Coding System. In Manual and Investigator’s Guide; Research Nexus: Salt Lake City, UT, USA, 2002. [Google Scholar]
  44. Kim, J.; Kima, B.; Roy, P.P.; Jeong, D. Efficient Facial Expression Recognition Algorithm Based on Hierarchical Deep Neural Network Structure. IEEE Access 2019, 7, 41273–41285. [Google Scholar] [CrossRef]
  45. Pham, T.T.D.; Kim, S.; Lu, Y.; Jung, S.; Won, C. Facial Action Units-Based Image Retrieval for Facial Expression Recognition. IEEE Access 2019, 7, 5200–5207. [Google Scholar] [CrossRef]
  46. Kherchaoui, S.; Houacine, A. Face detection based on a model of the skin color with constraints and template matching. In Proceedings of the International Conference on Machine and Web Intelligence, Algiers, Algeria, 3–5 October 2010. [Google Scholar]
  47. Budiman, I.; Herlianto; Suhartono, D.; Purnomo, F.; Shodiq, M. The effective noise removal techniques and illumination effect in face recognition using Gabor and Non-Negative Matrix Factorization. In Proceedings of the International Conference on Informatics and Computing, Shanghai, China, 23–25 October 2016. [Google Scholar]
  48. Burton, A.M.; Bindemann, K. The role of view in human face detection. Vis. Res. 2009, 49, 2026–2036. [Google Scholar] [CrossRef] [PubMed]
  49. Batista, J.P. Locating Facial Features Using an Anthropometric Face model for Determining the Gaze of Faces in Image Sequences. In Proceedings of the International Conference on Image Analysis and Recognition, Montreal, QC, Canada, 5–7 August 2007. [Google Scholar]
  50. Edwards, G.J.; Taylor, C.J.; Cootes, T.F. Interpreting face images using active appearance models. In Proceedings of the 3rd IEEE International Conference on Automatic Face and Gesture Recognition, Nara, Japan, 14–16 April 1998. [Google Scholar]
  51. Dornaika, F.; Ahlberg, J. Efficient active appearance model for real-time head and facial feature tracking. In Proceedings of the IEEE International SOI Conference, Newport Beach, CA, USA, 14–17 October 2003. [Google Scholar]
  52. Tzimiroupoulos, G.; Pantic, M. Fast Algorithms for Fitting Active Appearance Modes to Unconstrained Images. Int. J. Comput. Vis. 2017, 122, 17–33. [Google Scholar] [CrossRef]
  53. van der Maaten, L.; Hendriks, E. Action Unit classification using active appearance models and conditional random fields. Cogn. Process. 2012, 13, 507–518. [Google Scholar] [CrossRef] [PubMed]
  54. Komalasari, D.; Widyanto, M.R.; Basaruddin, T.; Liliana, D.Y. Shape analysis using generalized Procrustes analysis on Active Appearance Model for facial expression recognition. In Proceedings of the International Conference on Electrical Engineering and Computer Science, Palembag, Indonesia, 16 August 2017. [Google Scholar]
  55. Liao, C.T.; Wu, Y.K.; Lai, S.H. Locating facial feature points using support vector machines. In Proceedings of the 9th International Workshop on Cellular Neural Networks and Their Applications, Hsinchu, Taiwan, 28–30 May 2005. [Google Scholar]
  56. Gavrilescu, M. Proposed architecture of a fully integrated modular neural network-based automatic facial emotion recognition system based on Facial Action Coding System. In Proceedings of the International Conference on Communications (COMM), Bucharest, Romania, 29–31 May 2014. [Google Scholar]
  57. Lucey, P.; Cohn, J.F.; Kanade, T.; Saragih, J.; Ambadar, Z.; Matthews, I. The Extended Cohn-Kanade Dataset (CK+): A complete dataset for action unit and emotion-specified expression. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition—Workshops, San Francisco, CA, USA, 13–18 June 2010. [Google Scholar]
  58. Pantic, M.; Valstar, M.; Rademaker, R.; Maat, L. Web-based database for facial expression analysis. In Proceedings of the IEEE International Conference on Multimedia and Expo, Amsterdam, The Netherlands, 6–8 July 2005. [Google Scholar]
  59. Lyons, M.J.; Akamatsu, S.; Kamachi, M.; Gyoba, J. Coding Facial Expressions with Gabor Wavelets. In Proceedings of the 3rd IEEE International Conference on Automatic Face and Gesture Recognition, Nara, Japan, 14–16 April 1998. [Google Scholar]
  60. Sneddon, I.; McRorie, M.; McKeown, G.; Hanratty, J. The Belfast Induced Natural Emotion Database. IEEE Trans. Affect. Comput. 2012, 3, 32–41. [Google Scholar] [CrossRef]
  61. Kossaifi, J.; Tzimiropoulos, G.; Todorovic, S.; Pantic, M. AFEW-VA database for valence and arousal estimation in-the-wild. Image Vis. Comput. 2017, 65, 23–36. [Google Scholar] [CrossRef]
  62. Dhall, A.; Goecke, R.; Lucey, S.; Gedeon, T. Collecting Large, Richly Annotated Facial-Expression Databases from Movies. IEEE MultiMedia 2012, 19, 34–41. [Google Scholar] [CrossRef] [Green Version]
  63. Mohammadpour, M.; Khaliliardali, H.; Hashemi, S.M.R.; AlyanNezhadi, M.M. Facial emotion recognition using deep convolutional networks. In Proceedings of the 4th IEEE International Conference on Knowledge-Based Engineering and Inovation, Teheran, Iran, 22 December 2017. [Google Scholar]
  64. Ghosh, S.; Lakshana, E.; Scherer, S.; Morency, L.P. A multi-label convolutional neural network approach to cross-domain action unit detection. In Proceedings of the 2015 International Conference on Affective Computing and Intelligent Interaction, Xi’an, China, 21–24 September 2015. [Google Scholar]
  65. Pham, T.T.D.; Won, C.S. Facial Action Units for Training Convolutional Neural Networks. IEEE Access 2019, 7, 77816–77824. [Google Scholar] [CrossRef]
  66. Buhmann, M.D. Radial Basis Functions: Theory and Implementations; Cambridge University: Cambridge, UK, 2003. [Google Scholar]
  67. Miljanovic, M. Comparative analysis of Recurrent and Finite Impulse Response Neural Networks in Time Series Prediction. Indian J. Comput. Eng. 2012, 3, 1. [Google Scholar]
  68. Saxen, H.; Petterson, F. A methodology for developing Nonlinear Models by Feedforward Neural Networks. In Proceedings of the International Conference on Adaptive and Natura Computing Algorithms, Ljubljana, Slovenia, 14–16 April 2009. [Google Scholar]
  69. Huang, G.B. Learning capability and storage capacity of two-hidden-layer feed-forward networks. IEEE Trans. Neural Netw. 2003, 14, 274–281. [Google Scholar] [CrossRef]
  70. Hara, K.; Saito, D.; Shouno, H. Analysis of function of rectified linear unit used in deep learning. In Proceedings of the International Joint Conference on Neural Networks, Killarney, Ireland, 12–16 July 2015. [Google Scholar]
  71. Bridle, J.S. Probabilistic Interpretation of Feedforward Classification Network Outputs, with Relationships to Statistical Pattern Recognition; Springer: New York, NY, USA, 1990; pp. 227–236. [Google Scholar]
  72. Li, J.; Cheng, J.H.; Huang, F. Brief Introduction of Back Propagation (BP) Neural Network Algorithm and Its Improvement. In Advances in Computer Science and Information Engineering; Springer: New York, NY, USA, 2012; pp. 553–558. [Google Scholar]
  73. Masood, S.; Doja, M.N.; Chandra, P. Analysis of weight initialization techniques for gradient descent. In Proceedings of the Annual IEEE India Conference, New Delhi, India, 17–20 December 2015. [Google Scholar]
  74. World Medical Association, World Medical Association Declaration of Helsinki: Ethical principles for medical research involving human subjects. JAMA 2013, 310, 2191–2194. [CrossRef] [PubMed]
  75. Harmon-Jones, C.; Bastian, B.; Harmon-Jones, E. The Discrete Emotions questionnaire: A New Tool for Measuring State Self-Reported Emotions. PloS One 2016, 11, e0159915. [Google Scholar] [CrossRef] [PubMed]
  76. Pampuchidou, A.; Marias, K.; Tsiknakis, M.; Simos, P.; Yang, F.; Lemaitre, G.; Meriaudeau, F. Video-based depression detection using local curvelet binary patterns in pairwise orthogonal planes. In Proceedings of the Annual International Conference of the IEEE Engineering in Medicine and Biology Society, Orlando, FL, USA, 17–20 August 2016. [Google Scholar]
  77. Simundic, A.M. Measures of Diagnostic Accuracy: Basic Definitions. Electr. J. Int. Federation Clin. Chem. Lab. Med. 2009, 19, 203–211. [Google Scholar]
  78. Bianco, S.; Cadene, R.; Celona, L.; Napoletano, P. Benchmark Analysis of Representative Deep Neural Network Architectures. IEEE Access 2018, 6, 64270–64277. [Google Scholar] [CrossRef]
Figure 1. Facial landmarks for AAM algorithm and different facial expressions.
Figure 1. Facial landmarks for AAM algorithm and different facial expressions.
Sensors 19 03693 g001
Figure 2. AU classification accuracy for cross-database tests.
Figure 2. AU classification accuracy for cross-database tests.
Sensors 19 03693 g002
Figure 3. Comparison of state-of-the-art methods for AU classification. (a) Average classification accuracy [%], (b) Average classification time [miliseconds].
Figure 3. Comparison of state-of-the-art methods for AU classification. (a) Average classification accuracy [%], (b) Average classification time [miliseconds].
Sensors 19 03693 g003
Figure 4. Overall architecture for predicting DASS levels by analyzing facial features using FACS.
Figure 4. Overall architecture for predicting DASS levels by analyzing facial features using FACS.
Sensors 19 03693 g004
Figure 5. Prediction accuracy using different combinations of neural networks. T: tanh; S: sigmoid; R: ReLU; SMAX: softmax; 1L: one hidden layer; 2L: two hidden layers; : common FFNN for all three emotional states; : dedicated FFNN for each emotional state.
Figure 5. Prediction accuracy using different combinations of neural networks. T: tanh; S: sigmoid; R: ReLU; SMAX: softmax; 1L: one hidden layer; 2L: two hidden layers; : common FFNN for all three emotional states; : dedicated FFNN for each emotional state.
Sensors 19 03693 g005
Figure 6. Face Depression Anxiety Stress Scale Neural Network (FDASSNN) configuration and hyperparameters.
Figure 6. Face Depression Anxiety Stress Scale Neural Network (FDASSNN) configuration and hyperparameters.
Sensors 19 03693 g006
Figure 7. Prediction accuracy based on the number of neurons in hidden layers.
Figure 7. Prediction accuracy based on the number of neurons in hidden layers.
Sensors 19 03693 g007
Figure 8. User interface for the application designed to predict DASS levels based on FACS.
Figure 8. User interface for the application designed to predict DASS levels based on FACS.
Sensors 19 03693 g008
Figure 9. DASS levels distribution.
Figure 9. DASS levels distribution.
Sensors 19 03693 g009
Figure 10. DASS levels prediction accuracy (intrasubject methodology).
Figure 10. DASS levels prediction accuracy (intrasubject methodology).
Sensors 19 03693 g010
Figure 11. Prediction accuracy based on the number of seconds analyzed from the test video samples (intrasubject methodology).
Figure 11. Prediction accuracy based on the number of seconds analyzed from the test video samples (intrasubject methodology).
Sensors 19 03693 g011
Figure 12. Prediction accuracy for major depressive disorder (MDD), generalized anxiety disorder (GAD), and post-traumatic stress disorder (PTSD) (intrasubject methodology).
Figure 12. Prediction accuracy for major depressive disorder (MDD), generalized anxiety disorder (GAD), and post-traumatic stress disorder (PTSD) (intrasubject methodology).
Sensors 19 03693 g012
Figure 13. DASS levels prediction accuracy (intersubject methodology).
Figure 13. DASS levels prediction accuracy (intersubject methodology).
Sensors 19 03693 g013
Figure 14. Prediction accuracy based on the number of seconds analyzed from the test video samples (intersubject methodology).
Figure 14. Prediction accuracy based on the number of seconds analyzed from the test video samples (intersubject methodology).
Sensors 19 03693 g014
Figure 15. Prediction accuracy for MDD, GAD, and PTSD (intersubject methodology).
Figure 15. Prediction accuracy for MDD, GAD, and PTSD (intersubject methodology).
Sensors 19 03693 g015
Figure 16. Prediction accuracy based on the number of subjects used for training (intersubject methodology).
Figure 16. Prediction accuracy based on the number of subjects used for training (intersubject methodology).
Sensors 19 03693 g016
Figure 17. Stress prediction comparison to state-of-the-art methods (ANUStressDB [12]). (a) Average accuracy, sensitivity, and specificity, (b) Average prediction time (seconds).
Figure 17. Stress prediction comparison to state-of-the-art methods (ANUStressDB [12]). (a) Average accuracy, sensitivity, and specificity, (b) Average prediction time (seconds).
Sensors 19 03693 g017
Figure 18. Depression prediction comparison to state-of-the-art methods (AVEC2014 [19]). (a) Average accuracy, sensitivity, and specificity, (b) Average prediction time (seconds).
Figure 18. Depression prediction comparison to state-of-the-art methods (AVEC2014 [19]). (a) Average accuracy, sensitivity, and specificity, (b) Average prediction time (seconds).
Sensors 19 03693 g018
Figure 19. Depression prediction comparison to state-of-the-art methods (own database). (a) Average accuracy, sensitivity, and specificity, (b) Average prediction time (seconds).
Figure 19. Depression prediction comparison to state-of-the-art methods (own database). (a) Average accuracy, sensitivity, and specificity, (b) Average prediction time (seconds).
Sensors 19 03693 g019
Figure 20. Anxiety prediction comparison to state-of-the-art methods (own database). (a) Average accuracy, sensitivity, and specificity, (b) Average prediction time (seconds).
Figure 20. Anxiety prediction comparison to state-of-the-art methods (own database). (a) Average accuracy, sensitivity, and specificity, (b) Average prediction time (seconds).
Sensors 19 03693 g020
Figure 21. Stress prediction comparison to state-of-the-art methods (own database). (a) Average accuracy, sensitivity, and specificity, (b) Average prediction time (seconds).
Figure 21. Stress prediction comparison to state-of-the-art methods (own database). (a) Average accuracy, sensitivity, and specificity, (b) Average prediction time (seconds).
Sensors 19 03693 g021
Table 1. DASS levels prediction accuracy, sensitivity, specificity, and average duration of analyzed video sample in intrasubject methodology.
Table 1. DASS levels prediction accuracy, sensitivity, specificity, and average duration of analyzed video sample in intrasubject methodology.
DSTRDSCDSCDSRDSRDSC + DSRLMM
DSTSDSCDSRDSRDSCDSC + DSR
MMACSESPACSESPACSESPACSESPACSESP
Depression(1)91.1919190.489.789.982.181.681.885.284.484.986.385.786(2)
(2)80.279.979.978.5787871.971.771.674.673.774.47675.575.7(3)
(3)82.181.181.781.180.780.872.771.872.375.975.375.57776.876.6(1)
(4)9493.993.892.891.992.684.98484.787.586.88788.687.988.1(3)
(5)95.293.994.393.292.592.886.185.885.690.390.19091.791.291.3(4)
Avg.89.388.188.387.286.686.979.678.979.182.782.482.383.983.183.6
ADVS1212171614
Anxiety(1)85.184.884.784.183.583.774.974.674.778.377.377.879.579.279.1(2)
(2)787878.277.27776.767.967.567.771.770.971.372.972.772.6(1)
(3)72.271.471.670.870.670.36362.162.86665.865.667.166.466.9(4)
(4)76.174.975.474.773.974.364.163.263.867.166.666.968.468.168.1(5)
(5)84.183.783.882.982.582.472.872.172.575.575.375.276.975.976.6(4)
Avg.79.278.478.677.97777.768.667.968.371.771.571.47372.272.6
ADVS1414191716
Stress(1)95.194.994.994.293.293.887.186.286.789.588.789.190.590.290.1(2)
(2)89.288.888.98887.887.677.776.877.58180.580.682.481.582(3)
(3)89.188.488.988.387.688.180.880.380.582.98282.584.183.883.9(2)
(4)89.287.988.387.186.186.877.476.577.181.781.181.58382.182.8(5)
(5)959494.593.593.193.383.983.283.688.287.78889.688.689.1(4)
Avg.92.190.991.190.289.789.881.481.281.184.783.884.385.985.785.7
ADVS1414191716
AC: accuracy; SE: sensitivity; SP: specificity; Avg: average; (1): normal level; (2): mild level; (3): moderate level; (4): severe level; (5): extremely severe level; DSTR: dataset training; DSTS: dataset testing; ADVS: average duration of analyzed video sample (seconds); MM: measure method; LMM: DASS level most mistaken for.
Table 2. DASS levels prediction accuracy, sensitivity, specificity, and average duration of analyzed video samples in intersubject methodology.
Table 2. DASS levels prediction accuracy, sensitivity, specificity, and average duration of analyzed video samples in intersubject methodology.
DSTRDSCDSCDSRDSRDSC + DSRLMM
DSTEDSCDSRDSRDSCDSC + DSR
MMACSESPACSESPACSESPACSESPACSESP
Depression(1)81.278.079.880.176.978.768.165.467.067.365.965.672.270.070.5(2)
(2)71.970.670.170.668.368.852.950.051.053.150.151.357.155.955.4(1)
(3)73.871.372.172.671.471.355.552.253.655.453.554.360.357.858.8(2)
(4)85.182.783.283.982.382.973.870.272.773.170.471.177.774.875.9(3)
(5)87.183.285.285.984.584.475.073.773.975.772.674.779.877.678.6(4)
Avg.79.877.278.178.676.777.265.162.363.664.962.563.469.467.267.8
ADVS1116181413
Anxiety(1)75.272.073.774.071.272.165.661.663.864.961.163.069.467.868.1(2)
(2)68.565.567.367.064.665.750.349.249.049.946.248.454.852.652.8(3)
(3)63.060.961.461.658.860.553.051.851.253.749.952.658.455.557.3(2)
(4)64.263.162.362.760.860.950.749.449.350.347.049.154.651.353.3(3)
(5)72.470.771.471.369.069.455.152.253.255.152.454.060.157.458.4(4)
Avg.68.766.467.267.364.965.754.952.853.354.851.353.459.556.958.0
ADVS1216201414
Stress(1)86.083.284.885.081.783.574.170.872.274.471.772.879.478.377.5(2)
(2)78.876.576.877.673.675.960.456.459.060.858.259.665.464.163.8(3)
(3)80.077.978.178.875.077.671.069.969.070.869.269.275.172.473.4(4)
(4)78.276.576.377.075.175.369.066.767.668.665.867.173.070.571.6(3)
(5)84.581.382.983.582.281.675.473.874.075.171.273.479.978.278.4(4)
Avg.81.579.179.880.477.578.770.067.568.470.067.268.474.672.773.0
ADVS1215181414
AC: accuracy; SE: sensitivity; SP: specificity; Avg.: average; (1): normal level; (2): mild level; (3): moderate level; (4): severe level; (5): extremely severe level; DSTR: dataset training; DSTS: dataset testing; DAVS: average duration of analyzed video sample (seconds); MM: measure method; LMM: DASS level most mistaken for.
Table 3. Correlations between FACS and DASS levels.
Table 3. Correlations between FACS and DASS levels.
Emotional StateAU (Intensity Level D or E)Accuracy (%)
DepressionAU6, AU12, AU15, AU2684.51
AnxietyAU2, AU9, AU25, AU4572.92
StressAU1, AU6, AU12, AU1588.41
Table 4. Correlations between emotions and DASS levels.
Table 4. Correlations between emotions and DASS levels.
Emotional StateInduced EmotionAccuracy (%)
DepressionHappiness, sadness87.71
AnxietySurprise, disgust82.13
StressSadness, disgust93.21

Share and Cite

MDPI and ACS Style

Gavrilescu, M.; Vizireanu, N. Predicting Depression, Anxiety, and Stress Levels from Videos Using the Facial Action Coding System. Sensors 2019, 19, 3693. https://doi.org/10.3390/s19173693

AMA Style

Gavrilescu M, Vizireanu N. Predicting Depression, Anxiety, and Stress Levels from Videos Using the Facial Action Coding System. Sensors. 2019; 19(17):3693. https://doi.org/10.3390/s19173693

Chicago/Turabian Style

Gavrilescu, Mihai, and Nicolae Vizireanu. 2019. "Predicting Depression, Anxiety, and Stress Levels from Videos Using the Facial Action Coding System" Sensors 19, no. 17: 3693. https://doi.org/10.3390/s19173693

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop