Next Article in Journal
Identification of Reference Buildings in Mediterranean Countries: The HAPPEN Project Approach
Next Article in Special Issue
Learning Robust Shape-Indexed Features for Facial Landmark Detection
Previous Article in Journal
An Application of Android Sensors and Google Earth in Pavement Maintenance Management Systems for Developing Countries
Previous Article in Special Issue
Hybrid Approach for Facial Expression Recognition Using Convolutional Neural Networks and SVM
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

An Approach for Selecting the Most Explanatory Features for Facial Expression Recognition

by
Pedro D. Marrero-Fernandez
1,
Jose M. Buades-Rubio
2,*,
Antoni Jaume-i-Capó
2,3 and
Tsang Ing Ren
1
1
Centro de Informática, Universidade Federal de Pernambuco (UFPE), Recife 50670-901, Brazil
2
Computer Graphics, Vision and Artificial Intelligence Group (UGIVIA), Department of Mathematics and Computer Science, University of the Balearic Islands, 07122 Palma, Spain
3
Laboratory of Artificial Intelligence Applications (LAIA@UIB), University of the Balearic Islands, 07122 Palma, Spain
*
Author to whom correspondence should be addressed.
Appl. Sci. 2022, 12(11), 5637; https://doi.org/10.3390/app12115637
Submission received: 18 March 2022 / Revised: 27 May 2022 / Accepted: 28 May 2022 / Published: 1 June 2022
(This article belongs to the Special Issue Research on Facial Expression Recognition)

Abstract

:
The objective of this work is to analyze which features are most important in the recognition of facial expressions. To achieve this, we built a facial expression recognition system that learns from a controlled capture data set. The system uses different representations and combines them from a learned model. We studied the most important features by applying different feature extraction methods for facial expression representation, transforming each obtained representation into a sparse representation (SR) domain, and trained combination models to classify signals, using the extended Cohn–Kanade (CK+), BU-3DFE, and JAFFE data sets for validation. We compared 14 combination methods for 247 possible combinations of eight different feature spaces and obtained the most explanatory features for each facial expression. The results indicate that the LPQ (83%), HOG (82%), and RAW (82%) features are those features most able to improve the classification of expressions and that some features apply specifically to one expression (e.g., RAW for neutral, LPQ for angry and happy, LBP for disgust, and HOG for surprise).

1. Introduction

Computers are quickly becoming a ubiquitous part of our lives, and we spend a large amount of time interacting with computers of one type or another; however, the devices that we use are indifferent to our affective states, as they are emotionally blind. Successful human–human communication relies on the ability to read affective and emotional signals. Human–computer interaction, which does not consider the affective states of its users, loses a large part of the information available from the interaction.
At present, expression recognition can be achieved using neural networks. A neural network behaves like a black-box model; therefore, it is difficult to understand how it performs its classifications. Recent works have attempted to create attention maps to understand the results obtained by neural networks [1,2]; however, these maps do not produce reasonable interpretations of the types of features learned by the neural network.
Other solutions, such as sparse representation, help in studying the classification performance. In this context, a trustworthy intelligent system must be able to explain its decisions and actions to human users through the use of techniques that produce more understandable models, while maintaining high performance levels [3].
Thus, in this paper, we analyze the importance of each feature used for facial expression recognition. Feature-relevance techniques seem to be some of the most-used schemes for post hoc explanation [4,5], as they can provide an explicit description of the inner behavior of the model, contributing to the goal of designing an explainable intelligent system.
In our work, we focus on determining which feature sub-spaces are the most advantageous, where sparse representation (SR) was used to determine the features.
The remainder of this paper is organized as follows: Section 2 introduces the related work. Section 3 describes the methodology, local and geometric features, expression classification, and combination framework. Section 4 describes the specific experimental process. Section 5 analyzes and discusses the results. The research is summarized, and our conclusions are offered in Section 6.

2. Related Work

Guidotti et al. [6] carried out an exhaustive literature review to explain learning systems. In [7], Letham et al. generated an explanation using rules and Bayesian analysis by making if/else decisions. LIME was proposed as a solution in [8], with the aim of explaining the predictions of any model in an interpretable manner.
Chandrashekar et al. [9] provided an introduction to feature-selection techniques. They concluded that “more information is not always good in machine learning applications”. This is an important issue, as it means that adding more features does not always elicit better results. In [10], Weitz et al. used LIME and layer-wise relevance propagation (LRP) to explain how a neural network can distinguish between pain, happiness, and disgust.
In [11], Gund et al. used videos from CK+ and a temporal convolutional network to determine the most important facial landmarks (i.e., 68 OpenFace landmarks) that can be used to infer facial expressions. In [12], Ter Burg concluded that human participants found the explanations for the geometric feature-based DNN better overall than the Grad-CAM explanations for the CNN. In Wang et al., M2Lens [13] contributes to visualizing and explaining multi-modal models for sentiment analysis, which is related to facial expression recognition.
Lian et al. [14] and Kim et al. [15] determined the regions used for the classification of individual expressions. Deramgozin et al. [16] determined how a CNN classifies different expressions, both locally and globally. However, none of the reviewed works have determined which characteristics are the most important for better facial expression recognition. Zhu et al., in their recent publications, have attempted to focus the majority of their processing on a target region of interest, using a so-called visual attention mechanism [1,2]. Although the existing attention methods have contributed greatly to facial expression recognition, according to Bonnard et al. [2], there is a problem associated with the insufficient utilization of spatial features. In our opinion, there is a need to begin to understand the different types of features that contribute to face-recognition analysis, which could ultimately improve the actual systems in practice.
Several papers related to feature extraction for facial expression recognition (FER) can be found in the literature. Ying et al. [17] used local binary patterns (LBP) and Image Raw (RAW) to train two classifiers using a sparse representation-based classifier (SRC). For each of these schemes (LBP + SRC and RAW + SRC), the approximation error signal was obtained for each class. This error was used as a fuzzy measure for the evaluation of a decision rule. The residual ratio was calculated as the ratio between the second smallest residual and the smallest residual for the LBP + SRC and RAW + SRC methods. If the results for the two classifiers were not the same, the classification with the larger residual ratio was chosen. In the study by Li et al. [18], the classifiers were trained using local phase quantization (LPQ) and Gabor wavelets + Adaboost (GW). The Adaboost algorithm was used to select the most effective 100 features from each Gabor filter. As reported in [17], the classification result with the larger residual ratio was chosen if the classification using the two classifiers differed [17]. Ouyang [19] used a histogram of oriented gradients (HOG) and LBP. This approach was based on the idea that features are complementary, as HOG mainly extracts a contour-based shape, while LBP primarily extracts the texture information of the images. Kittler et al. [20] showed that the output of each classifier could be used to evaluate a decision rule, then they applied combination rules. Of these, the product rule (PR) and the SR provided the best results. Ptucha et al., Ji et al., and Tsalakanidou et al. have also employed dynamic features in [21,22,23]. In these works, the variability of facial changes was studied using regions or points of interest in the face images.
The notion of SRs—finding sparse solutions to under-determined systems—has been applied in several scientific fields. Olshausen et al. used sparse models that are similar in nature to the network of neurons in V1—the first layer of the visual cortex in the human brain—and, more generally, to the mammalian brain [24,25]. Patterns of light are represented by a series of innate or learned basis functions, whereby sparse linear combinations form surrogate input stimuli to the brain. Similarly, for many input signals of interest, such as natural images, a small number of exemplars can form a surrogate representation of a new test image.
In SR systems, new test images are efficiently represented by sparse linear coefficients on a dictionary D of over-complete basis functions. Specifically, SR systems comprise an input sample x R m (where m is the number of features), along with a dictionary D of n samples, D R m × n . SR solves the coefficients α R n that satisfy the l 1 minimization problem x = D .
Wright et al., Weifeng et al., and Zhen et al. showed the advantages of exploiting sparsity in pattern classification, which has been demonstrated extensively for FER problems [26,27,28]. The experimental results of [26] showed that the magnitude of the representation error in the facial feature vectors obtained by SR provided a good metric with which to classify facial expressions.
Different expression representation techniques have been created, and many more are expected to be created in the next few years. Representations may be features designed by experts (feature engineering methods), as was demonstrated by Ptucha et al., Mollahosseini et al., and Zhang et al. [21,29,30,31], or embedded vectors obtained from training a deep neural network. The main objective of this work was to define a new facial expression recognition system that uses the representations obtained from different sources and combines them using a learned model. For this goal, we performed the following steps: (1) we first obtained several representations of facial expressions by applying different representation methods; (2) we then transformed each obtained representation into the same domain, the domain of representation errors of sparse representation (SR); finally, (3) in this new space, we trained a combination model to classify the signals. We used the extended Cohn–Kanade data set (CK+), the BU-3DFE data set, and the JAFFE data set for validation. The major innovation of this work is the exploration of feature extractors to select those features that are the most explanatory for facial expression recognition, using a sparse representation-based classifier.

3. Methodology

In this section, we detail the feature extraction and expression classification methods. The feature extraction methods used can be grouped according to local and geometric features, as well as global features extracted from pre-trained deep models. Local features use neighbor information, while geometric features consider points of interest in the face. Pre-trained neural network models provide a k-dimensional feature vector extracted from the last layer. Figure 1 represents the scheme of the proposed approach.

3.1. Feature Extraction Methods

In this work, we group the feature methods into local, geometric, and global. The local features were defined through the extraction of supervised features from facial patches. The facial patches were defined as regions in the face that were active during different facial expressions. It has been reported that some facial patches are common during the production of all basic expressions, while others are confined to a single expression [32]. The results indicated that these active patches were positioned below the eyes, between the eyebrows, and around the nose and mouth. To extract these patches from the face image, we first located the facial components. The locations of active patches were defined with respect to the positions of landmarks, which were estimated using Open-Face [33]. Happy and Routray [34] have observed that features formed from fewer facial patches can replace high-dimensional features without a significant decrease in recognition accuracy. Geometric features were defined according to the distances and regions between the different landmarks [23]. The global methods employed unsupervised features obtained by the pre-trained deep learning models, VGG and VGGFace, over the entire image [35].

3.1.1. Local Features

The local features were established through the extraction of supervised features from facial patches, defined as the regions of the face that are active during different facial expressions. We used the following methods for the extraction of local features:
  • Gabor wavelet filters (GW): This technique has previously been applied successfully for facial expression recognition [18,30]. In our work, 68 landmark points were selected. For each point, a patch sized 2 k + 1 × 2 k + 1 was used to compute the feature vector. Four scales and eight orientations were used to calculate the Gabor kernels, where k = 7 . This selection generated a vector of 2176 ( 68 × 4 × 8 ) elements.
  • Local binary patterns (LBP): this system has been widely used as a robust illumination-invariant feature descriptor [19,27,36,37]. This operator generates a binary number by comparing neighboring pixel values with the center pixel value. The uniform LBP and rotation-invariant uniform LBP [38] were also used in the experiment. These methods generated a vector that was the same size as the image ( 256 × 256 ).
  • Histograms of oriented gradients (HOG): This concept has also been successfully applied in facial expression recognition [19,39]. The basic idea of HOG features is that the local object’s appearance and shape can often be well-characterized by the distribution of local intensity gradients or edge directions, even without precise knowledge of the corresponding gradient or edge positions. The orientation analysis is robust to changes in illumination, as the histogram possesses translational invariance. The associated vector size was 256 × 256 × 2 .
  • Local phase quantization (LPQ): The LPQ feature [18,40] is a blur-robust image descriptor. The LPQ descriptor is based on the intensity of low-frequency phase components with respect to a centrally symmetric blur. Therefore, LPQ employs the phase information obtained from the short-term Fourier transform, which is locally computed on a window around each pixel of the image. The vector size was 256 × 256 .
  • RAW: The image intensity has also been used as a feature vector (RAW) [17,41]. Here, the vector size was 256 × 256 .

3.1.2. Geometric Features

Geometric measurements (GEO) were computed using 68 landmarks [23]. Figure 2 shows the order of the landmarks selected in our work. The order and number of landmarks varied in some of the previous works in the literature. We maintained the appropriate correspondence in all cases. Table 1 details the 17 geometric distance measures used.
We also used shape measurements for the eye, nose, and mouth regions. We defined these regions using the landmarks in each region. For example, the left eye region is defined by landmarks 37, 38, 39, 40, 41, and 42. Measures were defined for each region, as follows: solidity ( M 18 ) returns a scalar specifying the proportion of the pixels in the convex hull that are also in the region, computed as A r e a / C o n v e x A r e a ; the axes relationship ( M 19 ) is the ratio between the lengths of the minor and major axes of an ellipse, which has the same normalized second central moments as the region, computed as A x i s M i n / A x i s M a x ; tshe circularity factor ( M 20 ) is computed as 4 π A r e a s / P e r m s 2 ; eccentricity ( M 21 ) is a scalar that specifies the eccentricity of the ellipse, which has the same second moments as the region; and extent ( M 22 ) is a scalar that specifies the ratio of pixels in the region to pixels in the total bounding box. The distances of the centroids of each region to the center of the nose were also calculated. Table 2 summarizes the vector size used for each feature; the total vector size was 329,992.

3.1.3. Pre-Trained Deep Models

Several deep learning algorithms have been proposed and applied to FER [29,44,45], particularly from an explainable artificial intelligence (XAI) point of view [46,47,48]. Our interest was in the combination of different representation spaces. Thus, we selected pre-trained models to obtain facial expression representations. The training of deep learning architectures for FER poses a problem when using data sets such as CK+, JAFFE, and BU-3DFE, due to the small number of elements in these sets. Models trained with such data sets are prone to overfitting, which can hinder objective analysis when determining the contribution of the characteristics of the model to the system. Therefore, we selected general classification models (i.e., object classification and facial classification models) that were pre-trained on extensive databases, including the VGG-face [35] and VGG [49] models. In the case of the VGG-face model, as a feature, we selected the output of the Rectifier Linear Unit (ReLU) layer 33. For the VGG model, we selected the output of the ReLU layer 34. The size of both vectors was 4096 .

3.2. Expression Classification

Consider a set of training signals D = D 1 , D 2 , , D k R m × p from k different classes, where the columns of each sub-matrix D j = d 1 j , d 2 j , , d n j j R m × n j are signals from the class w j . Ideally, we have sufficient training samples of class w j , such that a test signal x R m that belongs to the same class can be approximated by a linear combination of the training samples from D j , which can be written as:
x = i = 1 n j α i j d i j ,
which can be rewritten as x = D δ j α R m , where δ j α = 0 , 0 , α 1 j , α 2 j , , α n j j , 0 , 0 T R p is a vector of coefficients having many values equal to zero, except for those associated with the class w j . As a valid test sample, x can be sufficiently represented only using the training samples from the same class; this representation is the sparsest among all others, finding the identity of x is equal to finding the sparsest solution of Equation (2). This is equivalent to solving the following optimization problem ( l 0 -minimization):
α = a r g m i n α R p α 0   s . t .   D α = x .
However, solving the l 0 minimization of an undetermined system of linear equations is NP-hard. If the sought solution α is sparse, the solution of the l 0   minimization problem, as defined in Equation (2), is equal to the solution of the following l 1 minimization problem [50]:
α = a r g m i n α R p α 1   s . t .   D α = x .
Then, the estimate x using the coefficients corresponding to a given class w j , x x j ^ = D δ j α , is possible. This is consistent with the previous findings, and also mimics the behavior of simple cells in the visual cortex. The error of the representation, e j = x x j ^ , can be used to determine the class of the signal x .

3.3. Proposed Combination Framework

When different classifiers based on SR are considered, the reconstruction error can be used to evaluate the combination rules. This supposes that the probability of success of each classifier D i for each class, P w j | D i , is the same. Combination methods of this type are called “class-conscious” [51].
Depending on its intrinsic characteristics, each expression can best be characterized in a particular subspace or subset. For example, expressions that involve some form of facial movement (e.g., opening the mouth) may be better described by the shape spaces, which are recorded as changes in gradient. On the other hand, changes in the frequency intensity of the image may be best characterized by texture analysis methods. This property suggests that classifiers exist that are specialists (experts) for some classes. As such, these classifiers should have greater weight in the final decision regarding classification.
For the calculation of the weights, a new feature space based on the output of each of the classifiers was generated. The variable d i , j x denotes the support that classifier D i gives to the hypothesis that x comes from class w j . The larger the support, the more likely it is that the class label belongs to w j . In this approach, d i , j x are features in a new space, defined as the intermediate feature space [51]. The support for class w j is calculated as:
μ j x = i = 1 L θ i , j d i , j x .
Linear regression is the most commonly used procedure to derive the weights for this model [21]. Algorithm 1 describes each step in the classification of a test pattern. Note that the output of each classifier d i , j , for the class w j , creates a new feature space. For each output subspace, an estimated model θ j weighs the decision of each classifier in the class w j . For the experiments, the fuzzy integral (FI) [51], linear opinion pools (LOP) [52], SVM, and naïve Bayes method were used to adjust the output of the classifiers (see Equation (8)). Other techniques for combining experts exist, but they need to be trained using a large number of samples [53].
Algorithm 1. Sparse representation fusion classification (SRFC). D 1 , . D L , with D i R n i × m , are the dictionaries for each feature space; g 1 x , , g L x , with g i x R n i and g i x g j x , are the feature extraction methods
1: Calculate sparse representation:
For i = 1 , , L
α i = a r g m i n α R p α 0 s . t . g i x = D i α (5)
2: Calculate the vote of each representation to each class:
For i = 1 , , L and j = 1 , , C
r i , j = g i x D i δ j α i 2 2 , (6)
where δ j selects the entries of α corresponding to the class j , and r j represents the residual test sample g i x with the linear combination D i δ j α . To obtain the vote for each class, the softmax function is applied to the inverse of the normalization of r i , j :
d i , j = σ s o f t m a x 1 r i , j r . , j 1 , (7)
where r . , j refers to the column j as a vector and d i , j represents the decision profile.
3: Trained combination rules:
μ j x = i = 1 L θ i , j d i , j x . (8)
The weights θ j are estimated for the decision profile for class w j .
4: Classification:
j ^ = a r g m a x j 1 , , C μ j . (9)
5: Return
The estimated class j ^ for the signal x

4. Experiments

In order to determine the influence of different features, we performed four experiments. (1) Statistical analysis was employed to determine which combination of rules presented the best classification results, for which purpose we evaluated 247 combinations of feature extraction methods—i.e., we considered all possible combinations of 8 features, excluding empty sets and those using only one feature set (28 − 1 − 8 = 247)—and 14 combination rules in three different data sets. (2) To investigate the generalization performance of the proposed method vs. individual classification methods, we performed an inter-database experiment. (3) We analyzed the influence of the methods for each class; and (4) we compared the obtained results with those using state-of-the-art methods on the same data set and experimental protocol. The results obtained from these experiments are described in detail in the next section.

4.1. Data Set

To evaluate our proposed method, we used three public data sets: the extended Cohn–Kanade (CK+) [54], JAFFE [55], and BU-3DFE [43]. In all experiments, person-independent FER scenarios were used [56]. Therefore, the subjects in the training set were completely different from the subjects within the test set (i.e., the subjects used for training were not used for testing). Following the recommendations in [38], each face image used in our experiments was cropped, based on the locations of the eye. The landmarks were obtained using Open-Face. The cropped face images were rescaled to 256 × 256 pixels. Figure 3 shows the various classification results.
  • CK+ data set: This includes 593 image sequences from 123 subjects. From the 593 sequences, we selected 325 sequences of 118 subjects, each of which met the criteria for one of the seven expressions [54]. The selected 325 sequences consisted of 45 angry, 18 contempt, 58 disgust, 25 fear, 69 happy, 28 sadness, and 82 surprise sequences [54]. In the neutral face scenario, we selected the first frame of the sequence of 33 randomly selected subjects.
  • BU-3DFE data set: This is known as the most challenging and difficult data set, mainly due to the presence of a variety of ethnic/racial ancestries and expression intensities [43]. A total of 700 expressive face images (1 intensity × 6 expressions × 100 subjects) and 100 neutral face images (each of which was of one subject) [43] were used.
  • JAFFE data set: This contains 10 female subjects and 213 facial expression images [55]. The number of images corresponding to each of the seven categories of expression (neutrality, happiness, sadness, surprise, anger, disgust, and fear) was almost always the same. Each actor repeated the same expression several times (i.e., two, three, or four times).
Figure 3. Classification results. The two top rows show different users, and the two bottom rows show each user with two different expressions. We denote the expression marked by the classifier system in red. Original images extracted from BU-3DFE [42,43].
Figure 3. Classification results. The two top rows show different users, and the two bottom rows show each user with two different expressions. We denote the expression marked by the classifier system in red. Original images extracted from BU-3DFE [42,43].
Applsci 12 05637 g003aApplsci 12 05637 g003b

4.2. Protocol

In this work, eight feature extraction methods—GW, LBP, HOG, LPQ, RAW, GEO, VGG, and VGGF—were used, according to the methodology proposed. These features have been used widely in the literature for FER [17,18,19,27,30,39,41,57,58]. We generated 247 possible scenarios, which were all combinations of the selected feature extraction method. We denote the classification schemes as follows: GW+SR with W, LBP+SR with B, HoG+SR with H, LPQ+SR with P, RAW+SR with R, GEO+SR with G, VGG+SR with V, and VGGF+SR with F.
Next, for example, a possible combination scenario might be denoted as W/B/H, which corresponds to GW/LBP/HOG+SR. We tested 14 combination rules for each scenario: five soft-level rules—product rule (RP), sum rule (RS), max rule (RMX), min rule (RMI), and median rule (RMD) [20]; three hard-level rules—weighted majority vote rule (WMV), recall rule (REC), and naïve Bayes rule (NB) [59]; and six trainable methods—fuzzy integral (FI) [51], linear opinion pools (LOP) [52], Bayes model (MB), SVM linear kernel model (ML), and SVM polynomial kernel model (MP). The results of these combinations were also compared to those of the individual methods.
The parameters for each of the methods were selected according to those obtained using state-of-the-art methods. The GW representation was obtained using five scales and eight orientations, in order to construct a set of Gabor filter banks for 25 × 25 and k = 7 neighborhoods [60]; the image resolution was 256 × 256 pixels. For the extraction of LBP features, as used by Huang et al. [38] and Ying et al. [17], we adopted a uniform LBP operation, with parameters of P = 8 and R = 2 in images sized 256 × 256 pixels. The histogram was extracted for each patch sized 25 × 25 . For HOG, the bin number was set to 9 , the cell size was 16 × 16 pixels, and a block size of 2 × 2 was adopted for each selected landmark point [19]. For the extraction of the LPQ [28], we set M = 5 and a = 1 / 5 . The histograms were extracted from each selected landmark point for those patches sized 25 × 25 . In the case of VGGF, as a feature, we selected the output of the ReLU layer 33; while, in the case of VGG, we selected the output of the ReLU layer 34.

4.3. Metrics

For this analysis, different metrics were used. Accuracy was calculated as the average number of successes, divided by the total number of observations (in this case, each face was considered an observation). The precision, recall, F1-score, and confusion matrix were also used for an analysis of the effectiveness of the system. Demšar [61] recommended the Friedman test, followed by the pairwise Nemenyi test, to compare multiple pieces of data. The Friedman test is a non-parametric alternative to the ANOVA test. The null hypothesis of the test H 0 is that all classifier models are equivalent. In our work, the Friedman test was used to identify the best classification scheme between different combination rules and all feature extraction methods. Similar to previously published works [31,62,63,64], leave-one-subject-out (LOSO) cross-validation was adopted in the evaluation.

4.4. Experimental Environment

One of the most important aspects to consider when evaluating the use of multiple classifier systems is that of time. A critical component of the system is the feature extraction method ( F s i ) and its representation by SRC. As the number of methods increases, the amount of time also increases. It was expected that method F s 1   + SRC has time t 1 , method F s 2   + SRC has time t 2 , method F s n + SRC has time t n , and the system time is t = t i . However, the system components are independent and depend only on the input image (i.e., there are no interdependencies); therefore, if the number of methods in the pool is not large, each one can be assigned to a processing unit. In this case, the mean time could be shown as t = m a x t i + c , where c is a constant. At present, such solutions are viable and easily accessible.
The hardware used to carry out the computation was a desktop computer with the following specifications: CPU: Intel i9-9900KF (16) @ 5.000 GHz; GPU: NVIDIA GeForce RTX 2060; memory: 7003 MiB/32,035 MiB; OS: Ubuntu 20.04.4; LTS: x86_64. With this configuration, the total computing time was 12.3 h (including feature extraction, sparse representation, training, and metrics).

5. Results and Discussion

In this section, we present the results and discuss the influence of the combination rule, the differences in the use of individual and multiple classification methods, and the contribution of each feature.

5.1. Statistical Analysis of the Combination Rule

Table 3, Table 4 and Table 5 show the accuracy values for the 14 combination rules (columns) and the best combinations of classifiers in different subspaces (rows) for the CK+ data set, BU-3DFE data set, and JAFFE data set, respectively. Regarding the CK+ data set (Table 3), an accuracy of more than 0.985 was achieved; in the BU-3DFE data set (Table 4), we achieved an accuracy of more than 0.821 ; and for JAFFE (Table 5), the accuracy was 0.776 (in all three cases, more than 20 combinations were selected). The best accuracy for each combination is underlined. The last row of tables shows the average accuracy across the 247 combinations. With the large span of classification accuracy values, it is unlikely that these values will be commensurable; however, although the average values across the feature combination do not serve as a valid performance metric, they give a rough indication of the achievements of the combinations. The tables show that the trainable combination rules of MP and ML presented the best results in most cases. The JAFFE data set had the fewest images ( 213 images), thus limiting the training phase of the trainable combiners. In this case, the best option seemed to be the use of non-trainable combiners.
To determine the best choice for CK+ and BU-3DFE, we calculated the ranks of the combiners. For example, for the combined features R/W/H/P/V/F in the CK+ dataset, the order by rank was as follows: MP (the best), ML, MB, RP, RS, RMD, MV, WMV, LOP, REC, NB, RMX, FI, and RMI (the worst). In case of a tie, the ranks were shared. The average ranks across the combination of features in CK+ (BU-3DFE) were as follows: RP 6.709 (5.830), RS 6.749 (5.873), RMX 11.699 (12.694), RMI 12.956 (12.484), RMD 6.749 (5.873), MV 7.820 (9.982), WMV 7.524 (7.401), REC 7.512 (6.328), NB 10.757 (9.423), LOP 8.124 (7.294), FI 11.682 (12.611), MB 2.767 (3.792), MP 1.984 (2.678), and ML 1.970 (2.739). These results indicate that MP, ML, and MB were the best combiners in both cases.
The Friedman non-parametric test was executed on the ranks, followed by a multiple comparisons test. The Friedman test score was 2326.60 (2249.70), giving a p -value of approximately 0 (0), indicating significant differences among the ranks for the CK+ (BU-3DFE) data set. The Nemenyi post hoc test and Bonferroni–Dunn post hoc test were applied, in order to determine those methods that had significant differences.
The result of the Nemenyi post hoc test (two-tailed test) showed that there were significant differences between the MB, MP, and ML methods and all of the other methods, with a significance level of α < 0.05 . For MB, MP, and ML, we applied the Bonferroni–Dunn post hoc test (one-tailed test) to strengthen the power of the test hypotheses. At a significance level of 0.05, the Bonferroni–Dunn post hoc test did not show significant differences between the MB, MP, and ML methods. Therefore, we can conclude that, in general, these methods had similar behavior in this case. For the case of the BU-3DFE data set, similar results were obtained. The proposed methods (MB, MP, and ML) were significantly superior to the others for combination rules in FER problems; however, when very little training data sets are available, the use of non-trainable combination rules is suggested.
Table 6, Table 7 and Table 8 show the accuracy, precision, recall, and F1-score measurements of the best schemes for each individual and combination rule in the CK+ data set, BU-3DFE data set, and JAFFE data set, respectively. In the case of the CK+ data set (Table 6), the R/W/H/P/G/F+MP scheme presented an accuracy of 0.992 and also obtained the best results in terms of precision, recall, and F1-score, with values of 0.991, 0.985, and 0.999, respectively. In the case of the BU-3DFE data set (Table 7), the R/W/B/P/G+MP scheme presented an 0.828 accuracy, with the best values of precision, recall, and F1-score also being obtained by this scheme (0.893, 0.833, and 0.965, respectively). In both cases, the MP combination rule achieved the best results. In the case of the JAFFE data set (Table 8), the W/H/P/F+ML scheme obtained an accuracy of 0.828; however, the R/H/P/G+RP scheme showed the best results for precision, recall, and F1 (0.871, 0.809, and 0.959, respectively). This was because, as mentioned earlier, the trainable combination rules are inadequate when there are insufficient data in the training data set.
In summary, our results exceeded the results of the state-of-the-art for these types of features. In all cases, the combination rules improved upon the accuracy of the single methods. Figure 4 shows that the classification error of the individual methods was greater than that of the combination schemes.

5.2. Multiple vs. Individual Classification Methods

For the experiments, the BU-3DFE and JAFFE data sets were used to construct the training and test sets, respectively. The training and test sets contained six expression classes—anger, disgust, fear, happiness, sadness, and surprise—which were common in both data sets. The experimental results presented in Table 9 show that the fusion method R/W/B/P/G+MP achieved an accuracy of 0.54 and an F1-score of 0.721, higher than those of the other methods used for comparison. Ramis et al. [65] presented a method to combine multiple datasets and conducted an exhaustive evaluation of a proposed system based on a CNN that analyzed and compared performance using single- and cross-dataset approaches with other architectures. They reported that they trained with BU-4FDE and tested with JAFFE, obtaining an accuracy of 0.4317. In our current proposal, we study the most important features by applying different feature extraction methods for facial expression representation, transforming each obtained representation into a sparse representation (SR) domain, and training combination models to classify signals, using the extended Cohn–Kanade (CK+), BU-3DFE, and JAFFE data sets for validation. In general, combining the feature methods increased the system performance. This supports the theory that trainable fusion methods can find experts for each subspace of features and efficiently combine them. In all cases, the combination rules were better at identification than the individual methods. However, the values of the metrics were lower than in the case in which face images from a single data set were used for both the training set and test set. This performance degradation can mainly be attributed to the fact that the face images were collected under two different controlled conditions. For a better generalization ability across image acquisition conditions, it is necessary to collect large training data sets under various image acquisition conditions [66].
In summary, our findings confirm the increased accuracy of the proposed combination system, with respect to the individual systems. The best schemes (R/W/H/P/G/F+MP and R/W/B/P/G+MP) combined the best and the worst individual schemes. This increases the diversity of the system, offering the opportunity to select the best representation subspaces for each class.

5.3. Analysis of the Influence of the Feature Methods

The visualization of learning models (i.e., the generated subspaces) using classifiers such as SVM and the naïve Bayes method is important. To analyze the contribution of each feature to each class, a logistic regression model was trained on the decision profile. From the weights obtained for each class, a weight map was created. The weight maps obtained for each data set are shown in Figure 5. On the horizontal axis, we can see the weight of the feature space for each of the classes. On the vertical axis, we can see the importance of each subspace for features over a class.
As can be seen, the LPQ, HOG, and RAW methods presented high specialization values for some of the classes. For example, for the CK+ data set, LPQ had great decision power over the angry and happy classes, the feature space RAW was most successful in the neutral class, and HOG specialized in the surprise and sad classes. In general, we can observe that the spaces of characteristics do not always specialize in the same classes. This is because there are significant differences between the images in the different data sets.
We calculated the frequencies of the feature methods from Table 3, Table 4 and Table 5. These frequencies are shown in Figure 6 and Table 10. It can be observed that the LPQ, HOG, and RAW methods were present in more than 80 percent of the selected combinations. This demonstrates that they had significant weight in the classification of some expressions.
In summary, the LPQ, HOG, and RAW methods were present in most of the selected combinations, and some features were most efficient for particular expressions (e.g., RAW for neutral, LPQ for angry and happy, LBP for disgust, and HOG for surprise).

6. Conclusions

In this work, we studied the different contributions of various combination methods to a facial recognition system. To achieve this, we utilized the concept of sparse representations. The results demonstrated that the combination of classifiers can lead to improved performance when compared to individual classifiers. In particular, the MB, MP, and ML combination methods were significantly superior to other combination rules in facial expression recognition problems.
In all cases, the combination of different features improved the results, compared to when using only one feature. Individually, the best features were HoG (in 82% of combinations) and LPQ (in 83% of combinations); however, when combining different features, the best selection was RAW/GW/LBP/LPQ/GEO. Curiously, the best schemes (R/W/H/P/G/F+MP and R/W/B/P/G+MP; accuracy of 0.992 and 0.828, respectively) combined the best and the worst individual schemes. This increases the diversity of the system, providing the opportunity to select the best representation subspace for each class.
Different features were found to specialize in different expressions; for example, the feature space RAW specializes in the neutral class, LPQ specializes in the angry and happy classes, LBP specializes in the disgust class, and HOG specializes in the surprise and sad classes, while the fear class did not have a predominant feature space.
We observed that the LPQ, HOG, and RAW methods were present in more than 80% of the selected combinations. This demonstrates that they play a significant role in the classification of some expressions.
We presented the most explanatory features for facial expression recognition. Feature relevance techniques appear to be some of the most-used schemes for post hoc explanations [4,5], as they are capable of providing an explicit description of the inner behavior of the model. There was no combination of features in a model that appeared to be clearly the best, leading to the conclusion that the recognition of a facial expression is the result of a combination of many features.
To the best of our knowledge, there has been no other study focused on the explainability of features. In this work, a system based on SRC was presented to provide an explanation, thus determining the most important features for each expression. This information can be useful in designing neural networks or performing fine-tuning in existing ones. CNN-based methods can be stripped from the top classification layer and that vector is, in fact, a feature representation, but we cannot have any knowledge of how these features are calculated and every time the CNN is trained, those features can change. Our method is based on well-known feature extraction methods; therefore, we can establish what features are more revealing. We can take advantage of this knowledge to improve the performance of all the models, including the CNN-based methods.
In our future work, we will explore the most focused-upon face regions for each feature vector, in order to determine the most interesting regions used for features and for each expression.

Author Contributions

Conceptualization, J.M.B.-R. and A.J.-i.-C.; methodology, P.D.M.-F., A.J.-i.-C. and T.I.R.; software, P.D.M.-F.; formal analysis, P.D.M.-F. and J.M.B.-R.; data curation, P.D.M.-F. and J.M.B.-R.; writing—original draft preparation, P.D.M.-F.; writing—review and editing, T.I.R. and A.J.-i.-C.; funding acquisition, A.J.-i.-C. All authors have read and agreed to the published version of the manuscript.

Funding

The research was funded by Project PID2019-104829RA-I00 “EXPLainable Artificial INtelligence systems for health and well-beING (EXPLAINING)” funded by MCIN/AEI/10.13039/501100011033.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data involved in the article are all shown in the figures and tables. However, they are also available from the first author on reasonable request.

Conflicts of Interest

The authors declare that they have no competing interests.

References

  1. Zhu, X.; He, Z.; Zhao, L.; Dai, Z.; Yang, Q. A Cascade Attention Based Facial Expression Recognition Network by Fusing Multi-Scale Spatio-Temporal Features. Sensors 2022, 22, 1350. [Google Scholar] [CrossRef] [PubMed]
  2. Bonnard, J.; Dapogny, A.; Dhombres, F.; Bailly, K. Privileged Attribution Constrained Deep Networks for Facial Expression Recognition. arXiv 2022, arXiv:2203.12905. [Google Scholar]
  3. Adadii, A.; Mohammed, B. Peeking Inside the Black-Box: A Survey on Explainable Artificial Intelligence (XAI). IEEE Access 2018, 6, 52138–52160. [Google Scholar] [CrossRef]
  4. Barredo Arrieta, A.; Díaz-Rodríguez, N.; Del Ser, J.; Bennetot, A.; Tabik, S.; Barbado, A.; Garcia, S.; Gil-Lopez, S.; Molina, D.; Benjamins, R.; et al. Explainable Explainable Artificial Intelligence (XAI): Concepts, taxonomies, opportunities and challenges toward responsible AI. Inf. Fusion 2020, 58, 82–115. [Google Scholar] [CrossRef] [Green Version]
  5. Petrović, N.; Moyà-Alcover, G.; Jaume-i-Capó, A.; González-Hidalgo, M. Sickle-cell disease diagnosis support selecting the most appropriate machine learning method: Towards a general and interpretable approach for cell morphology analysis from microscopy images. Comput. Biol. Med. 2020, 126, 104027. [Google Scholar] [CrossRef]
  6. Guidotti, R.; Monreale, A.; Ruggieri, S.; Turini, F.; Giannotti, F.; Pedreschi, D. A survey of methods for explaining black box models. ACM Comput. Surv. 2018, 51, 1–42. [Google Scholar] [CrossRef] [Green Version]
  7. Letham, B.; Rudin, C.; McCormick, T.H.; Madigan, D. Interpretable classifiers using rules and bayesian analysis: Building a better stroke prediction model. Ann. Appl. Stat. 2015, 9, 1350–1371. [Google Scholar] [CrossRef]
  8. Ribeiro, M.T.; Singh, S.; Guestrin, C. “Why should I trust you?” Explaining the predictions of any classifier. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, San Francisco, CA, USA, 13–17 August 2016; pp. 1135–1144. [Google Scholar] [CrossRef]
  9. Chandrashekar, G.; Sahin, F. A survey on feature selection methods. Comput. Electr. Eng. 2014, 40, 16–28. [Google Scholar] [CrossRef]
  10. Weitz, K.; Hassan, T.; Schmid, U.; Garbas, J.-U. Deep-learned faces of pain and emotions: Elucidating the differences of facial expressions with the help of explainable AI methods Tief erlernte Gesichter von Schmerz und Emotionen: Aufklärung der Unterschiede von Gesichtsausdrücken mithilfe erklärbarer K. tm-Technisches Mess. 2019, 86, 404–412. [Google Scholar] [CrossRef]
  11. Gund, M.; Bharadwaj, A.R.; Nwogu, I. Interpretable emotion classification using temporal convolutional models. In Proceedings of the 25th International Conference on Pattern Recognition (ICPR), Milan, Italy, 10–15 January 2021; pp. 6367–6374. [Google Scholar] [CrossRef]
  12. Burg, K.T. Explaining Dnn Based Facial Expression Classifications. Bachelor’s Thesis, Utrecht University, Utrecht, The Netherlands, 2021. Volume 7. [Google Scholar]
  13. Wang, X.; He, J.; Jin, Z.; Yang, M.; Wang, Y.; Qu, H. M2Lens: Visualizing and Explaining Multimodal Models for Sentiment Analysis. IEEE Trans. Vis. Comput. Graph. 2021, 28, 802–812. [Google Scholar] [CrossRef]
  14. Lian, Z.; Li, Y.; Tao, J.H.; Huang, J.; Niu, M.Y. Expression Analysis Based on Face Regions in Read-world Conditions. Int. J. Autom. Comput. 2020, 17, 96–107. [Google Scholar] [CrossRef] [Green Version]
  15. Jung, H.; Lee, S.; Yim, J.; Park, S.; Kim, J. Joint fine-tuning in deep neural networks for facial expression recognition. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; pp. 2983–2991. [Google Scholar]
  16. Deramgozin, M.; Jovanovic, S.; Rabah, H.; Ramzan, N. A Hybrid Explainable AI Framework Applied to Global and Local Facial Expression Recognition. In Proceedings of the 2021 IEEE International Conference on Imaging Systems and Techniques (IST), Kaohsiung, Taiwan, 24–26 August 2021; pp. 1–5. [Google Scholar] [CrossRef]
  17. Ying, Z.-L.; Wang, Z.-W.; Huang, M.-W. Facial expression recognition based on fusion of sparse representation. In Advanced Intelligent Computing Theories and Applications. With Aspects of Artificial Intelligence; Springer: Berlin/Heidelberg, Germany, 2010; pp. 457–464. [Google Scholar]
  18. Li, L.; Ying, Z.; Yang, T. Facial expression recognition by fusion of gabor texture features and local phase quantization. In Proceedings of the 2014 12th International Conference on Signal Processing (ICSP), Hangzhou, China, 19–23 October 2014; pp. 1781–1784. [Google Scholar]
  19. Ouyang, Y.; Sang, N.; Huang, R. Accurate and robust facial expressions recognition by fusing multiple sparse representation based classifiers. Neurocomputing 2015, 149, 71–78. [Google Scholar] [CrossRef]
  20. Kittler, J.; Hatef, M.; Duin, R.P.W.; Matas, J. On combining classifiers. Pattern Anal. Mach. Intell. IEEE Trans. 1998, 20, 226–239. [Google Scholar] [CrossRef] [Green Version]
  21. Ptucha, R.; Savakis, A. Fusion of static and temporal predictors for unconstrained facial expression recognition. In Proceedings of the 2012 19th IEEE International Conference on Image Processing (ICIP), Orlando, FL, USA, 30 September–3 October 2012; pp. 2597–2600. [Google Scholar]
  22. Ji, Y.; Idrissi, K. Automatic facial expression recognition based on spatiotemporal descriptors. Pattern Recognit. Lett. 2012, 33, 1373–1380. [Google Scholar] [CrossRef]
  23. Tsalakanidou, F.; Malassiotis, S. Real-time 2D+ 3D facial action and expression recognition. Pattern Recognit. 2010, 43, 1763–1775. [Google Scholar] [CrossRef]
  24. Olshausen, B.A.; Field, D.J. Sparse coding with an overcomplete basis set: A strategy employed by V1? Vis. Res. 1997, 37, 3311–3325. [Google Scholar] [CrossRef] [Green Version]
  25. Olshausen, B.A.; Field, D.J. Emergence of simple-cell receptive field properties by learning a sparse code for natural images. Nature 1996, 381, 607–609. [Google Scholar] [CrossRef]
  26. Wright, J.; Yang, A.Y.; Ganesh, A.; Sastry, S.S.; Ma, Y. Robust face recognition via sparse representation. Pattern Anal. Mach. Intell. IEEE Trans. 2009, 31, 210–227. [Google Scholar] [CrossRef] [Green Version]
  27. Weifeng, L.; Caifeng, S.; Yanjiang, W. Facial expression analysis using a sparse representation based space model. In Proceedings of the 2012 IEEE 11th International Conference on Signal Processing (ICSP), Beijing, China, 21–25 October 2012; Volume 3, pp. 1659–1662. [Google Scholar]
  28. Zhen, W.; Zilu, Y. Facial expression recognition based on local phase quantization and sparse representation. In Proceedings of the 2012 Eighth International Conference on Natural Computation (ICNC), Chongqing, China, 29–31 May 2012; pp. 222–225. [Google Scholar]
  29. Mollahosseini, A.; Chan, D.; Mahoor, M.H. Going deeper in facial expression recognition using deep neural networks. In Proceedings of the 2016 IEEE Winter Conference on Applications of Computer Vision (WACV), Lake Placid, NY, USA, 7–10 March 2016; pp. 1–10. [Google Scholar]
  30. Zhang, S.; Li, L.; Zhao, Z. Facial expression recognition based on Gabor wavelets and sparse representation. In Proceedings of the 2012 IEEE 11th International Conference on Signal Processing (ICSP), Beijing, China, 21–25 October 2012; Volume 2, pp. 816–819. [Google Scholar]
  31. Ptucha, R.; Savakis, A. Manifold based sparse representation for facial understanding in natural images. Image Vis. Comput. 2013, 31, 365–378. [Google Scholar] [CrossRef]
  32. Zhong, L.; Liu, Q.; Yang, P.; Liu, B.; Huang, J.; Metaxas, D.N. Learning active facial patches for expression analysis. In Proceedings of the 2012 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Providence, RI, USA, 16–21 June 2012; pp. 2562–2569. [Google Scholar]
  33. Baltrušaitis, T.; Robinson, P.; Morency, L.-P. OpenFace: An open source facial behavior analysis toolkit. In Proceedings of the IEEE Winter Conference on Applications of Computer Vision, Lake Placid, NY, USA, 7–10 March 2016. [Google Scholar]
  34. Happy, S.L.; Routray, A. Automatic Facial Expression Recognition Using Features of Salient Facial Patches. Affect. Comput. IEEE Trans. 2015, 6, 1–12. [Google Scholar] [CrossRef] [Green Version]
  35. Parkhi, O.M.; Vedaldi, A.; Zisserman, A. Deep Face Recognition. In Proceedings of the BMVC, Swansea, UK, 7–10 September 2015; Volume 1, p. 6. [Google Scholar]
  36. Liu, P.; Han, S.; Meng, Z.; Tong, Y. Facial expression recognition via a boosted deep belief network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 1805–1812. [Google Scholar]
  37. Liu, Z.; Song, X.; Tang, Z. A novel SRC fusion method using hierarchical multi-scale LBP and greedy search strategy. Neurocomputing 2015, 151, 1455–1467. [Google Scholar] [CrossRef]
  38. Huang, M.-W.; Wang, Z.; Ying, Z.-L. A new method for facial expression recognition based on sparse representation plus LBP. In Proceedings of the 2010 3rd International Congress on Image and Signal Processing (CISP), Yantai, China, 16–18 October 2010; Volume 4, pp. 1750–1754. [Google Scholar]
  39. Dalal, N.; Triggs, B. Histograms of oriented gradients for human detection. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition CVPR 2005, San Diego, CA, USA, 20–25 June 2005; Volume 1, pp. 886–893. [Google Scholar]
  40. Ojansivu, V.; Heikkilä, J. Blur Insensitive Texture Classification Using Local Phase Quantization. In Image and Signal Processing; Springer: Berlin/Heidelberg, Germany, 2008; ISBN 978-3-540-69904-0. [Google Scholar]
  41. Yuan, X.-T.; Liu, X.; Yan, S. Visual classification with multitask joint sparse representation. Image Process. IEEE Trans. 2012, 21, 4349–4360. [Google Scholar] [CrossRef]
  42. Yin, L.; Chen, X.; Sun, Y.; Worm, T.; Reale, M. A high-resolution 3d dynamic facial expression database. In Proceedings of the 2008 8th IEEE International Conference on Automatic Face and Gesture Recognition, FG 2008, Amsterdam, The Netherlands, 17–19 September 2008; pp. 1–6. [Google Scholar] [CrossRef]
  43. Yin, L.; Wei, X.; Sun, Y.; Wang, J.; Rosato, M.J. A 3D facial expression database for facial behavior research. In Proceedings of the 7th International Conference on Automatic Face and Gesture Recognition, FGR 2006, Southampton, UK, 10–12 April 2006; pp. 211–216. [Google Scholar]
  44. Liu, M.; Li, S.; Shan, S.; Wang, R.; Chen, X. Deeply learning deformable facial action parts model for dynamic expression analysis. In Proceedings of the Asian Conference on Computer Vision, Singapore, 1–5 November 2014; pp. 143–157. [Google Scholar]
  45. Marrero Fernandez, P.D.; Guerrero Pena, F.A.; Ing Ren, T.; Cunha, A. FERAtt: Facial Expression Recognition with Attention Net. arXiv 2019, arXiv:1902.03284. [Google Scholar]
  46. Manresa-Yee, C.; Ramis, S. Assessing Gender Bias in Predictive Algorithms Using EXplainable AI; Association for Computing Machinery: New York, NY, USA, 2021. [Google Scholar] [CrossRef]
  47. Wang, Z.; Qinami, K.; Karakozis, I.C.; Genova, K.; Nair, P.; Hata, K.; Russakovsky, O. Towards fairness in visual recognition: Effective strategies for bias mitigation. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 8916–8925. [Google Scholar] [CrossRef]
  48. Xu, T.; White, J.; Kalkan, S.; Gunes, H. Investigating Bias and Fairness in Facial Expression Recognition. In Proceedings of the European Conference on Computer Vision, Glasgow, UK, 23–28 August 2020. [Google Scholar]
  49. Simonyan, K.; Zisserman, A. Very Deep Convolutional Networks for Large-Scale Image Recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  50. Donoho, D.L. For most large underdetermined systems of linear equations the minimal ?1-norm solution is also the sparsest solution. Commun. Pure Appl. Math. 2006, 59, 797–829. [Google Scholar] [CrossRef]
  51. Kuncheva, L.I.; Bezdek, J.C.; Duin, R.P.W. Decision templates for multiple classifier fusion: An experimental comparison. Pattern Recognit. 2001, 34, 299–314. [Google Scholar] [CrossRef]
  52. Jacobs, R.A. Methods for combining experts’ probability assessments. Neural Comput. 1995, 7, 867–888. [Google Scholar] [CrossRef]
  53. Jordan, M.I.; Jacobs, R.A. Hierarchical mixtures of experts and the EM algorithm. Neural Comput. 1994, 6, 181–214. [Google Scholar] [CrossRef] [Green Version]
  54. Lucey, P.; Cohn, J.F.; Kanade, T.; Saragih, J.; Ambadar, Z.; Matthews, I. The Extended Cohn-Kanade Dataset (CK+): A complete dataset for action unit and emotion-specified expression. In Proceedings of the 2010 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), San Francisco, CA, USA, 13–18 June 2010; pp. 94–101. [Google Scholar]
  55. Lyons, M.; Akamatsu, S.; Kamachi, M.; Gyoba, J. Coding facial expressions with gabor wavelets. In Proceedings of the Third IEEE International Conference on Automatic Face and Gesture Recognition, Nara, Japan, 14–16 April 1998; pp. 200–205. [Google Scholar]
  56. Zeng, Z.; Pantic, M.; Roisman, G.; Huang, T.S. A survey of affect recognition methods: Audio, visual, and spontaneous expressions. Pattern Anal. Mach. Intell. IEEE Trans. 2009, 31, 39–58. [Google Scholar] [CrossRef]
  57. Li, X.; Hong, X.; Moilanen, A.; Huang, X.; Pfister, T.; Zhao, G.; Pietikäinen, M. Reading Hidden Emotions: Spontaneous Micro-expression Spotting and Recognition. arXiv 2015, arXiv:1511.00423. [Google Scholar]
  58. Ouyang, Y.; Sang, N.; Huang, R. Robust automatic facial expression detection method based on sparse representation plus LBP map. Opt. J. Light Electron. Opt. 2013, 124, 6827–6833. [Google Scholar] [CrossRef]
  59. Kuncheva, L.I.; Rodríguez, J.J. A weighted voting framework for classifiers ensembles. Knowl. Inf. Syst. 2014, 38, 259–275. [Google Scholar] [CrossRef]
  60. Zavaschi, T.H.H.; Britto, A.S.; Oliveira, L.E.S.; Koerich, A.L. Fusion of feature sets and classifiers for facial expression recognition. Expert Syst. Appl. 2013, 40, 646–655. [Google Scholar] [CrossRef]
  61. Demšar, J. Statistical comparisons of classifiers over multiple data sets. J. Mach. Learn. Res. 2006, 7, 1–30. [Google Scholar]
  62. Huang, X.; Zhao, G.; Zheng, W.; Pietikäinen, M. Spatiotemporal local monogenic binary patterns for facial expression recognition. Signal Process. Lett. IEEE 2012, 19, 243–246. [Google Scholar] [CrossRef]
  63. Li, Y.; Wang, S.; Zhao, Y.; Ji, Q. Simultaneous facial feature tracking and facial expression recognition. Image Process. IEEE Trans. 2013, 22, 2559–2573. [Google Scholar]
  64. Lee, S.H.; Plataniotis, K.; Konstantinos, N.; Ro, Y.M. Intra-Class Variation Reduction Using Training Expression Images for Sparse Representation Based Facial Expression Recognition. Affect. Comput. IEEE Trans. 2014, 5, 340–351. [Google Scholar] [CrossRef]
  65. Ramis, S.; Buades, J.M.; Perales, F.J.; Manresa-Yee, C. A Novel Approach to Cross dataset studies in Facial Expression Recognition. Multimed. Tools Appl. 2022. [Google Scholar] [CrossRef]
  66. Littlewort, G.; Bartlett, M.S.; Fasel, I.; Susskind, J.; Movellan, J. Dynamics of facial expression extracted automatically from video. Image Vis. Comput. 2006, 24, 615–625. [Google Scholar] [CrossRef]
Figure 1. Proposed classifier system, divided into three sections: feature extraction methods, sparse representation classification, and combination rules.
Figure 1. Proposed classifier system, divided into three sections: feature extraction methods, sparse representation classification, and combination rules.
Applsci 12 05637 g001
Figure 2. The 68 landmarks used in our work. Each landmark is marked with its number. Original image extracted from BU-3DFE [42,43].
Figure 2. The 68 landmarks used in our work. Each landmark is marked with its number. Original image extracted from BU-3DFE [42,43].
Applsci 12 05637 g002
Figure 4. Classification error of the best schemes for each combination rule and individual schemes: (a) CK+ data set; (b) BU-3DFE data set; (c) JAFFE data set.
Figure 4. Classification error of the best schemes for each combination rule and individual schemes: (a) CK+ data set; (b) BU-3DFE data set; (c) JAFFE data set.
Applsci 12 05637 g004
Figure 5. Weight map learning for the combination of all features. NE: neutral; AN: anger; DI: disgust; FE: fear; HA: happy; SA: sadness; SU: surprise. (a) CK+ data set; (b) BU-3DFE data set; and (c) JAFFE data set.
Figure 5. Weight map learning for the combination of all features. NE: neutral; AN: anger; DI: disgust; FE: fear; HA: happy; SA: sadness; SU: surprise. (a) CK+ data set; (b) BU-3DFE data set; and (c) JAFFE data set.
Applsci 12 05637 g005
Figure 6. Frequency of features in selected combinations (Table 3, Table 4 and Table 5) for the CK+, BU-3DFE, and JAFFE data sets. RAW: (R), GW: (W), LBP: (B), HOG: (H), LPQ: (P), GEO: (G), VGG: (V), VGGF: (F).
Figure 6. Frequency of features in selected combinations (Table 3, Table 4 and Table 5) for the CK+, BU-3DFE, and JAFFE data sets. RAW: (R), GW: (W), LBP: (B), HOG: (H), LPQ: (P), GEO: (G), VGG: (V), VGGF: (F).
Applsci 12 05637 g006
Table 1. Geometric facial measurements. d i j : the Euclidean distance between landmarks i and j ; α : the angle between two lines; ξ i j : the line defined by i and j ; l i j k the length of the curve defined by i , j , and k ; o 1 and o 2 : the centers of the right and left eye, respectively.
Table 1. Geometric facial measurements. d i j : the Euclidean distance between landmarks i and j ; α : the angle between two lines; ξ i j : the line defined by i and j ; l i j k the length of the curve defined by i , j , and k ; o 1 and o 2 : the centers of the right and left eye, respectively.
MeasurementNameDistances Involved
M 1 Inner eyebrow displacement d 40 , 22 , d 43 , 23
M 2 Outer eyebrow displacement d 18 , o 1 , d 28 , o 2
M 3 Inner eyebrow corners dist. d 22 , 23
M 4 Eyebrow from nose root dist. d 22 , 28 , d 23 , 28
M 5 Eye opening d 39 , 41 , d 44 , 48
M 6 Eye shape d 39 , 41 / d 37 , 40 , d 44 , 48 / d 29 , 46
M 7 Nose length d 29 , 31
M 8 Nose width d 32 , 36
M 9 Lower lip boundary length l 55 , 56 , 57 , 58 , 59 , 60 , 49
M 10 Mouth corners dist. d 49 , 55
M 11 Mouth opening d 63 , 67
M 12 Mouth shape d 63 , 67 / d 49 , 55
M 13 Nose–mouth corners angle α ξ 32 , 49 , ξ 36 , 55
M 14 Mouth corners to eye dist. d o 1 , 49 , d o 2 , 55
M 15 Mouth corners to nose dist. d 49 , 34 , d 55 , 34
M 16 Upper lip to nose dist. d 52 , 34
M 17 Lower lip to nose dist. d 67 , 34
Table 2. Feature summary. For each feature, we provide its type and size.
Table 2. Feature summary. For each feature, we provide its type and size.
TypeFeatureVector Size
LocalGW 2176
LocalLBP 256 × 256
LocalHOG 256 × 256 × 2
LocalLPQ 256 × 256
LocalRAW 256 × 256
GeometricGEO 68 × 2
Table 3. Accuracy selection was greater than 0.985 for the 14-rule combination methods and 247 combinations of eight subspaces in the CK+ data set. CSC: a combination of soft classifiers; CHC: a combination of hard classifiers; TC: trainable classifiers. Best classifier result is underlined. Best feature/classifier is in bold type.
Table 3. Accuracy selection was greater than 0.985 for the 14-rule combination methods and 247 combinations of eight subspaces in the CK+ data set. CSC: a combination of soft classifiers; CHC: a combination of hard classifiers; TC: trainable classifiers. Best classifier result is underlined. Best feature/classifier is in bold type.
FeaturesCSCCHCTC
RPRSRMXRMIRMDRMWMVRECNBLOPFIMBMPML
R/B/P/V0.9860.9860.9500.9130.9860.9750.9720.9690.9580.9780.9500.9780.9750.975
R/P/V/F0.9610.9610.9360.9080.9610.9440.9470.9580.9390.9550.9360.9800.9890.983
R/W/B/P/V0.9750.9750.9470.9300.9750.9720.9640.9500.9500.9750.9470.9860.9780.978
R/W/B/V/F0.9690.9690.9470.9330.9690.9640.9640.9610.9550.9610.9470.9860.9830.980
R/W/H/P/F0.9690.9690.9470.9410.9690.9580.9550.9610.9470.9640.9470.9890.9830.983
R/W/P/G/F0.9470.9470.9020.8940.9470.9530.9530.9610.9410.9250.9020.9720.9860.986
R/B/H/V/F0.9690.9690.9530.9330.9690.9660.9660.9530.9500.9660.9530.9750.9800.986
R/H/P/V/F0.9640.9640.9440.9160.9640.9500.9550.9530.9270.9530.9440.9830.9860.986
R/P/G/V/F0.9390.9390.8830.8800.9390.9330.9410.9610.9080.9300.8830.9750.9890.980
W/B/H/P/V0.9750.9750.9550.9270.9750.9640.9640.9530.9500.9640.9550.9860.9830.983
W/B/H/V/F0.9690.9690.9550.9300.9690.9610.9580.9470.9360.9580.9550.9780.9860.983
R/W/B/H/P/V0.9780.9780.9470.9330.9780.9750.9720.9550.9550.9720.9470.9890.9860.983
R/W/B/H/P/F0.9750.9750.9530.9440.9750.9780.9690.9530.9640.9750.9530.9860.9830.983
R/W/B/H/V/F0.9750.9750.9470.9390.9750.9640.9610.9500.9500.9610.9470.9890.9800.986
R/W/B/P/G/V0.9580.9580.9080.8880.9580.9720.9660.9500.9410.9410.9080.9860.9830.983
R/W/B/P/V/F0.9750.9750.9470.9270.9750.9750.9690.9610.9610.9610.9470.9890.9800.980
R/W/H/P/G/V0.9580.9580.8970.8850.9580.9550.9530.9500.9190.9330.8970.9800.9800.986
R/W/H/P/G/F0.9550.9550.9050.8940.9550.9580.9550.9530.9410.9330.9050.9780.9860.986
R/W/H/P/V/F0.9660.9660.9360.9270.9660.9660.9550.9530.9390.9550.9360.9860.9920.989
R/B/H/P/V/F0.9720.9720.9530.9250.9720.9750.9690.9550.9500.9660.9530.9890.9860.986
R/B/H/G/V/F0.9610.9610.8990.8880.9610.9640.9640.9530.9470.9330.8990.9750.9860.986
R/H/P/G/V/F0.9440.9470.8940.8830.9470.9580.9550.9470.9020.9270.8940.9780.9890.986
W/B/H/P/V/F0.9720.9720.9550.9250.9720.9720.9660.9550.9390.9530.9550.9830.9800.986
R/W/B/H/P/G/F0.9660.9640.9110.8940.9640.9660.9610.9530.9640.9250.9110.9860.9830.983
R/W/B/H/P/V/F0.9720.9720.9470.9330.9720.9660.9690.9550.9550.9580.9470.9890.9800.983
R/W/B/P/G/V/F0.9580.9580.9080.8880.9580.9660.9580.9550.9580.9330.9080.9860.9830.983
R/W/H/P/G/V/F0.9500.9500.8970.8850.9500.9530.9470.9530.9270.9300.8970.9800.9890.989
W/B/H/P/G/V/F0.9550.9550.8910.8800.9550.9610.9610.9530.9390.9300.8910.9830.9860.986
Average0.9460.9460.9100.9020.9460.9410.9460.9480.9290.9400.9100.9660.9710.971
Table 4. Accuracy selection greater than 0.821 for 14-rule combination methods and 247 combinations of eight subspaces in the BU-3DFE data set. CSC: a combination of soft classifiers; CHC: a combination of hard classifiers; TC: trainable classifiers. Best classifier result is underlined. Best feature/classifier is in bold type.
Table 4. Accuracy selection greater than 0.821 for 14-rule combination methods and 247 combinations of eight subspaces in the BU-3DFE data set. CSC: a combination of soft classifiers; CHC: a combination of hard classifiers; TC: trainable classifiers. Best classifier result is underlined. Best feature/classifier is in bold type.
FeaturesCSCCHCTC
RPRSRMXRMIRMDRMWMVRECNBLOPFIMBMPML
R/W/P/G0.7740.7780.7240.6970.7780.7720.7830.7810.7710.7710.7240.8070.8240.826
R/W/B/P/G0.8070.8070.7410.7090.8070.7810.7980.8020.7790.7880.7410.8070.8280.826
R/W/B/P/V0.8020.8000.7520.7670.8000.7900.7970.7950.7780.7970.7520.8220.8160.817
R/W/H/P/G0.7880.7910.7290.7140.7910.8050.8050.8090.7790.7840.7290.8090.8260.824
R/H/P/G/V0.7840.7840.7240.7140.7840.7790.7910.7900.7780.7760.7240.8030.8240.819
W/H/P/G/V0.7780.7780.7190.6980.7780.7790.7880.8000.7840.7810.7190.7950.8220.819
W/H/P/V/F0.7950.7980.7340.7480.7980.7980.7930.7970.7810.7830.7340.8220.8140.810
R/W/B/H/P/G0.8090.8070.7430.7120.8070.8020.8000.8190.7830.7980.7430.8190.8280.822
R/W/B/H/P/F0.8170.8160.7450.7520.8160.8100.8070.8170.7950.8090.7450.8210.8220.824
R/W/B/P/G/V0.7980.8000.7410.7210.8000.7880.7950.7950.7830.7830.7410.8140.8260.826
R/W/H/P/G/V0.8000.8000.7280.7210.8000.7980.8090.8140.7910.7900.7280.8160.8260.824
R/W/H/P/G/F0.8030.8030.7280.7220.8030.7980.8070.8280.7840.7930.7280.8160.8120.807
R/B/H/P/V/F0.8070.8070.7640.7670.8070.8030.8100.8140.8020.8000.7640.8220.8100.810
R/H/P/G/V/F0.8100.8090.7220.7240.8090.7860.8030.8120.7790.7760.7220.8170.8220.824
W/B/H/P/G/V0.8000.7980.7410.7070.7980.7950.7950.8070.7900.7840.7410.8090.8220.821
W/B/H/P/G/F0.8030.8030.7260.7100.8030.8050.8070.8190.7900.7900.7260.8120.8220.822
W/B/H/P/V/F0.8050.8030.7500.7590.8030.8050.8090.8090.7900.7860.7500.8280.8140.814
R/W/B/H/P/G/V0.8020.8000.7430.7220.8000.7910.8020.8160.7900.7860.7900.8210.8220.819
R/W/B/H/P/G/F0.8210.8210.7410.7220.8210.8160.8190.8120.7950.7980.7910.8280.8260.822
R/W/B/H/P/V/F0.8020.8030.7500.7600.8030.8120.8160.8170.7980.7950.7910.8240.8160.816
Average0.7790.7790.7270.7270.7790.7620.7740.7770.7660.77400.7290.7890.7940.794
Table 5. Accuracy selection greater than 0.776 for 14-rule combination methods and 247 combinations of eight subspaces in the JAFFE data set. CSC: a combination of soft classifiers; CHC: a combination of hard classifiers; TC: trainable classifiers. Best classifier result is underlined. Best feature/classifier is in bold type.
Table 5. Accuracy selection greater than 0.776 for 14-rule combination methods and 247 combinations of eight subspaces in the JAFFE data set. CSC: a combination of soft classifiers; CHC: a combination of hard classifiers; TC: trainable classifiers. Best classifier result is underlined. Best feature/classifier is in bold type.
FeaturesCSCCHCTC
RPRSRMXRMIRMDRMWMVRECNBLOPFIMBMPML
R/B/H0.7460.7460.7260.7060.7460.7010.7110.7060.6920.7610.7260.7510.7260.741
B/H/P0.7460.7460.7060.6670.7460.7260.7060.6920.7010.7660.7060.7160.7360.736
R/W/H/G0.7510.7460.6470.6220.7460.7210.7310.7360.5970.7310.6470.7610.7360.736
R/H/P/G0.7710.7660.6720.6120.7660.7010.7210.7110.5970.7160.6720.7310.7360.736
R/H/P/F0.7160.7160.6970.6870.7160.7260.7110.6970.6270.7360.6970.7460.7660.761
W/B/H/P0.7210.7210.6820.6820.7210.6870.7060.7510.6470.7510.6820.7210.7610.756
W/H/P/F0.7460.7460.6470.6770.7460.7260.7060.7260.6420.7360.6470.7210.7610.776
R/W/H/P/G0.7510.7510.6470.6220.7510.7160.7310.7360.6120.7410.6470.7610.7460.746
R/W/H/G/F0.7660.7610.6420.6170.7610.7160.7260.7260.5620.7210.6420.7410.6920.687
R/B/H/G/F0.7610.7560.6770.6170.7560.7310.7160.7010.5820.7110.6770.7160.7210.726
R/H/P/G/F0.7660.7660.6670.6170.7660.7210.7210.7060.5870.7010.6670.7360.7510.751
W/H/P/V/F0.7410.7460.6520.6370.7460.7160.7310.7160.5870.7260.6520.7160.7660.756
R/W/B/H/P/F0.7410.7410.6920.6820.7410.7260.7210.7260.6370.7310.6920.7310.7460.761
R/W/B/H/G/V0.7310.7310.6520.6120.7310.7160.7210.7210.5220.6970.6520.7710.7260.731
R/W/B/H/G/F0.7610.7610.6570.6220.7610.7260.7310.7210.5870.7160.6570.7460.7160.726
R/W/B/H/V/F0.7610.7560.6920.6370.7560.7160.7310.7310.5720.7210.6920.7410.6970.692
R/W/H/P/G/F0.7660.7660.6420.6220.7660.7260.7410.7360.5820.7160.6420.7410.7360.736
R/B/H/P/G/F0.7660.7710.6770.6220.7710.7410.7260.7060.6070.7060.6770.7310.7460.736
R/B/H/P/V/F0.7610.7610.6870.6470.7610.7460.7510.7310.6020.7460.6870.7360.7210.711
R/W/B/H/P/G/F0.7710.7710.6570.6270.7710.7310.7310.7260.5970.7210.7560.7460.7560.756
R/W/B/H/P/V/F0.7610.7610.6920.6470.7610.7260.7260.7310.5920.7210.7260.7310.7210.721
R/B/H/P/G/V/F0.7610.7610.6570.5970.7610.7460.7510.7360.5620.7110.7410.7260.7360.731
Average0.7000.7000.6400.6040.7000.6850.7000.6970.5860.6910.6430.6930.6850.686
Table 6. The best schemes for each combination rule ( e · 1000 1 ) for the CK+ data set. Acc: accuracy; Prec: precision; Rec: recall. Best classifier result is underlined.
Table 6. The best schemes for each combination rule ( e · 1000 1 ) for the CK+ data set. Acc: accuracy; Prec: precision; Rec: recall. Best classifier result is underlined.
Combined SchemesIndividual Schemes
RP (R/B/P/V)RS (R/B/P/V)RMX (R/B)RMI (W/B)RMD (R/B/P/V)RM (R/W/B/P)WMV (R/B/H/F)REC (R/B/P/V)NB (R/B/H/F)LOP (R/W/B)FI (R/B)MB (R/W/H/P/F)MSVMP (R/W/H/P/G/F)MSVML (R/W/H/P/G/F)RBHWPGVF
Acc.0.9860.9860.9640.9750.9860.9800.9750.9690.9660.9780.9640.9890.9920.9890.9530.9470.9300.9470.8970.7180.7540.852
F10.9920.9920.9680.9820.9920.9870.9810.9750.9730.9820.9680.9890.9910.9890.9610.9660.9510.9640.9400.7740.7980.898
Prec0.9850.9850.9450.9680.9850.9770.9670.9570.9530.9680.9450.9800.9850.9800.9360.9420.9180.9380.9000.6740.6990.835
Rec0.9980.9980.9940.9960.9980.9970.9960.9950.9950.9970.9940.9980.9990.9980.9930.9920.9900.9920.9850.9360.9480.976
Table 7. The best schemes for each combination rule ( e · 1000 1 ) for the BU-3DFE data set. Acc: accuracy; Prec: precision; Rec: recall. Best classifier result is underlined.
Table 7. The best schemes for each combination rule ( e · 1000 1 ) for the BU-3DFE data set. Acc: accuracy; Prec: precision; Rec: recall. Best classifier result is underlined.
Combined SchemesIndividual Schemes
RP (R/B/H/P/G/F)RS (R/B/H/P/G/F)RMX (H/P)RMI (R/H/V)RMD (R/B/H/P/G/F)RM (R/W/B/H/P/G/F)WMV (R/W/B/H/P/G/F)REC (R/W/H/P/G/F)NB (R/B/H/P/F)LOP (R/W/B/H/P/F)FI (R/W/B/H/P/G/F)MB (W/B/H/P/V/F)MSVMP (R/W/B/P/G)MSVML (R/W/P/G)RBHWPGVF
Acc.0.8210.8210.7660.7780.8210.8160.8190.8280.8070.8090.7910.8280.8280.8260.7310.7090.7570.6900.7380.5290.6740.681
F10.8910.8900.8490.8580.8900.8900.8900.8940.8780.8830.8740.8920.8930.8920.8330.8100.8420.7990.8400.6540.7820.791
Prec0.8290.8280.7700.7800.8280.8300.8300.8350.8120.8190.8070.8300.8330.8300.7530.7200.7600.7060.7590.5370.6870.697
Rec0.9650.9650.9500.9540.9650.9640.9640.9650.9600.9620.9580.9650.9650.9660.9410.9330.9470.9290.9470.8570.9190.924
Table 8. The best schemes for each combination rule ( e · 1000 1 ) for the JAFFE data set. Acc: accuracy; Prec: precision; Rec: recall. Best classifier result is underlined.
Table 8. The best schemes for each combination rule ( e · 1000 1 ) for the JAFFE data set. Acc: accuracy; Prec: precision; Rec: recall. Best classifier result is underlined.
Combined SchemesIndividual Schemes
RP (R/H/P/G)RS (R/B/H/P/G/F)RMX (R/B)RMI (R/H)RMD (R/B/H/P/G/F)RM (R/B/H/P/V)WMV (R/B/H/G/V)REC (R/W/H/P/G/V)NB (B/H/P)LOP (B/H/P)FI (R/W/B/H/P/G/F)MB (R/W/B/H/G/V)MSVMP (R/H/P/F)MSVML (W/H/P/F)RBHWPGVF
Acc.0.7710.7710.7260.7260.7710.7560.7560.7560.7010.7660.7560.7710.7660.7760.6720.6720.6670.6320.6620.3930.4430.577
F10.8710.8700.8380.8300.8700.8670.8500.8480.8030.8590.8600.8560.8440.8590.7870.7850.7870.7330.7930.5410.5780.701
Prec0.8090.8070.7640.7500.8070.8040.7760.7680.7070.7880.7940.7790.7620.7840.6890.6940.6950.6290.6990.4160.4620.583
Rec0.9590.9580.9440.9420.9580.9570.9500.9510.9340.9550.9530.9540.9500.9560.9270.9240.9260.8980.9320.8040.8090.889
Table 9. Generalization performance on the two different data sets. The BU-3DFE and JAFFE data sets were used for training and testing, respectively. Best feature results are in bold type.
Table 9. Generalization performance on the two different data sets. The BU-3DFE and JAFFE data sets were used for training and testing, respectively. Best feature results are in bold type.
MethodAccuracyF1-Measure
R/B/H/W/P/G/V/F0.5230.726
R/H/P/F0.5170.703
R/W/B/H/P/G/F0.5270.722
R/W/B/P/G0.5470.721
R0.4330.635
B0.3680.494
H0.4880.626
W0.3530.594
P0.4630.592
G0.3780.582
V0.1890.332
F0.4680.720
Table 10. Frequency of features in selected combinations for the CK+, BU-3DFE, and JAFFE data sets. RAW: (R), GW: (W), LBP: (B), HOG: (H), LPQ: (P), GEO: (G), VGG: (V), VGGF: (F).
Table 10. Frequency of features in selected combinations for the CK+, BU-3DFE, and JAFFE data sets. RAW: (R), GW: (W), LBP: (B), HOG: (H), LPQ: (P), GEO: (G), VGG: (V), VGGF: (F).
FeatureCKBU-3DFEJAFFETotal
R0.860.760.820.82
W0.720.860.590.73
B0.660.620.590.40
H0.690.811.000.82
P0.831.000.680.83
G0.420.710.540.55
V0.830.620.280.59
F0.790.480.680.75
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Marrero-Fernandez, P.D.; Buades-Rubio, J.M.; Jaume-i-Capó, A.; Ing Ren, T. An Approach for Selecting the Most Explanatory Features for Facial Expression Recognition. Appl. Sci. 2022, 12, 5637. https://doi.org/10.3390/app12115637

AMA Style

Marrero-Fernandez PD, Buades-Rubio JM, Jaume-i-Capó A, Ing Ren T. An Approach for Selecting the Most Explanatory Features for Facial Expression Recognition. Applied Sciences. 2022; 12(11):5637. https://doi.org/10.3390/app12115637

Chicago/Turabian Style

Marrero-Fernandez, Pedro D., Jose M. Buades-Rubio, Antoni Jaume-i-Capó, and Tsang Ing Ren. 2022. "An Approach for Selecting the Most Explanatory Features for Facial Expression Recognition" Applied Sciences 12, no. 11: 5637. https://doi.org/10.3390/app12115637

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop