Next Article in Journal
Intercomparison of Soil Moisture Retrieved from GNSS-R and from Passive L-Band Radiometry at the Valencia Anchor Station
Next Article in Special Issue
Parallel Transmission of Distributed Sensor Based on SCTP and TCP for Heterogeneous Wireless Networks in IoT
Previous Article in Journal
A Robust Method for Automatic Panoramic UAV Image Mosaic
Previous Article in Special Issue
Household Power Demand Prediction Using Evolutionary Ensemble Neural Network Pool with Multiple Network Structures
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Improved Real-Time Facial Expression Recognition Based on a Novel Balanced and Symmetric Local Gradient Coding

1
College of Computer Science and Information Engineering, Tianjin University of Science and Technology, Tianjin 300457, China
2
Department of Electronic and Information Engineering, Chonbuk National University, Jeonbuk 561-756, Korea
*
Authors to whom correspondence should be addressed.
Sensors 2019, 19(8), 1899; https://doi.org/10.3390/s19081899
Submission received: 27 February 2019 / Revised: 17 April 2019 / Accepted: 17 April 2019 / Published: 22 April 2019

Abstract

:
In the field of Facial Expression Recognition (FER), traditional local texture coding methods have a low computational complexity, while providing a robust solution with respect to occlusion, illumination, and other factors. However, there is still need for improving the accuracy of these methods while maintaining their real-time nature and low computational complexity. In this paper, we propose a feature-based FER system with a novel local texture coding operator, named central symmetric local gradient coding (CS-LGC), to enhance the performance of real-time systems. It uses four different directional gradients on 5 × 5 grids, and the gradient is computed in the center-symmetric way. The averages of the gradients are used to reduce the sensitivity to noise. These characteristics lead to symmetric of features by the CS-LGC operator, thus providing a better generalization capability in comparison to existing local gradient coding (LGC) variants. The proposed system further transforms the extracted features into an eigen-space using a principal component analysis (PCA) for better representation and less computation; it estimates the intended classes by training an extreme learning machine. The recognition rate for the JAFFE database is 95.24%, whereas that for the CK+ database is 98.33%. The results show that the system has advantages over the existing local texture coding methods.

1. Introduction

Facial expressions include a wealth of information, which directly reflects people’s psychological characteristics, providing a very important means of communication. Psychologists have reported that communicating emotional information consists of approximately 7% of the language used, 38% of the tone used, and 55% of the facial expressions used [1]. It can be seen, in the course of human communication, if they cannot communicate with language, they can express basic emotions through their facial expressions. Therefore, FER has been widely studied and has become a research hotspot [2].
In addition to its relevance in human to human communication, advanced intelligent interactions between humans and machines is also an important goal driving the study of FER [3]. At present, FER systems are widely used in medical, education, driving [4] and other fields. Especially because facial expressions contain abundant psychological information, this work could be assistive to both researchers and the broader community applying FER methods.
FER system includes expression image preprocessing, face detection, region segmentation, expression feature extraction, classification, and other components. Expression feature extraction is essential for FER systems [5]. Various feature extraction algorithms, based on the local feature extraction methods, provide strong stability to light and changes of the scene orientation. Local binary pattern (LBP) [6] calculates the order of gradient between each pixel in the image and its local neighborhood points. Then the binary order relation is coded to form a local binary pattern. Finally, the multi-region histogram is used as the feature description of the image. Jabid and Chae [7] further improved the LBP algorithm to Local Directional Pattern (LDP), which has good robustness to illumination and less computation. The Gabor wavelet [6] kernel function has the same characteristics as the two-dimensional reflector of simple cells in human cerebral cortex, which can better describe texture features. Histograms of oriented gradients (HOG) [8] is also a common feature extraction algorithm. HOG features are constructed by calculating and counting the gradient direction histogram of the local area of the image. It is robust to geometric and optical changes of image. Adaptive Neuro-fuzzy Inference Systems (ANFIS) [9] is an image compression prediction coding algorithm for constructing predictors, which is also used in the field of facial expression recognition. In addition, there are local Gabor binary patterns (LGBP) [10], scale-invariant feature transform (SIFT) [11] and some corresponding improved algorithms [12]. These have been widely used in face recognition, texture classification, facial expression recognition, and other fields. These traditional methods tend to be more effective in specific small sample sets. In recent time, deep learning has become very useful for many computer vision applications [13,14]. However, due to the large amount of data required for deep learning as well as long training time, a traditional algorithm is selected for further improvement. In local feature operators, LBP operator and Gabor wavelet transform are representative methods. However, Gabor wavelet transform takes a long time and has a large feature dimension. The method of reducing the dimension will affect the accuracy of facial expression. The LBP operator and some improved methods are not perfect in facial expression recognition. These operators are unable to accurately describe texture features such as facial muscles and wrinkles.
In order to address this problem, in 2014, Ying et al. have proposed a Local Gradient Coding (LGC) [15] algorithm. The image texture feature is extracted by encoding the strength of the neighborhood from the four gradient directions of the horizontal, vertical, and diagonal lines in the 3 × 3 neighborhood. This method fully considers the influence of texture changes on the expression and can improve the recognition rate. However, the neighborhood of the algorithm is fixed and the image texture feature has performance limitations on large scale problems. It is easy to lose part of the gradient direction information and is susceptible to noise. In order to improve this deficiency, in 2017, Shen et al. proposed an enhanced LGC facial expression recognition (LGC-FN) [16] based on a 5 × 5 neighborhood. In the 5 × 5 neighborhood, feature values are calculated in three directions, horizontal and two diagonal, around the target pixel. Finally, classification is performed using a support vector machine (SVM). This method has further improved the recognition rate.
In the neighborhood division, the LGC-FN uses the 5 × 5 neighborhood to calculate the eigenvalues, but in the calculation process, there are two key limitations. First, vertical gradient information is not considered. Although this information could be added in a simple way, it results in a higher computational dimension; this is not ideal for many real-time applications. Second, the horizontal, vertical, and diagonal information are not balanced.
In order to solve these problems, we propose an innovative algorithm, called Central Symmetric Local Gradient Coding (CS-LGC), which defines the neighborhood as a 5 × 5 grid, uses the concept of center symmetry, and extracts the gradient information in the horizontal, vertical, and two diagonal directions. This algorithm presents several advantages. First, it extracts the gradient information in four directions as an average so that the obtained feature values of the target pixels are more representative. Second, the gradient information is more balanced and comprehensive, and the computational dimension does not increase. Finally, the algorithm has good generalization ability and can be widely used in the field of facial expression recognition.
The rest of this paper is arranged as follows: the materials and methods for this research are presented in Section 2, where Section 2.1 introduces the basic theory of the related research methods and Section 2.2 presents our proposed method. Section 3 presents the results of experimental studies to evaluate the performance of the proposed method; an analysis of the experimental results is presented in Section 4. Section 5 presents a summary of the proposed approach and a direction for future research.

2. Materials and Methods

2.1. Related Work

2.1.1. The LBP Operator Theory

The LBP operator is the first operator to use local features for biometric recognition. Usually, it is defined in the 3 × 3 neighborhood, and the pixel value of its center point is taken as the threshold value, which is subtracted from the surrounding points. If the result is greater than or equal 0, it is set to 1; otherwise is set to 0. Finally, the result is arranged in order to compute a binary value of the center point. This value is then converted to its decimal representation, that is, the corresponding gray value of the point. The specific process of the LBP algorithm is shown in Figure 1. The formula of the operator is shown in (1):
LBP ( g c ) =   p = 0 p 1 s ( g i g c ) · 2 p   ; s ( x ) = { 1     x 0 0     x < 0
In the above formulas, c denotes the central pixel, g c denotes the gray value of the central pixel, g i denotes the gray value of the surrounding pixels, and p is the sampling point number. The LBP algorithm also has many derivative algorithms, such as Center-Symmetric Local Binary Patterns (CSLBP) operators. A CSLBP operator applies the idea of centrosymmetry to an LBP operator. The computational complexity of LBP operators is reduced.

2.1.2. The LGC Operator Theory

The neighborhood size of the LGC operator is a 3 × 3 grid, as shown in Figure 2. The LGC operator compares the gray values in the horizontal, vertical, and diagonal directions for eight adjacent pixels to obtain an 8-bit binary number. The binary coding is converted into a decimal number and used to compute the LGC histogram statistics and link all of the sub-image LGC histogram features to form the entire face of identification features. The operator is defined by (2) below, where P is the sampling point number and R is the radius of the neighborhood area. Usually, R = 1, P = 8:
L G C P R = s ( g 1 g 3 ) 2 7 + s ( g 4 g 5 ) 2 6 + s ( g 6 g 8 ) 2 5 + s ( g 1 g 6 ) 2 4 + s ( g 2 g 7 ) 2 3 + s ( g 3 g 8 ) 2 2 + s ( g 1 g 8 ) 2 1 + s ( g 3 g 6 ) 2 0
This encoding method is consistent with the changing trend of human eyes, mouth, forehead, and other characteristic points; in addition, it fully considers the effect of texture changes on expression. Compared with traditional LBP operators, LGC operators have stronger discrimination capabilities.
LGC operators also have some extension operators, such as LGC-HD (LGC based on horizontal and diagonal gradient prior principle) operators and LGC-FN (LGC based on a 5 × 5 neighborhood horizontal and diagonal gradient prior principle) operators mentioned below. An LGC-HD operator gives priority to gradients in the horizontal and diagonal directions, which reduces the feature dimension. The formula is shown in (3), where R = 1, P = 6:
L G C H D P R = s ( g 1 g 3 ) 2 4 + s ( g 4 g 5 ) 2 3 + s ( g 6 g 8 ) 2 2 + s ( g 1 g 8 ) 2 1 + s ( g 3 g 6 ) 2 0

2.1.3. The LGC-FN Operator Theory

The LGC-FN operator expands the neighborhood to a 5 × 5 grid, and calculates the feature values in three directions, horizontal and two diagonal, around the target pixel. The model diagram is shown in Figure 3 and the coding formula is (4). Among them, R = 2, P = 8:
L G C F N P R = s ( g 1 g 2 ) 2 7 + s ( g 3 g 4 ) 2 6 + s ( g 5 g 6 ) 2 5 + s ( g 7 g 8 ) 2 4     + s ( g 1 g 8 ) 2 3 + s ( g 3 g 6 ) 2 2 + s ( g 2 g 7 ) 2 1 + s ( g 4 g 5 ) 2 0

2.1.4. The LGC-AD Operator Theory

The LGC-AD operator directly adds the vertical gradient information to the LGC-FN operator. Model diagram refers to Figure 3. The formula is shown in (5).
L G C A D P R = s ( g 1 g 2 ) 2 11 +   s ( g 3 g 4 ) 2 10 +   s ( g 5 g 6 ) 2 9 +   s ( g 7 g 8 ) 2 8 + s ( g 1 g 7 ) 2 7 +   s ( g 3 g 5 ) 2 6 +   s ( g 4 g 6 ) 2 5 +   s ( g 2 g 8 ) 2 4 + s ( g 1 g 8 ) 2 3 +   s ( g 3 g 6 ) 2 2 +   s ( g 2 g 7 ) 2 1 +   s ( g 4 g 5 ) 2 0

2.2. Proposed Method

We propose a balanced and symmetrical local texture coding-based FER system. The structure diagram of this system is shown in Figure 4.

2.2.1. FER Data Characteristics

There are still some problems with the LGC operator and its extensions. LGC operator compares three groups of horizontal and three groups of vertical pixels, but only two groups of diagonal pixels are compared. As a result, the facial feature extraction in the diagonal direction of the gradient information is reduced, which negatively influences on the recognition accuracy.
The LGC-HD operator deletes the gradient information in the vertical direction, reduces the dimension of the feature, and has no significant impact on the recognition accuracy in a 3 × 3 neighborhood. However, the LGD-FN operator extends its neighborhood to a 5 × 5 grid. We believe that the information in the vertical expression cannot be completely ignored. In addition, the correlation between horizontal and central pixels selected by LGC-FN operators is not the strongest, and the characteristics of central pixels cannot be described in more detail. Finally, if this supplementary information is directly included, it leads to higher operational dimensions.

2.2.2. CS-LGC

In order to solve these problems, we use the idea of central symmetry, and propose a centrally-symmetric local gradient encoding (CS-LGC) method, which calculates the eigenvalues of the selected pixels in the direction of the target pixels from the horizontal, vertical, 45°, and 135°. The proposed method is as follows; the results of CS-LGC and other operators’ feature extraction are shown in Figure 5:
C S L G C P R = s ( g 13     g 16 ) 2 7 +   s ( g 14     g 15 ) 2 6 +   s ( g 9     g 12 ) 2 5 +   s ( g 10     g 11 ) 2 4 + s ( g 1     g 8 ) 2 3 +   s ( g 3     g 6 ) 2 2 +   s ( g 2     g 7 ) 2 1 + s ( g 4     g 5 ) 2 0
First, we select a 5 × 5 neighborhood and perform the CS-LGC operation on pixels in the 0°, 90°, 45°, and 135° directions, and compare the values of the corresponding location pixels in this order as shown in Figure 6. Then, according to (6), we compute a set of binary strings. Finally, the binary strings are converted into a decimal number to get the final pixel value. Among them, R = 2, P = 8.
For example, as shown in Figure 7, we use the CS-LGC operator to calculate the target pixel to get a binary string of 01100001. Converting it to a decimal value of 0 × 128 + 1 × 64 + 1 × 32 + 0 × 16 + 0 × 8 + 0 × 4 + 0 × 2 + 1 × 1 = 97, the feature value of the final target pixel is 97.

2.2.3. CS-LGC Feature Extraction

Histograms can effectively describe the global features of image texture, but histogram statistics directly on the whole expression image will lose a lot of structural details. Therefore, the expression image is divided into several equal-sized and non-overlapping sub-blocks, and then the histogram of each sub-block area is counted. Finally, the histograms of each sub-block are cascaded, and the cascaded histograms represent the features of the expression image. In this paper, the specific steps of feature extraction process are as follows:
(1)
Facial expression image is divided into m × n sub-blocks.
(2)
Each sub-block is coded by CS-LGC operator, and the texture image of CS-LGC is obtained.
(3)
The CS-LGC histogram of each sub-block is counted.
(4)
Cascade the histograms of m × n sub-blocks. The cascaded histogram shows the features of the expression image. The extraction process is shown in Figure 8.

2.2.4. Dimensionality Reduction

Because the dimensions of the CS-LGC feature vector are relatively high, in order to not affect the processing efficiency (in addition to removing redundancy and noise) in the facial expression feature, we use a Principal Component Analysis (PCA) [17] algorithm to reduce the dimensions of the feature matrix. PCA is a common data analysis method. It transforms the original data into a set of linear independent representations of each dimension through a linear transformation, and can be used to extract the main feature components of the data. When the contribution rate is 0.95, the main features extracted by the PCA algorithm can already describe the overall characteristics. At this rate, the characteristic dimension is not large, and the recognition performance is not very different from the performance when the contribution rate is greater than 0.95. Therefore, in the experiment, we select 0.95 as the contribution rate. It can effectively improve the efficiency of the FER system.

2.2.5. Training and Classification

In the experiment, we use the Extreme Learning Machine (ELM) algorithm [18] for training and classification. A SLFNs with L hidden nodes is described as follows:
f L ( x ) = i   =   1 L β i G ( a i   , b i   , x ) = h ( x ) β
where G (ai, bi, x) is the activation function with the hold parameters ai and bi.
Compared with other training algorithms, the ELM algorithm does not need to adjust the input weights and the bias of the implied network neurons, so that the unique solution can be obtained with higher efficiency. Since its initial introduction, the network has shown better performance than traditional BP algorithm and SVM in solving complex pattern recognition problems because of its simple structure, fast speed, and good generalization performance [19]. Therefore, the ELM algorithm is selected for training and classification in this paper.

3. Results

In order to evaluate the performance of the proposed operator on facial expression images, we have conducted experiments on the JAFFE database [20] (The database was planned and assembled by Michael Lyons, Miyuki Kamachi, and Jiro Gyoba. The photos were taken at the Psychology Department in Kyushu University Fukuoka, Japan) and the CK+ database [21] (The database was planned and assembled by Patrick Lucey, Jeffrey F. Cohn, Takeo Kanade, Jason Saragih, Zara Ambadar, University of Pittsburgh, Pittsburgh, PA, USA). The experimental software environment consists of MATLAB R2016b (MATLAB is produced by MathWorks, Natick, MA, USA). The hardware environment’s CPU is an Intel (R) Core (TM) i7-7700HQ @ 2.80 GHz; the system’s memory is 8 GB. The operating system was 64-bit Microsoft Windows 10.

3.1. Experiments on the CK+ Database

The CK+ database is based on the Cohn-Kanade Dataset and was released in 2010. The database contains 123 subjects, 593 image sequences, and the last frame of each image sequence has the label of action units. The size of each image is 640 × 490. Of the 593 image sequences, there are 327 sequences with emotion. The database contains women, men, Europeans, Africans, and some other people. Everyone has six expressions: angry, disgusted, fearful, happy, sad, and surprised. Some samples in CK+ database are presented in Figure 9.
We have experimented with all six expressions of 20 people in the CK+ library. Each person randomly selected three pictures for each expression. A sample of 360 images has been selected. We detect the face of the original image, and normalize face image to 128 × 128 pixels.

3.1.1. Choose the Best Image Blocking Method

Because the number of sub-images affects the size of the image feature matrix, using image partitioning and a sub-image feature matrix to represent a facial expression image can more fully represent its information. This can increases the effective recognition of features and better reflects the differences of the local features of a facial expression. In order to discuss the influence of image segmentation on the recognition algorithm, the effect of image segmentation on the recognition rate is explored in this paper as follows. When the number of sub-images is too large, the feature matrix becomes very large, requiring substantial computation time. Therefore, in the trial, we select the following blocking methods from the original image: 1 × 2, 1 × 4, 2 × 2, 2 × 4, 4 × 2, and 4 × 4. In this experiment, we select 20 images of each expression for training, and the remaining 40 images for testing. The experimental results are shown in Table 1. From this, we can see that when the block mode is 1 × 4, the recognition rate is the highest, reaching 98.33%. Therefore, we use the 1 × 4 block method in the subsequent experiments. We divide 128 × 128 facial expression images into 4 blocks according to 1 × 4 block method. The histogram size of each region feature image is 1 × 4096.

3.1.2. Compare the Performance of Different Operators

We select training samples n = 6, 8, 10, 12, 14, 16, 18, 20 for this experiment. We compare the CS-LGC operator with LBP, CSLBP, LGC, LGC-HD, LGC-FN, and LGC-AD operators. The experimental results are shown in Table 2 and Figure 10.
As shown in Figure 11, we have plotted the Receiver Operating Characteristic (ROC) curves. In this experiment, we select 20 samples of each expression for training, and the remaining 40 are tested. The curve of the CS-LGC is below the other curves, which means that the CS-LGC method providese the best performance.

3.2. Experiments on the JAFFE Database

The JAFFE database has a total of 213 emoticons of 10 women; there are seven expressions in the pictures: anger, disgust, fear, happiness, neutral, sadness, and surprise. Each expression has three or four samples. The size of each image is 256 × 256. Some samples in JAFFE database are presented in Figure 12. We first face detection in the original image, and normalize face image to 128 × 128 pixels. We use the 1 × 4 block manner for the experiments. Due to the small amount of data in the JAFFE database, in order to ensure the reliability of the experiment, we use the Leave-p-out Cross Validation method to conduct the experiments.

Compare the Performance of Different Operators

In this experiment, three samples of for each of the seven expressions are organized into a collection. Each of these collections is further divided into 10 groups. Next, nine groups of experiments are performed: leave-1-out, leave-2-out, … and leave-9-out. Finally, the average of five, randomly selected experimental results for each group of experiments is used to compute the final recognition rate. We compare the CS-LGC operator with LBP, CSLBP, LGC, LGC-HD, LGC-FN and LGC-AD operators. The experimental results are shown in Table 3 and Figure 13.
As shown in Figure 14, we have plotted the Receiver Operating Characteristic (ROC) curves. In this experiment, we select 10 samples of each expression for training, and the remaining 20 are tested. Although the CS-LGC curve has some intersection with the other curves, it is still below the other curves, which indicates its strong performance.

3.3. The Processing Time of Different Operators

In the experiment, we compare the processing time of the LBP, CSLBP, LGC, LGC-HD, LGC-FN, LGC-AD and CSLBC operators. Table 4 shows the processing time of each operator. The LGC-AD operator has the longest processing time among the derivative LGC operators, which is due to the high feature dimension. CS-LGC operators introduce the idea of centrosymmetric. On the basis of guaranteeing the Omni-directional selection of information, the feature size is reduced and the operation efficiency is improved.

4. Discussion

From Table 2 and Table 3, we can see that compared with other operators, the CS-LGC operator has certain advantages in the recognition rate for the CK+ database and the JAFFE database, which we attribute mainly to two factors. First, the LBP operator only considers the relationship between the central pixel and the adjacent pixel, ignoring the correlation between the adjacent pixels. The coding method of the LGC and its extension operators is consistent with the changing trend of human eyes, mouth, and other features. Fully considering the influence of texture changes on expression, the LGC and its extension operators are more suitable for facial expression recognition. Secondly, in the 3 × 3 neighborhood there is little difference in the recognition accuracy between the LGC and LGC-HD operators. Some studies also show that there is less information in the expression of the vertical gradient, but this does not mean that the information in the expression of the vertical gradient can be completely ignored. When the neighborhood is extended to a 5 × 5 grid, the influence of the vertical gradient information on the recognition results increases with the increase of information. However, adding the vertical gradient information directly to the operator increases the feature dimension. The vertical gradient information is added using the principle of centrosymmetric; hence, the CS-LGC operator fully considers the vertical gradient information, the gradient information between centrosymmetric relational pixels, and the relationship between the adjacent pixels. While guaranteeing a lower training dimension, the vertical gradient information is added to improve the recognition rate.
At the same time, we also find that the impact of the CS-LGC operator is more apparent in the JAFFE database, mainly because the facial expression gradient information in this database is rich, and some other restrictions, such as gender, race, and environment, affect together. However, from the overall performance of the operator, introducing the concept of centrosymmetric can extract more uniform and effective gradient information, which is helpful to further study the horizontal, vertical, and two diagonal directions of the neighborhood.

5. Conclusions

Overall, the CS-LGC operator has certain advantages in the field of facial expression recognition. Compared with other operators, the CS-LGC operator has better performance and generalization ability; its lower computational cost supports its broad use in the field of facial expression recognition. Compared with deep learning methods, traditional computing approaches offer significant performance advantages, as they are simple and effective. At the same time, we also recognize that the formation of facial expression recognition in practice is a dynamic process. Although this operator is suitable for static facial expression recognition, further research is needed to recognize dynamic expressions in video sequences.

Author Contributions

Conceptualization, J.Y. and D.S.P.; methodology, J.Y. and X.W.; software, S.H.; validation, X.W., J.W. and S.H.; formal analysis, X.W.; investigation, J.W.; resources, Y.W.; data curation, X.W.; writing—original draft preparation, X.W.; writing—review and editing, J.Y. and D.S.P.; visualization, X.W.; supervision, J.Y.

Funding

This research was funded by The Research Project of Tianjin Municipal Commission of Education, grant number 2017KJ033.

Acknowledgments

This work is supported by The Research Project of Tianjin Municipal Commission of Education No. 2017KJ033.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Mehrabian, A. Communication without words. In Communication Theory; Transaction Publishers: Piscataway, NJ, USA, 2008; pp. 193–200. [Google Scholar]
  2. Stefanos, E.; Ognjen, R.; Maja, P. Discriminative shared Gaussian processes for multiview and view-invariant facial expression recognition. IEEE Trans. Image Process. 2014, 24, 189–204. [Google Scholar]
  3. Shan, C.; Gong, S.; Mcowan, P.W. Facial expression recognition based on Local Binary Patterns: A comprehensive study. Image Vis. Comput. 2009, 27, 803–816. [Google Scholar] [CrossRef] [Green Version]
  4. Jeong, M.; Ko, B.C. Driver’s Facial Expression Recognition in Real-Time for Safe Driving. Sensors 2018, 18, 4270. [Google Scholar] [CrossRef] [PubMed]
  5. Guo, Y.; Yu, L.; Oerlemans, A.; Lao, S.; Song, W.; Lew, M.S. Deep learning for visual understanding: A review. Neurocomputing 2016, 187, 27–48. [Google Scholar] [CrossRef]
  6. Evangelos, S.; Hatice, G.; Andrea, C. Automatic Analysis of Facial Affect: A Survey of Registration, Representation, and Recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2015, 37, 1113–1133. [Google Scholar]
  7. Jabid, T.; Kabir, M.H.; Chae, O. Robust Facial Expression Recognition Based on Local Directional Pattern. ETRI J. 2010, 32, 784–794. [Google Scholar]
  8. Dahmane, M.; Meunier, J. Emotion recognition using dynamic grid-based HoG features. In Proceedings of the IEEE International Conference on Automatic Face & Gesture Recognition & Workshops, Santa Barbara, CA, USA, 21–25 March 2011. [Google Scholar]
  9. Perikos, I.; Paraskevas, M.; Hatzilygeroudis, I. Facial expression recognition using adaptive neuro-fuzzy inference systems. In Proceedings of the 2018 IEEE/ACIS 17th International Conference on Computer and Information Science (ICIS), Singapore, 6–8 June 2018; pp. 1–6. [Google Scholar]
  10. Zhang, W.; Shan, S.; Wen, G.; Chen, X.; Zhang, H. Local Gabor Binary Pattern Histogram Sequence (LGBPHS): A novel non-statistical model for face representation and recognition. In Proceedings of the Tenth IEEE International Conference on Computer Vision, Beijing, China, 17–21 October 2005. [Google Scholar]
  11. Berretti, S.; Amor, B.B.; Daoudi, M.; Bimbo, A.D. 3D facial expression recognition using SIFT descriptors of automatically detected keypoints. Vis. Comput. 2011, 27, 1021. [Google Scholar] [CrossRef]
  12. Bettadapura, V. Face expression recognition and analysis: The state of the art. arXiv 2012, arXiv:1203.6722. [Google Scholar]
  13. Mavani, V.; Raman, S.; Miyapuram, K.P. Facial Expression Recognition Using Visual Saliency and Deep Learning. In Proceedings of the IEEE International Conference on Computer Vision Workshop, Munich, Germany, 8–14 September 2018. [Google Scholar]
  14. Oyedotun, O.K.; Demisse, G.; Shabayek, A.E.R.; Aouada, D.; Ottersten, B. Facial Expression Recognition via Joint Deep Learning of RGB-Depth Map Latent Representations. In Proceedings of the IEEE International Conference on Computer Vision Workshop, Munich, Germany, 8–14 September 2018. [Google Scholar]
  15. Tong, Y.; Chen, R.; Cheng, Y. Facial expression recognition algorithm using LGC based on horizontal and diagonal prior principle. Opt.-Int. J. Light Electron Opt. 2014, 125, 4186–4189. [Google Scholar] [CrossRef]
  16. Shen, S.; Si, S.S. Facial expression recognition based on LGC in 5 × 5 neighborhood. Intell. Comput. Appl. 2017, 7, 47–48. [Google Scholar]
  17. Turk, M.A.; Pentland, A.P. Face recognition using eigenfaces. In Proceedings of the 1991 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Lahaina, HI, USA, 3–6 June 1991; pp. 586–591. [Google Scholar]
  18. Huang, G.B.; Zhu, Q.Y.; Siew, C.K. Extreme learning machine: A new learning scheme of feedforward neural networks. In Proceedings of the IEEE International Joint Conference on Neural Networks, Montreal, QC, Canada, 31 July–4 August 2005. [Google Scholar]
  19. Liu, Z.T.; Sui, G.T.; Li, D.Y.; Tan, G.Z. A novel facial expression recognition method based on extreme learning machine. In Proceedings of the 2015 34th Chinese Control Conference (CCC), Hangzhou, China, 28–30 July 2015; pp. 3852–3857. [Google Scholar]
  20. Lyons, M.J.; Akamatsu, S.; Kamachi, M.; Gyoba, J. Coding facial expressions with Gabor wavelets. In Proceedings of the IEEE International Conference on Automatic Face & Gesture Recognition, Washington, DC, USA, 20–21 May 2002. [Google Scholar]
  21. Lucey, P.; Cohn, J.F.; Kanade, T.; Saragih, J.; Ambadar, Z.; Matthews, I. The Extended Cohn-Kanade Dataset (CK+): A complete dataset for action unit and emotion-specified expression. In Proceedings of the Computer Vision & Pattern Recognition Workshops, San Francisco, CA, USA, 13–18 June 2010. [Google Scholar]
Figure 1. The example LBP operator.
Figure 1. The example LBP operator.
Sensors 19 01899 g001
Figure 2. Sample mask for 3 × 3 operators.
Figure 2. Sample mask for 3 × 3 operators.
Sensors 19 01899 g002
Figure 3. Sample mask for 5 × 5 operators.
Figure 3. Sample mask for 5 × 5 operators.
Sensors 19 01899 g003
Figure 4. Symmetrical local texture coding-based FER system block diagram.
Figure 4. Symmetrical local texture coding-based FER system block diagram.
Sensors 19 01899 g004
Figure 5. All kinds of algorithm result graphs.
Figure 5. All kinds of algorithm result graphs.
Sensors 19 01899 g005
Figure 6. The design of the central symmetric LGC (CS-LGC).
Figure 6. The design of the central symmetric LGC (CS-LGC).
Sensors 19 01899 g006
Figure 7. CS-LGC operator operation example.
Figure 7. CS-LGC operator operation example.
Sensors 19 01899 g007
Figure 8. CS-LGC feature extraction process.
Figure 8. CS-LGC feature extraction process.
Sensors 19 01899 g008
Figure 9. The examples of CK+ face database.
Figure 9. The examples of CK+ face database.
Sensors 19 01899 g009
Figure 10. Recognition rate (%) of different operators under the CK+ database.
Figure 10. Recognition rate (%) of different operators under the CK+ database.
Sensors 19 01899 g010
Figure 11. ROC curves comparison of different methods on CK+ database.
Figure 11. ROC curves comparison of different methods on CK+ database.
Sensors 19 01899 g011
Figure 12. The examples of JAFFE face database [11].
Figure 12. The examples of JAFFE face database [11].
Sensors 19 01899 g012
Figure 13. Recognition rate (%) of different operators under the JAFFE database.
Figure 13. Recognition rate (%) of different operators under the JAFFE database.
Sensors 19 01899 g013
Figure 14. ROC curves comparison of different methods on JAFFE database.
Figure 14. ROC curves comparison of different methods on JAFFE database.
Sensors 19 01899 g014
Table 1. Recognition rate of different block manner using CS-LGC operator on CK+ database.
Table 1. Recognition rate of different block manner using CS-LGC operator on CK+ database.
Block Manner1 × 21 × 42 × 22 × 44 × 24 × 4
Recognition rate (%)97.7898.3397.7896.1197.7896.11
Table 2. Recognition rate (%) of LGC-FN, LGC-AD and CS-LGC under the CK+ database.
Table 2. Recognition rate (%) of LGC-FN, LGC-AD and CS-LGC under the CK+ database.
Number of Training Samples (N)Operators
LBPCSLBPLGCLGC-HDLGC-FNLGC-ADCS-LGC
N = 636.1136.7337.0435.8037.9638.8939.51
N = 843.9144.5544.8744.2345.1945.5145.83
N = 1049.0051.6751.6748.0053.3353.3354.33
N = 1256.2557.6457.6457.6458.6859.0360.07
N = 1464.8667.3966.3065.2264.8667.0367.75
N = 1676.1476.8976.8976.5277.2776.8978.41
N = 1885.7186.5187.7087.3087.3088.1088.49
N = 2096.4797.0897.5097.0897.9297.9298.33
Table 3. Recognition rate (%) of different operators under the JAFFE database.
Table 3. Recognition rate (%) of different operators under the JAFFE database.
Number of Training Samples P (One Copy Contains Three Samples)Operators
LBPCSLBPLGCLGC-HDLGC-FNLGC-ADCS-LGC
P = 134.9239.1539.3740.7441.3740.7441.90
P = 248.1051.9052.7352.9854.3754.7655.48
P = 357.8263.9569.2570.7571.2972.7973.47
P = 465.2472.3874.7676.1979.5279.3779.84
P = 566.4873.5279.2479.0579.4380.0080.95
P = 673.0980.0082.1483.3384.5285.7186.66
P = 775.2381.9086.0388.9890.4890.4891.11
P = 877.1486.6790.4890.4890.9690.9691.91
P = 984.7688.5788.5790.4892.3891.4395.24
Table 4. The processing time of different algorithms.
Table 4. The processing time of different algorithms.
AlgorithmsLBPCSLBP LGCLGC-HDLGC-FNLGC-ADCS-LGC
Processing Time (s)0.0170.0190.0160.0160.0140.0320.015

Share and Cite

MDPI and ACS Style

Yang, J.; Wang, X.; Han, S.; Wang, J.; Park, D.S.; Wang, Y. Improved Real-Time Facial Expression Recognition Based on a Novel Balanced and Symmetric Local Gradient Coding. Sensors 2019, 19, 1899. https://doi.org/10.3390/s19081899

AMA Style

Yang J, Wang X, Han S, Wang J, Park DS, Wang Y. Improved Real-Time Facial Expression Recognition Based on a Novel Balanced and Symmetric Local Gradient Coding. Sensors. 2019; 19(8):1899. https://doi.org/10.3390/s19081899

Chicago/Turabian Style

Yang, Jucheng, Xiaojing Wang, Shujie Han, Jie Wang, Dong Sun Park, and Yuan Wang. 2019. "Improved Real-Time Facial Expression Recognition Based on a Novel Balanced and Symmetric Local Gradient Coding" Sensors 19, no. 8: 1899. https://doi.org/10.3390/s19081899

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop