Next Article in Journal
The Research of Relationship among Smile Developing Software, Internet Addiction, and Attachment Style
Next Article in Special Issue
Some Algorithms for Computing Short-Length Linear Convolution
Previous Article in Journal
A Deep Reinforcement Learning-Based Power Resource Management for Fuel Cell Powered Data Centers
Previous Article in Special Issue
Robust Room Impulse Response Measurement Using Perfect Periodic Sequences for Wiener Nonlinear Filters
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Optical Flow Filtering-Based Micro-Expression Recognition Method

1
School of Software, Nanchang University, Nanchang 330047, China
2
Department of Safety Management, Jiangxi Police College, Nanchang 330004, China
*
Author to whom correspondence should be addressed.
Electronics 2020, 9(12), 2056; https://doi.org/10.3390/electronics9122056
Submission received: 28 October 2020 / Revised: 23 November 2020 / Accepted: 1 December 2020 / Published: 3 December 2020
(This article belongs to the Special Issue Theory and Applications in Digital Signal Processing)

Abstract

:
The recognition accuracy of micro-expressions in the field of facial expressions is still understudied, as current research methods mainly focus on feature extraction and classification. Based on optical flow and decision thinking theory, we propose a novel micro-expression recognition method, which can filter low-quality micro-expression video clips. Determined by preset thresholds, we develop two optical flow filtering mechanisms: one based on two-branch decisions (OFF2BD) and the other based on three-way decisions (OFF3WD). In OFF2BD, which use the classical binary logic to classify images, and divide the images into positive or negative domain for further filtering. Differ from the OFF2BD, OFF3WD added boundary domain to delay to judge the motion quality of the images. In this way, the video clips with low degree of morphological change can be eliminated, so as to directly improve the quality of micro-expression features and recognition rate. From the experimental results, we verify the recognition accuracy of 61.57%, and 65.41% for CASMEII, and SMIC datasets, respectively. Through the comparative analysis, it shows that the scheme can effectively improve the recognition performance.

1. Introduction

Micro-expressions are always subtle and unconscious, which the duration of facial expression action is less than 0.2 s [1]. Micro-expressions can mine the true emotional state hidden in the human and demonstrates important application value in medical, polygraph, legal case and other fields. Therefore, micro-expressions recognition has gradually become a hot research issue. In early period, Ekman et al. [2] designed a Facial Action Coding System (FACS) based on the correlation between facial morphology changes to facilitate in-depth research in the field of it, and each of facial action coding is called an action unit (AU). After that, a lot of relative works have been researched on micro-expressions with the gradual maturity of artificial intelligence technology. However, the dataset of micro-expressions is difficult to collect because of the short duration and unconscious characteristics. Therefore, many micro-expression datasets have been constructed for the convenience of researchers for better in-depth research, such as SMIC [3], CASME II [4], SAMM [5], CAS(ME)2 [6], etc.
The main research idea of traditional micro-expression is to mine micro-expression information by extracting the features of the image combined with the classifier method. Pfister et al. [7] proposed an automatic micro-expression recognition system based on a temporal interpolation model and a spontaneous micro-expression corpus. This work is one of the earliest to realize automatic micro-expression recognition. Domestically, as representatives of micro-expression research and development, Fu et al. [8,9] made a lot of efforts to facilitate further research on micro-expression. With the gradual maturity of deep learning technology, researchers at home and abroad began to use deep learning methods [10,11,12] for micro-expression image coding and micro-expression recognition. Those methods used CNN and extended methods to replace the traditional texture feature extraction method for feature extraction and recognition.
According to the research, it can be found that the current research on micro-expression recognition at home and abroad is mainly related to the following two links: The first is how to extract high-quality texture features, and the second is to choose which classifier has a good classification effect. However, there are several problems with the current research thinking. 1. Traditional research methods mainly focus on the internal connection between the feature value in the images. Actually, the micro-expression is essentially the combined behavior of the morphological semantics of the various organs of the face, so that the influence of the morphological semantic quality of micro-expression is ignored. 2. Because of the small amount in the micro-expression dataset, the potential of deep learning method in the field of micro expression can’t be fully exploited, which also has low interpretability as a brute force method. In response to this problem, this paper introduces the decision-making thinking of rough set theory, which proposes an optical flow filtering method based on optical flow characteristics. This method eliminates video clips with low morphological change to improve the feature expression ability of the micro-expression to be recognized, thereby achieving the purpose of improving the performance of micro-expression recognition.
Therefore, the main contributions are as follows:
  • From the morphological change perspective, we propose a novel image weighting method based on optical flow to assign semantic quality to images between video clips;
  • Two optical flow filtering algorithms are proposed based on two-branch decisions and three-way decisions and a framework of micro-expression recognition, which establishes an interpretable and robust micro-expression recognition theory;
  • Compared with the state-of-the-art solution, the experiment proves the superiority of the effectiveness and efficiency of this method.
The rest of the paper is organized as follows: Section 2 conduct research work on related technologies. Section 3 describes the details of our method. Section 4 reports and discusses the experimental results. Section 5 gives conclusion and future work.

2. Related Works

Compared with macro-expression, the micro-expression has the characteristics of short duration and low intensity of facial movements, which leads to greater challenges in the facial features of micro-expressions. For the research on micro-expressions, the current research can be summarized as end-to-end research mode, which mainly focus on pre-processing, diversified feature extraction, and joint classifiers to micro-expression recognition. Nowadays, more and more new research methods are proposed with the maturity of science and technology.
I. Face detection and alignment. Face detection is essential for many facial applications (such as facial recognition and facial expression analysis). Accurate facial recognition through face detection methods can effectively reduce the impact of noise on micro-expressions in human face. The current face detection methods mainly include: 1. Use feature points to align faces and determine key points in the face area; 2. Use deep learning methods for face detection.
Cootes et al. [13] proposed the active shape model (ASM), which uses a 68-point subjective shape model to locate the key points of the human face to obtain the key points of the face area. This is an early widely used method for aligning face parts. Through the improvement of ASM, Cootes et al. [14] further proposed the active appearance model (AAM), which not only includes the establishment of shape models, but also the establishment of texture models and hybrid models. Subsequently, Cootes et al. [15] proposed a constrained local model (CLM), which abandoned the global texture method of the AAM algorithm, and modeled the local area in the neighborhood of candidate matching feature points and Space position is constrained.
With the maturity of deep learning technology, deep learning is widely applied in different fields [16,17]. Yang et al. [18] proposed a cascaded structure to detect tiny faces in the complex environments. Zhang et al. [19] exploits the inherent correlation between the tasks of detection and alignment, and combined the CNN to proposed MTCNN. MTCNN utilizes a cascaded multi-task framework to boost up the performance of both face detection and alignment. In addition, fast R-CNN [20] or faster R-CNN [21] has also been applied to face detection, which has good improvement in detection speed.
II Feature extraction and derivation. Feature extraction has been a hot issue in the field of micro-expression research. The performance of micro-expression recognition is mainly reflected in the extraction of texture features according to current research. Ojala et al. [22] proposed a kind of binary vector encoding (LBP) of local pixels, and describes the texture information of the current image by generating a histogram. LBP generally used for feature extraction of static image. Expand based on LBP operator, Zhao et al. [23] proposed a dynamic spatiotemporal feature operator LBPTOP. Compared with LBP Operator, LBPTOP takes into account the influence of video spatiotemporal information, thus adding two new feature planes XT and YT of projection on T time axis. Based on the above two classic operators, a variety of feature extraction variant operators have been derived. Huang et al. [24] proposed a spatiotemporal Local Binary Pattern with Integral Projection (STLBP-IP), which uses a differential image-based integral projection method to obtain appearance and motion features in horizontal and vertical directions. Subsequently, Huang et al. [25] proposed a spatial and temporal complete local quantization module (STCLQP) for facial micro-expression analysis, which optimizes the efficiency of micro-expression recognition by extracting three levels of feature vectors of pixel coincidence, amplitude and direction. With the further research, the team [26] once again proposed the local binary mode based on the spatiotemporal radon transform. Guo et al. [27] proposed an extended local binary on the tri-orthogonal plane Value model (ELBPTOP). In order to improve the efficiency of dynamic feature extraction, the LBPSIP [28] (Local Binary Patterns with Six Intersection Points) operator is proposed through the optimization on the LBP-TOP operator. This method transforms the histogram features of the three planes of LBPTOP into a two-dimensional vector encoding between 6 points intersecting XY, XT and YT, and uses the generated histogram as the final dynamic texture feature. Guo et al. [29] proposed to replace the classic LBPTOP operator with CBPTOP operator for image feature extraction, and then use ELM for feature recognition of micro-expressions.
III Research on micro-expression framework. With the gradual maturity of feature operators, in order to break this solidified research model, researchers began to conduct in-depth research on the recognition framework of micro-expression. Jia et al. [30] used the LBP operator to extract the macro-information and the LBP-TOP operator to extract the micro-information to construct the conversion model of the micro expression macro and micro information. Yan et al. [31] restricted local model (CLM) algorithm combined with local binary mode (LBP) and optical flow (of) to identify and analyze the correlation between the frames of face region (ROIs). Wang et al. [32] combining the weights calculated by the cumulative optical flow with the spatiotemporal features extracted by LBPTOP obtained the weighted spatiotemporal features, and the weighted features are input into a support vector machine for micro-expression recognition. Wang et al. [33] convert a color video clip into a four-dimensional vector, where the first two dimensions are spatial information, the third dimension is temporal information, and the fourth dimension is color information. Liong et al. [34] extracted the features of the vertex frame by selecting the Apex frame and the offset frame image of each video and combining the bi-directional weighted optical flow (Bi-WOOF).
IV Application of decision thinking theory. The decision thinking in rough set theory is the core idea of two-branch decision and three-way decision. Two-branch decision emphasizes the right or wrong of things and is absolutely. However, the three-way decision is a decision-making model based on human cognition, which believes that people can immediately make judgments about things that are fully certain to accept or reject in the actual decision-making process. Additionally, for things that cannot be made immediately, people tend to postpone the judgment of the event. Yang et al. [35] applied the three-way decision in medical diagnosis. Liang et al. [36] introduced the method of two-branch decision and three-way decision into the field of clustering, and defined a cluster calculation method based on clustering.
In addition, there are analysis some state-of-the-art-review in the fields of micro-expression, as shown in Table 1. According to the above research results, this paper combines multiple decision-making methods with classic feature extraction methods to construct a new micro-expression research framework.

3. Proposed Methods

Our proposed method consists of three main phases: Weight generation phase, optical flows filtering phase and micro-expression recognition phase. The main framework of our method is shown in Figure 1.

3.1. Weight Generation Phase

The analysis of SMIC [3] and CASMEII [4] can be found that the video clips contains a large number of images with no expression and low morphological semantic change. The region of the expression is mainly focus on the beginning of Onset-Frame to the end of Offset-Frame, and the peak of semantics is reached in Apex-Frame. Therefore, the features extracted through low motion semantic images will result in a large amount of low-quality and imbalance data. In order to conquered the problem, this paper proposes a low motion semantic image filtering method based on optical flow and decision thinking theory, which is used for the experimental preprocessing of video clips.
A pre-processing step was needed to ensure that have common features because of the resolution and other external features of each dataset used in experiments are various. First, Multi-task Cascaded Convolutional Networks (MTCNN) [19] module was adopted to detect face areas from the input images. Second, it crops only facial area from the entire image to make them the same size and not affected by unnecessary parts such as hair or accessories. Additionally, the cropped face images are resized to a size of 300 × 250. In this manner, the variations of spatial appearance among different video clips are normalized. Subsequently, the dense optical flows [42] are extracted from the aligned video clips to represent motion information of each pixel in micro-expression video clips. The optical flows ξ o i ( x , y ) are computed from each pair of neighboring images. When input t-frames video V   = { v 1 , v 2 , , v t   } into the method, t 1 frames optical flows O = { o 1 , o 2 , , o t 1 } will be obtained, Where o i is the optical flow image between v i and v i + 1 . The optical flows ξ o i ( x , y ) consist of horizontal displacements and vertical displacements. Therefore, the optical flows ξ o i ( x , y ) can be expressed as:
ξ o i ( x , y ) = ( h o i ( x , y ) , v o i ( x , y ) )
where h o i ( x , y ) and v o i ( x , y ) indicate the horizontal and vertical displacements respectively.
According to the optical flow of obtaining t 1 frames, for each image of optical flow, the Euclidean distance of the horizontal and vertical displacement of each pixel is obtained as the motion intensity of the current pixel. Then, the optical flow motion intensity W o i of the entire optical flow image of W × H pixels can be calculated as:
W o i = x = 0 W y = 0 H h o i ( x , y ) 2 + v o i ( x , y ) 2
where o i are the i-frame optical flow and i = 1 , 2 , , t 1 , W represents for optical flow o i width, H represents for optical flow o i height.
For the optical flow o i in the set O , given a semantic quality metric W ¯ o i called the optical flow average motion intensity to measure the quality of o i . W ¯ o i represents the average unit pixel semantic quality of the optical flow o i w.r.t the v i to v i + 1 .The calculation formula as shown in Equation (3):
W ¯ o i = W o i W × H
where W o i represents the motion intensity of i -frame optical flow, W represents for optical flow o i width, H represents for optical flow o i height.
Therefore, the motion intensity of optical flow can be measured by the average motion intensity value of optical flow in Equation (3). The average motion intensity reflects the strength of each group of neighboring video images. That is, the larger the average motion intensity of the optical flow, the larger the semantic quality between neighboring video frames. Namely, the smaller the average motion intensity of the optical flow, the smaller the semantic quality between neighboring video frames. Therefore, the average motion intensity of optical flow can be used to measure semantic quality information.
In order to better classify the semantic quality, the semantic quality of optical flow is evenly distributed in [ 0 , 1 ] by defining a normalization weighting formula. That is to say, the larger the average motion intensity of the optical flow, the larger the normalized semantic weight value; on the contrary, the smaller the average motion intensity of the optical flow, the smaller the normalized semantic weight value. For each optical flow frame of the average motion intensity W ¯ o i , the evaluation weighting of optical flow weighting filtering algorithm as shown in Equation (4).
ω o i = e W ¯ o i θ + 1
where θ is a parameter to adjust the influence of the weighted over the index.
Therefore, all the video frames obtained the semantic quality weight information ω o i ( i = 1 , 2 , , t 1 ) between neighboring video frames, the weight information represent the strength of the face area change between the neighboring video frames.

3.2. Optical Flows Filtering Phase

3.2.1. Optical Flows Filtering Algorithm Based on Two-Branch Decisions

The two-branch decisions have important application value in the choice of uncertainty about the data removal and retention. Therefore, determine the semantic quality of the video frames by defining the weight threshold δ ( δ ( 0 , 1 ) ) , which facilitates the classification of high-quality or low-quality semantic images. In solving uncertainty problems, classic binary logic embodies its important role. In distinguishing the motion intensity, the weight threshold is used to judge the semantic quality of the video frames, so that the optical flow weight between the video frames is divided into an object set P O S δ ( O ) or N E G δ ( O ) .
When ω o i   δ , the optical flow frame o i is divided into the positive domain of the set ω o i and this domain is denoted as P O S δ ( O ) ,recorded as o i P O S δ ( O )   ;
When ω o i < δ , the optical flow frame o i is divided into the negative domain of the set ω o i , and this domain is denoted as N E G δ ( O ) , recorded as o i N E G δ ( O ) .
Through the above method, the optical flow frames with high semantic change can be distinguished from the optical flow frames with low semantic change, so that the high semantic quality optical flow frames in set P O S δ ( O ) , and the low semantic quality optical flow frames in set N E G δ ( O ) .
With the average motion intensity weight ω o i ( i = 1 , 2 , , t 1 ) and threshold δ ( δ ( 0 , 1 ) ) , the motion intensity of the optical flow frame is divided into a set P O S δ ( O ) and a set N E G δ ( O ) . The more essential meaning is to distinguish images with small morphological changes between neighboring video frames from pictures with large morphological changes. This method uses the image of the first frame as the semantic benchmark to mine the morphological semantic quality of time series. Corresponding to this, the video frames V can be divided as follows according to the optical flow sets P O S δ ( O ) and N E G δ ( O ) as follows:
When o i P O S δ ( O ) , the video frame v i + 1 is divided into the high semantic quality domain, recorded as v i + 1 P O S δ ( V )   ;
When o i N E G δ ( O ) , the video frame v i + 1 is divided into the low semantic quality domain, recorded as v i + 1 N E G δ ( V )   ;
At this point, any video frame v i V in the video frames set V given semantic quality w.r.t V and based on optical flow weighting of the image semantic quality filtering algorithm is completed, which lays a solid foundation for the establishment of the morphological semantic database in Section 4 below. The pipeline of OFF2BD algorithm is illustrated in Algorithm 1.
Algorithm 1 Optical Flows Filtering Algorithm Based on Two-Branch Decisions (OFF2BD)
Input: video frames V   = { v 1 , v 2 , , v t   }, threshold δ
Output: High-quality semantic image set V after optical flow filtering
1. For v i = 1 , 2 , , | V | 1   /*Get initial optical flow set O */;
2. For v j = 2 , 3 , , | V | ;
3. Generate optical flow o i for v i and v j   ;
4. End for
5. End for
6. Obtain optical flow set O   ;
7. For o i = 1 , 2 , , | O |   /*Two-branch decisions iteration for optical flow in O */;
8. Calculate the average optical flow motion intensity W ¯ o i of optical flow o i w.r.t. Formulae (2) and (3);
9. Calculate the optical flow weight ω o i based on the average optical flow motion intensity W ¯ o i w.r.t. formula 4;
10. If ω o i     δ then;
11. put o i into P O S δ ( O )   ;
12. Else;
13. put o i into N E G δ ( O )   ;
14. End if;
15. End for;
16. Delete the optical flows set of N E G δ ( O ) and update the optical flow set O with P O S δ ( O )   ;
17. Update video set V according to the updated optical flow set O   ;
18. End for;
19. Output high-quality semantic image set V .

3.2.2. Optical Flows Filtering Algorithm Based on Three-Way Decisions

As the most important extension of the two-branch decisions, the three-way decisions [43] is an important granular calculation method derived from the rough set theory in machine learning, which is a new theory for solving uncertain problems proposed by the famous granular computing expert Yao [44] for the first time based on rough set theory and time research. Compared with the traditional two-branch decisions which requires the iterative incremental solution to solve the uncertainty of the problem, the three-way decision increases the delay behavior when the decision making is not accurate or cannot make accurate positive and negative decisions. The three-way decisions strategy not only reasonably explain the division behavior of the three decision domains (positive domain, negative domain and boundary domain) in the rough set theory, but also reduce the time consumption (abandoning incremental decision-making strategy (i.e., adding new knowledge constantly to make sure the uncertainty part turn to certain domain) in two-branch decisions.)
In dealing with uncertain problems, the three-way decisions method developed by the two-decision method is more suitable for human inertial thinking in solving problems. This approach is more representative of the problem-handling theory of contradictory relation and contrary relation problems. Therefore, in order to improve the semantic relationship between video frames and improve the quality of feature extraction, the quality thresholds α ( α ( 0 , 1 ) ) and β ( β ( 0 , 1 ) β > α ) of the video frames are defined for optical flow filtering. And then, the optical flows can be divided into an object set P O S ( α , β ) ( O ) or B N D ( α , β ) ( O ) or N E G ( α , β ) ( O ) .
When ω o i   β the optical flow frame o i is divided into the positive domain of the set ω o i and this domain is denoted as P O S ( α , β ) ( O ) , recorded as o i P O S ( α , β ) ( O )   ;
When α ω o i < β   , the optical flow frame o i is divided into the positive domain of the set ω o i and this domain is denoted as B N D ( α , β ) ( O ) ,recorded as o i B N D ( α , β ) ( O )   ;
When ω o i < α , the optical flow frame o i is divided into the negative domain of the set ω o i   , and this domain is denoted as N E G ( α , β ) ( O ) , recorded as o i N E G ( α , β ) ( O ) .
Three-way decisions methods for optical flow quality division are defined for ( α , β ) threshold, wherein the ( α , β ) are usually set to 0 < α < β < 1 . Compared with the two-branch decisions filtering method, this method adds the uncertainty domain B N D ( α , β ) ( O ) as a buffer area for uncertain semantic quality processing. Then, the B N D ( α , β ) ( O ) domain and the P O S ( α , β ) ( O ) domain are jointly used as the update domain, and iterative optical flow filtering analysis is performed until convergence is reached or a certain number of stops.
As the same time, the first frame of the image is used as a semantic reference combined with a convergent high-quality optical watershed to divide the collection of video frames. Corresponding to this, the video frames V can be divided as follows according to the optical flow set P O S ( α , β ) ( O )   , B N D ( α , β ) ( O ) and N E G ( α , β ) ( O ) as follows:
When o i P O S ( α , β ) ( O ) , the video frame v i + 1 is divided into the high semantic quality domain, recorded as v i + 1 P O S ( α , β ) ( V )   ;
When o i B N D ( α , β ) ( O ) , the video frame v i + 1 is divided into the fuzzy semantic quality domain, recorded as v i + 1 B N D ( α , β ) ( V )   ;
When o i N E G ( α , β ) ( O ) , the video frame v i + 1 is divided into the low semantic quality domain, recorded as v i + 1 N E G ( α , β ) ( V )   ;
When a specified number of iterative filtering or convergence is achieved, the P O S ( α , β ) ( V ) and B N D ( α , β ) ( V ) video frames set are jointly updated to the video frames set of the target domain V , and the video frame is reordered to obtain a high-quality semantic sequence video combination. The pipeline of OFF3WD algorithm is illustrated in Algorithm 2.
Algorithm 2 Optical flows filtering algorithm based on three-way decision (OFF3WD)
Input: video frames V   ={ v 1 , v 2 , , v t   }, threshold α , β   , iteration coefficient η   ;
Output: High-quality semantic image set V after optical flow filtering.
1. For k = 1 , 2 , , η   /*Filtering iteration for video frames in set V */;
2. For v i = 1 , 2 , , | V | 1   /*Get initial optical flow set O */;
3. For v j = 2 , 3 , , | V |   ;
4. Generate optical flow o i for v i and v j ;
5. End for
6. End for
7. Obtain optical flow set O   ;
8. For   o i = 1 , 2 , , | O |   /*Three-way decision iteration for optical flow in O */;
9. Calculate the average optical flow motion intensity W ¯ o i of optical flow o i w.r.t. Formulae (2) and (3);
10. Calculate the optical flow weight ω o i based on the average optical flow motion intensity W ¯ o i w.r.t. Formula (4);
11. If ω o i   β then;
12. put o i into P O S ( α , β ) ( O )   ;
13. If α ω o i < β then;
14. put o i into B N D ( α , β ) ( O ) ;
15. Else;
16. put o i into N E G ( α , β ) ( O )   ;
17. End if;
18. When | N E G ( α , β ) ( O ) | = 0 then;/*Reach the convergence condition*/
19. Combine P O S ( α , β ) ( O ) and B N D ( α , β ) ( O ) to update O   ;
20. Break;
21. Increase the α threshold naturally and delete the optical flows set of N E G ( α , β ) ( O )   ;
22. Combine P O S ( α , β ) ( O ) and B N D ( α , β ) ( O ) to update O   ;
23. Update video set V according to the updated optical flow set O   ;
24. End for;
25. End for;
26. Obtain the video frame set V according to the completed iteration condition or the converged optical flow set O   ;
27. Output high-quality semantic image set V   .

3.3. Micro-Expression Recognition Phase

3.3.1. Feature Extraction

In the field of micro-expression recognition, Local Binary Pattern (LBP) [22] is a common symbiotic pattern method that encodes local pixels in an image. The original LBP operator is defined as within the window of 3 × 3, taking the center pixel of the window as the threshold, and comparing the gray value of the adjacent 8 pixels with the center pixel in a clockwise direction, if the surrounding pixel value is greater than or equal to the center pixel value, the position of the pixel is marked as 1, otherwise it is 0. In this way, 8 pixels in the neighborhood of 3 × 3 can be compared to produce 8-bit binary numbers (usually converted to decimal numbers, that is, LBP codes, a total of 256 kinds), that is, the LBP value of the pixel at the center of the window is obtained, and this value is used to Reflect the texture information of this area. Such a 3 × 3 picture can be transformed into 256-dimensional texture features. The formula is shown in Equation (5).
H ( x c , y c ) = p = 0 p 1 2 p S ( i p i c )
S ( x ) = { 1 i f x 0 0 e l s e
where ( x c , y c ) represents the central pixel, i c is the gray value, i p is the gray value of the adjacent point, and S ( x ) is a function that conforms to Equation (6). According to the above formula method, the binary extraction process of the face picture is shown in Figure 2.
In order to solve the dynamic feature extraction, based on the LBP method proposed a dynamic feature extraction method LBPTOP (Local Binary Pattern from Three Orthogonal Planes). LBPTOP solves the problem that LBP can only extract two-dimensional features, and can be applied to feature extraction in three dimensions. LBPTOP extracts LBP features in three dimensions from the spatial plane XY and the spatial–temporal planes XT and YT respectively, and then combines them in series to form spatial-temporal features.
We use LBPTOP, LBPSIP, STCLQP, STLBPIP feature descriptors to encode the optical flow filtered video to express temporal and spatial information. These feature descriptors are all improved based on the LBPTOP method, and feature enhancement is performed on the three planes of the spatial plane XY and the spatial–temporal plane XT, YT. Therefore, the feature histograms of the based on the two-branch decisions spatial–temporal video V obtained after the optical flow filtering can be expressed as:
{ V δ = P O S δ ( V ) H δ = [ H V δ X Y , H V δ X T , H V δ Y T ]
where V δ represents the set of high-quality semantic frame spatiotemporal sequences obtained after the convergence of the OFF2BD algorithm. H δ are the functions to get the feature histograms from the V δ .
To further improve the morphological semantics quality of the spatial–temporal video V , the feature histograms based on OFF3WD can be expressed as:
{ V α , β = P O S α , β ( V ) B N D α , β ( V ) H α , β = [ H V α , β X Y , H V α , β X T , H V α , β Y T ]
where V α , β represents the set of high-quality semantic frame spatiotemporal sequences obtained after the convergence of the OFF3WD algorithm. H α , β are the functions to get the feature histograms from the V α , β .

3.3.2. Micro-Expression Recognition

For classic micro-expression feature recognition, Whitehill et al. [45] analyzed and compared it and found that SVM has relatively good recognition effect on micro-expression recognition compared with other traditional machine learning algorithms. FHOFO [37] uses linear SVM, 3point KNN and LDA to make correlation comparison among different micro-expression datasets. The experimental results show that compared with other classifiers, linear SVM has good generalization ability and stability. On the classification method, the leave-one-subject-out (LOSO) method was used to train and test the classifier. The LOSO-based validation is performed considering all images in the dataset. Given a dataset of n samples, the LOSO-based validation is performed with n iterations, such that in each iteration the classifier is trained with n-1 samples and tested on the remaining one sample. Using this strategy to evaluate the classification effect can more intuitively and accurately reflect the effect of the current experimental method. Therefore, this classification test method is currently used in most machine vision fields.

4. Experiments and Discussion

In this paper, The SMIC and CASME II will be used for optical flow filtering related experiments. All experiments in this paper are running on a PC of Windows 10, intel i7-8750H, 16G memory, python 3.7 environments, MATLAB 2016a and the python IDE environment is pyCharm professional 2019.3.

4.1. Datasets

The experiment selects two most popular spontaneous datasets, including CASME II and SMIC are used to evaluate the performance of the proposed method. The datasets statistics are summarized in Table 2.
SMIC [3]: SMIC was composed of 164 video clips of 16 subjects, which were divided into three different categories, such as Positive (51 samples), Negative (70 samples) and Surprise (43 samples). There are three versions of SMIC data: one is the high-speed camera (HS) of 100fps, the ordinary camera (VIS) version of 25 fps, and the near-infrared camera (NIR) version of 25 fps. HS cameras were used to record all data, while vis and NIR cameras were used to record data from the last eight subjects. Emotion classes are only based on participants’ self-report.
CASME II [4]: CASME II contains 247 micro-expression video clips from 26 subjects. All samples were recorded by high-speed cameras at 200 fps. The sampling resolution is 640 × 480 pixels. These samples were divided into five categories: Happiness (32 samples), Surprise (25 samples), Disgust (64 samples), Repression (27 samples) and Others (99 samples). Unlike SMIC, the AU label of CASME II follows the facial action coding system (FACS).

4.2. Experimental Setting

This section aims to verify the adaptability and performance of the method in this paper through relevant pairing and analysis experiments. MTCNN has good accuracy for face detection compared to traditional face alignment methods. Therefore, MTCNN is used for face positioning on the SMIC and CASME II data images, and the face region is cropped according to the final bounding box to obtain an average face area image size of 300 × 250.
Meanwhile, this article uses two optical flow filtering algorithms (OFF2BD, OFF3WD) to reduce the impact by low-quality semantic images. Through OFF2BD-LBPTOP, OFF3WD-LBPTOP, OFF2BD-LBPSIP, OFF3WD-LBPSIP, OFF2BD-STCLQP, OFF3WD-STCLQP, OFF2BD-STLBPIP, OFF3WD-STLBIP methods and classic micro-expressions recognition methods LBP-TOP, LBP-SIP, STCLQP, STLBP-IP for performance comparison. In order to verify the effectiveness of the optical flow filtering algorithm proposed in this paper on micro-expression recognition, several groups of experiments are designed as follows:
  • Sensitivity of the parameter θ to normalization function ω .
  • Sensitivity of the threshold δ to OFF2BD.
  • Sensitivity of the threshold ( α , β ) to OFF3WD.
  • The comparison of performance scores between refined algorithms (OFF2BD-LBPTOP, OFF3WD-LBPTOP, OFF2BD-LBPSIP, OFF3WD-LBPSIP, OFF2BD-STCLQP, OFF3WD-STCLQP, OFF2BD-STLBPIP, OFF3WD-STLBIP) and original algorithms (LBPTOP, LBPSIP, STCLQP, STLBPIP).
  • The confusion matrix of recognition accuracies for each micro-expression on CASMEII and SMIC dataset.
Choices of Parameters θ : The size of the cropped face image is equalized to 300 × 250 for SMIC and CASME II. A parameter θ is adopted in the computation of normalized weighting to adjust the influence of the magnitudes of optical flows over the index. Among then, through the image calculation of CSAMEII and SMIC the unit pixel magnitudes of each optical flow belong to the interval of [ 3 × 10 4 , 3.5 × 10 3 ] . Parameter θ setting range is (0,1), and its relationship with the normalized weight can be represented by Figure 3. As shown in Figure 3, when setting θ within a certain range values, e.g., setting θ < 0.005 , the Normalized Weight steady increase as the unit pixel magnitudes of optical flow increases and the difference between high-quality semantic optical flow weights and low-quality semantic weights is evenly expanded. When setting θ to larger values, the weight gap of high-quality semantic optical flow and that of low-quality semantic optical flow will be narrowed down, this will directly lead to the effect of the next optical flows weight filtering. Empirically, it is suggested that the parameter θ be set in the interval of [ 5 × 10 4 , 0.01 ] . Whereas θ = 0.001 , the weighted interval with a more uniform distribution will be obtained. Therefore, this paper chooses θ = 0.001 as the weight coefficient parameter of the experiment.
Performance Evaluation Index: Accuracy is a classic method to judge the performance of the model in machine learning. At present, most judgments on micro-expression recognition models are judged by recognition accuracy. The formula is as follows:
A c c = n * n * + m *
where n * is the number of correct micro-expressions predicted by the model, m * is the number of incorrect micro expressions predicted by the model, n * + m * is the total number of micro expressions in the micro-expression dataset.
F1 score (F1 Score) is an indicator used to measure the accuracy of the binary classification model in statistics, and it also takes into account the precision rate p and recall rate r of the classification model. Therefore, by calculating the recall rate r combined with the precision rate p to further calculate F1-score. In the prediction of the i -th micro-expression class, the number N i T P of true positive cases of the i -th micro-expression class that is predicted to be correct, the number N i F P of false positive cases that the i -th micro-expression class is predicted to be incorrect and the number N i F N of false negative cases that the non- i -th micro-expression class are predicted as i -th micro-expression class. The precision rate p i and the recall rate r i is calculated as follows:
p i = N i T P N i T P + N i F P
r i = N i T P N i T P + N i F N
where N i T P + N i F P is the total number of the real i -th micro-expression class, N i T P + N i F N is the number of i -th micro-expression class after prediction.
Calculation of micro-expression precision p and recall rate r of N ME classes, F1-score is defined as:
F = 1 N i = 1 N 2 p i × r i p i + r i
where p i and r i are the precision and recall of the i   -th micro-expression class.
The values of the above Accuracy and F1-score performance evaluation indicators are in the range [ 0 ,   1 ] , and the larger the indicator value, the better the recognition performance of the framework model for micro-expressions.

4.3. Experiments and Analysis

The LOSO method is used to the CASMEII and SMIC data are experimentally reproduced based on the LBP-TOP feature, and the selection of optimized relevant parameters value to further complete the parameter adjustment of the OFF2BD and OFF3BD algorithms. The video clips are divided to N × N × 2 blocks. The recognition results of micro-expressions are shown in Table 3, the best accuracy and F1-score for the CASME II dataset is achieved at 7 ×   7 blocks, the radius for the spatial plane (XY plane) and the radius for the spatial-temporal planes (XT and YT planes) is achieved at ( R x , R y , R t ) = ( 1 , 1 , 4 ) . Whereas the best accuracy and F1-score for the SMIC is achieved at 5 ×   5 blocks and ( R x , R y , R t ) = ( 2 , 2 , 3 ) .

4.3.1. Sensitivity of the Threshold δ to OFF2BD

In this part, the dataset is divided into train and test sets multiple times with different specifications by CASME II and SMIC. Through θ the comparison of the micro-expression recognition improvement rate obtained by the various threshold δ of multiple models in different dataset, and the robustness and generalization of the model of the OFF2BD algorithm under different thresholds are verified by the curve comparison of the test set and train set maps. LBPTOP and OFF2BD-LBPTOP are used as the benchmark method to compare the experimental algorithm. The improvement rate of the improved algorithm and the baseline algorithm is used as the evaluation index. The two-dimensional curve graph of threshold and improvement rate is fitted by many experiments on the data under different thresholds.
In this experiment, the CASME II and SMIC data are divided into multiple different dimensions of the dataset, and the effect of fitting the training data and test data of each dimension can be obtained, as shown in Figure 4. From the overall perspective, the analysis of the sets of data shows that the test data has a similar performance curve to the train data. In other words, the OFF2BD algorithm model has good generalization ability. From a partial analysis, when the threshold is at 0.5 ± 0.1 , the optical flow processing of OFF2BD algorithm has a good performance improvement. Among them, SMIC peaked around 0.43, and CASME II peaked around 0.5. When the threshold is ( 0.2 ,   0.4 ) , the micro-expression recognition performance gradually increases with the increase of the optical flow threshold, and when the threshold is greater than 0.55, the micro-expression recognition performance gradually decreases. Meanwhile, when the threshold is too high, as a large number of video clips are eliminated, some high-quality clips will be classified into negative domain, which will lead to a negative improvement rate values compared with the baseline. In summary, the OFF2BD algorithm further proves that the two-branch decision method is suitable for solving the uncertainty of image semantic quality.

4.3.2. Sensitivity of the Threshold ( α , β ) to OFF3WD

Through the analysis of the experimental results of OFF2BD, it is found that the threshold between (0.2, 0.4) has a significant improvement, and the performance gradually decreases after 0.55. The uncertainty domain BND is mainly used to solve the uncertainty of low-quality and high-quality data, and is used as a buffer domain of data to further distinguish the judgment after combining. Based on the above experience, this experiment chooses the threshold α ( 0.2 , 0.4 ) and threshold β ( 0.5 , 0.8 ) to experiment on the SMIC. Three-dimensional fitting is through the results generated by the OFF3WD algorithm, and the effect diagram is shown in Figure 5.
The α and β increase naturally according to the specified step length, and the range of the uncertainty domain is adjusted to select the optical flow intensity quality. As shown in Figure 5, the recognition rate of ME under the OFF3WD algorithm changes when the threshold range changes. Among them, it is a better effect improvement when the value of α is around 0.35 and the value of β is around 0.6. Additionally, the current threshold based on the OFF3WD algorithm also has a better improvement rate on CASME II ME recognition compared to the benchmark method.

4.3.3. Comparison with the State-of-the-Art Methods

LBPTOP was reproduced in this experiment and combined with the OFF2BD algorithm and the OFF3WD algorithm to obtain the ME recognition performance score. The experimental results are shown in Figure 6. The effects of OFF2BD-LBPTOP and OFF3WD-LBPTOP are significantly better than the traditional LBPTOP method on the test data. The accuracy has improved on the CASME II and SMIC about 2–3% by OFF2BD compared with the benchmark method. The OFF3WD algorithm has a more accurate judgment on the quality of optical flow semantics, which has increased by 0.5–2% compared with the OFF2BD algorithm. All in all, the experimental data preliminarily shows that the OFF2BD and OFF3WD algorithms can effectively mine the semantic information of micro-expressions and complete the image noise with low morphological semantic quality.
This section compares derivative algorithms (OFF2BD-LBPSIP, OFF3WD-LBPSIP, OFF2BD-STCLQP, OFF3WD-STCLQP, OFF2BD-STLBPIP, OFF3WD-STLBIP) and original algorithms (LBP-SIP, STCLQP, STLBP-IP) to further verify the improvement of the two filtering algorithms on the ME recognition. Two evaluation criteria (F1-score and Accuracy) were used to evaluate the performance scores. The experimental results are shown in Table 4. The proposed method achieves comparable recognition performances with the state-of-the-art methods on SMIC and CASME II. Compared with LBPSIP, the accuracy of OFF2BD and OFF3WD On CASME II is increased from 49.73% to 52.77% and 53.45%, which is 3–4% higher, and the F1 score is improved from 0.47 to 0.51. The accuracy of SMIC was improved by 2.94% and 4.15%, and the F1 score was improved from 0.46 to 0.5. Compared with STCLQP and STCLBPIP, the accuracy of the two algorithms on CASME II is improved by 2.48%, 2.89% and 2.06%, 1.02% respectively, and 0.82%, 1.39% and 1.79%, 3.90% on SMIC.
Among the feature methods of the current experiment, the two algorithms in this paper have achieved the greatest recognition accuracy and F1 score on STCLQP features. Especially, after three-way decision optical filtering, the recognition accuracy of STCLQP reached 65.41% on SMIC and 61.28% on CASME II, which have higher accuracies and F1-score than the state-of-the-art methods such as BI-WOOF and FHOFO. Simultaneously, the micro-expression feature obtained by the algorithm also has a good improvement effect on the F1-score.
In order to further evaluate the performance of the method in this paper to identify different micro-expressions, we perform optical flow filtering algorithm application and performance testing on the STCLQP features, and use the LOSO cross-validation method to generate a confusion matrix with the best recognition accuracy. Figure 7 and Figure 8 show the confusion matrix of OFF2BD-STCLQP and OFF3WD-STCLQP on SMIC and CASMEII respectively.
On the task of micro-expression detection, STCLQP can easily detect most micro-expression videos from non-micro-expression videos, and the OFF2BD and OFF3WD algorithms select high-quality semantic images, which further deepens the feature expression ability of STCLQP. For the micro-expression recognition on the SMIC database, OFF2BD has a corresponding improvement in the recognition accuracy of “positive” and “negative” classes compared with STCLQP, while the recognition of “surprise” shows a stable trend. OFF3WD further enhances the expression of morphological and semantic features of the “negative” class and make the recognition accuracy of the “negative” category reach 0.7. On the CASME II database, the micro-expression categories have become more complex and diverse, and the CASME II database contains five categories. From the confusion matrix of the CASMEII data set, it is found that the recognition and expression ability of the two algorithms proposed in this paper on STCLQP has generally increased. It shows the best improvement in identifying “happy” and “disgust”, and there are also minor improvements in “depressed”, “surprise” and “other”. Among the five classes, “depression” and “disgust” are the most difficult to identify micro-expression, while the classes of “surprise”, “happiness” and “other” are easy to identify. The algorithm improves the morphological semantic information of different classes, enhances the expression ability of different classes of features, and reduces the probability that the samples of “happiness”, “surprise”, “disgust” and “depression” are wrongly classified as “other” micro expression classes.

5. Conclusions

We propose an optical flow filtering micro-expression recognition algorithm (OFF2BD and OFF3WD) based on two-branch decision and three-way decision. The algorithm integrates the three-way decision method of rough set theory and adopts the three-way division idea of uncertain affairs, and improves the judgment of POS, BND and NEG domain conditions. The algorithm divides the morphological semantic quality of temporal images by redefining the weight value of optical flow to obtain a more robust and high-quality video images collection. We have seen that these two algorithms have improved their effects on the classic feature extraction methods. For STCLQP features, the optimized recognition rate and F1 score of the test are achieved on the SMIC data set. Optical flow filtering strengthens the feature intensity information on the time axis and eliminates images with low morphological semantic quality.
For future work, we pay more attention to the changes brought about by the facial action areas, and combine the idea of the OFF3WD algorithm to conduct semantic mining on the XY axis and the time axis T to reduce the impact of noise in different dimensions. In addition, not only limited to the method test of different features, we are designing the connection between different classifiers with our algorithm.

Author Contributions

Conceptualization, J.W. and J.X.; methodology, J.W., D.L., J.X. and M.T.; software, J.W. and D.L.; validation, J.W., D.L. and J.X.; formal analysis, D.L. and J.X.; investigation, J.W. and D.L.; resources, J.W. and J.X.; data curation, J.W. and J.X.; writing—original draft preparation, J.W.; writing—review and editing, D.L., J.X. and M.T.; visualization, J.W.; supervision, D.L. and J.X.; project administration, D.L. and J.X.; funding acquisition, J.X. and M.T. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by National Natural Science Foundation of China, grant number 61763031, Jiangxi Provincial Natural Science Foundation, grant number 20202BAB202018, Key Project of Jiangxi Provincial Department of Education Science and Technology Project “Public Security Big Data Analysis for Early Warning and Pre-control” under Grant GJJ191026.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Ekman, P.; Friesen, W.V. Nonverbal leakage and clues to deception. Psychiatry Interpers. Biol. Process. 1969, 32, 88–106. [Google Scholar] [CrossRef] [PubMed]
  2. Ekman, R. What the Face Reveals: Basic and Applied Studies of Spontaneous Expression Using the Facial Action Coding System (FACS); Oxford University Press: New York, NY, USA, 1997. [Google Scholar]
  3. Li, X.; Pfister, T.; Huang, X.; Zhao, G.; Pietikäinen, M. A spontaneous micro-expression database: Inducement, collection and baseline. In Proceedings of the 2013 10th IEEE International Conference and Workshops on Automatic Face and Gesture Recognition (FG), Shanghai, China, 22–26 April 2013; pp. 1–6. [Google Scholar]
  4. Yan, W.J.; Li, X.; Wang, S.J.; Zhao, G.; Liu, Y.J.; Chen, Y.H.; Fu, X. CASME II: An improved spontaneous micro-expression database and the baseline evaluation. PLoS ONE 2014, 9, e86041. [Google Scholar] [CrossRef] [PubMed]
  5. Davison, A.K.; Lansley, C.; Costen, N.; Tan, K.; Yap, M.H. SAMM: A spontaneous micro-facial movement dataset. IEEE Trans. Affect. Comput. 2016, 9, 116–129. [Google Scholar] [CrossRef] [Green Version]
  6. Qu, F.; Wang, S.J.; Yan, W.J.; Li, H.; Wu, S.; Fu, X. CAS(ME)2: A Database for Spontaneous Macro-Expression and Micro-Expression Spotting and Recognition. IEEE Trans. Affect. Comput. 2017, 9, 424–436. [Google Scholar] [CrossRef]
  7. Pfister, T.; Li, X.B.; Zhao, G.Y.; Pietikinen, M. Recognising spontaneous facial micro-expressions. In Proceedings of the 2011 IEEE International Conference on Computer Vision, Barcelona, Spain, 6–13 November 2011; pp. 1449–1456. [Google Scholar]
  8. Wang, S.J.; Yan, W.J.; Zhao, G.; Fu, X.; Zhou, C.G. Micro-Expression Recognition Using Robust Principal Component Analysis and Local Spatiotemporal Directional Features. In European Conference on Computer Vision; Springer: Cham, Switzerland, 2014; pp. 325–338. [Google Scholar]
  9. Wu, Q.; Shen, X.; Fu, X. The machine knows what you are hiding: An automatic micro-expression recognition system. In International Conference on Affective Computing and Intelligent Interaction; Springer: Berlin/Heidelberg, Germany, 2011; pp. 152–162. [Google Scholar]
  10. Khor, H.Q.; See, J.; Phan, R.C.W.; Lin, W. Enriched long-term recurrent convolutional network for facial micro-expression recognition. In Proceedings of the 2018 13th IEEE International Conference on Automatic Face and Gesture Recognition (FG 2018), Xi’an, China, 15–19 May 2018; pp. 667–674. [Google Scholar]
  11. Wang, S.J.; Li, B.J.; Liu, Y.J.; Yan, W.J.; Ou, X.; Huang, X.; Xu, F.; Fu, X. Micro-expression recognition with small sample size by transferring long-term convolutional neural network. Neurocomputing 2018, 312, 251–262. [Google Scholar] [CrossRef]
  12. Zhi, R.; Xu, H.; Wan, M.; Li, T. Combining 3D Convolutional Neural Networks with Transfer Learning by Supervised Pre-Training for Facial Micro-Expression Recognition. IEICE Trans. Inf. Syst. 2019, 102, 1054–1064. [Google Scholar] [CrossRef] [Green Version]
  13. Hill, A.; Cootes, T.F.; Taylor, C.J. Active shape models and the shape approximation problem. Image Vis. Comput. 1996, 14, 601–607. [Google Scholar] [CrossRef] [Green Version]
  14. Edwards, G.J.; Cootes, T.F.; Taylor, C.J. Advances in active appearance models. In Proceedings of the Seventh IEEE International Conference on Computer Vision, Kerkyra, Greece, 20–25 September 1999; pp. 137–142. [Google Scholar]
  15. Cristinacce, D.; Cootes, T. Automatic feature localisation with constrained local models. Pattern Recognit. 2008, 41, 3054–3067. [Google Scholar] [CrossRef]
  16. Yang, Z.; Leng, L.; Kim, B.G. StoolNet for Color Classification of Stool Medical Images. Electronics 2019, 8, 1464. [Google Scholar] [CrossRef] [Green Version]
  17. Leng, L.; Yang, Z.; Kim, C.; Zhang, Y. A Light-Weight Practical Framework for Feces Detection and Trait Recognition. Sensors 2020, 20, 2644. [Google Scholar] [CrossRef]
  18. Yang, Z.; Li, J.; Min, W.; Wang, Q. Real-Time Pre-Identification and Cascaded Detection for Tiny Faces. Appl. Sci. 2019, 9, 4344. [Google Scholar] [CrossRef] [Green Version]
  19. Zhang, K.; Zhang, Z.; Li, Z.; Qiao, Y. Joint face detection and alignment using multitask cascaded convolutional networks. IEEE Signal Process. Lett. 2016, 23, 1499–1503. [Google Scholar] [CrossRef] [Green Version]
  20. Sun, X.; Wu, P.; Hoi, S.C.H. Face detection using deep learning: An improved faster RCNN approach. Neurocomputing 2018, 299, 42–50. [Google Scholar] [CrossRef] [Green Version]
  21. Jiang, H.; Learned-Miller, E. Face Detection with the Faster R-CNN. In Proceedings of the 2017 12th IEEE International Conference on Automatic Face and Gesture Recognition (FG 2017), Washington, DC, USA, 30 May–3 June 2017; pp. 650–657. [Google Scholar]
  22. Ojala, T.; Pietikäinen, M.; Maenpaa, T. Multiresolution gray-scale and rotation invariant texture classification with local binary patterns. IEEE Trans. Pattern Anal. Mach. Intell. 2002, 24, 971–987. [Google Scholar] [CrossRef]
  23. Zhao, G.; Pietikäinen, M. Dynamic texture recognition using local binary patterns with an application to facial expressions. IEEE Trans. Pattern Anal. Mach. Intell. 2007, 29, 915–928. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  24. Huang, X.; Wang, S.J.; Zhao, G.; Piteikäinen, M. Facial micro-expression recognition using spatiotemporal local binary pattern with integral projection. In Proceedings of the IEEE International Conference on Computer Vision Workshops, Santiago, Chile, 11–18 December 2015; pp. 1–9. [Google Scholar]
  25. Huang, X.; Zhao, G.; Hong, X.; Zheng, W.; Pietikäinen, M. Spontaneous facial micro-expression analysis using spatiotemporal completed local quantized patterns. Neurocomputing 2016, 175, 564–578. [Google Scholar] [CrossRef]
  26. Huang, X.; Zhao, G. Spontaneous facial micro-expression analysis using spatiotemporal local radon-based binary pattern. In Proceedings of the 2017 International Conference on the Frontiers and Advances in Data Science (FADS), Xi’an, China, 23–25 October 2017; pp. 159–164. [Google Scholar]
  27. Guo, C.; Liang, J.; Zhan, G.; Liu, Z.; Pietikäinen, M.; Liu, L. Extended Local Binary Patterns for Efficient and Robust Spontaneous Facial Micro-Expression Recognition. IEEE Access 2019, 7, 174517–174530. [Google Scholar] [CrossRef]
  28. Wang, Y.; See, J.; Phan, W.; Oh, Y.H. LBP with Six Intersection Points: Reducing Redundant Information in LBP-TOP for Micro-expression Recognition. In Asian Conference on Computer Visio; Springer: Cham, Switzerland, 2014; pp. 525–537. [Google Scholar]
  29. Guo, Y.; Xue, C.; Wang, Y.; Yu, M. Micro-expression recognition based on CBP-TOP feature with ELM. Optik 2015, 126, 4446–4451. [Google Scholar] [CrossRef]
  30. Jia, X.; Ben, X.; Yuan, H.; Kpalma, K.; Meng, W. Macro-to-micro transformation model for micro-expression recognition. J. Comput. Sci. 2018, 25, 289–297. [Google Scholar] [CrossRef]
  31. Yan, W.J.; Chen, Y.H. Measuring dynamic micro-expressions via feature extraction methods. J. Comput. Sci. 2018, 25, 318–326. [Google Scholar] [CrossRef]
  32. Wang, L.; Xiao, H.; Luo, S.; Zhang, J.; Liu, X. A weighted feature extraction method based on temporal accumulation of optical flow for micro-expression recognition. Signal Process. Image Commun. 2019, 78, 246–253. [Google Scholar] [CrossRef]
  33. Wang, S.J.; Yan, W.J.; Li, X.; Zhao, G.; Zhou, C.G.; Fu, X.; Yang, M.; Tao, J. Micro-expression recognition using color spaces. IEEE Trans. Image Process. 2015, 24, 6034–6047. [Google Scholar] [CrossRef] [PubMed]
  34. Liong, S.T.; See, J.; Wong, K.S.; Phan, C.W. Less is more: Micro-expression recognition from video using apex frame. Signal Process. Image Commun. 2018, 62, 82–92. [Google Scholar] [CrossRef] [Green Version]
  35. Li, Z.; Xie, N.; Huang, D.; Zhang, G. A three-way decision method in a hybrid decision information system and its application in medical diagnosis. Artif. Intell. Rev. 2020, 25, 1–30. [Google Scholar] [CrossRef]
  36. Liang, W.; Zhang, Y.; Xu, J.; Lin, D. Optimization of Basic Clustering for Ensemble Clustering: An Information-theoretic Perspective. IEEE Access 2019, 7, 179048–179062. [Google Scholar] [CrossRef]
  37. Happy, S.L.; Routray, A. Fuzzy Histogram of Optical Flow Orientations for Micro-Expression Recognition. IEEE Trans. Affect. Comput. 2019, 10, 394–406. [Google Scholar] [CrossRef]
  38. Xia, Z.; Peng, W.; Khor, H.Q.; Feng, X.; Zhao, G. Revealing the Invisible with Model and Data Shrinking for Composite-Database Micro-Expression Recognition. IEEE Trans. Image Process. 2020, 29, 8590–8605. [Google Scholar] [CrossRef] [PubMed]
  39. Lu, H.; Kpalma, K.; Ronsin, J. Motion descriptors for micro-expression recognition. Signal Process. Image Commun. 2018, 67, 108–117. [Google Scholar] [CrossRef]
  40. Xu, F.; Zhang, J.; Wang, J.Z. Micro-expression identification and categorization using a facial dynamics map. IEEE Trans. Affect. Comput. 2017, 8, 254–267. [Google Scholar] [CrossRef]
  41. Xia, Z.; Hong, X.; Gao, X.; Feng, X.; Zhao, G. Spatiotemporal Recurrent Convolutional Networks for Recognizing Spontaneous Micro-Expressions. IEEE Trans. Multimed. 2020, 22, 626–640. [Google Scholar] [CrossRef] [Green Version]
  42. Black, M.J.; Anandan, P. The Robust Estimation of Multiple Motions: Parametric and Piecewise-Smooth Flow Fields. Comput. Vis. Image Underst. 1996, 63, 75–104. [Google Scholar] [CrossRef]
  43. Xu, J.; Miao, D.; Zhang, Y.; Zhang, Z. A three-way decisions model with probabilistic rough sets for stream computing. Int. J. Approx. Reason. 2017, 88, 1–22. [Google Scholar] [CrossRef]
  44. Yao, Y.Y. An outline of a theory of three-way decisions. In Rough Sets and Current Trends in Computing; Springer: Berlin/Heidelberg, Germany, 2012; pp. 1–17. [Google Scholar]
  45. Whitehill, J.; Littlewort, G.; Fasel, I.; Bartlett, M.; Movellan, J. Toward practical smile detection. IEEE Trans. Pattern Anal. Mach. Intell. 2009, 31, 2106–2111. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Framework of our proposed micro-expression recognition based on OFF3WD.
Figure 1. Framework of our proposed micro-expression recognition based on OFF3WD.
Electronics 09 02056 g001
Figure 2. Example of encoding a LBP feature.
Figure 2. Example of encoding a LBP feature.
Electronics 09 02056 g002
Figure 3. Correlation between normalized weight and unit pixel magnitudes of optical flows with different parameters θ .
Figure 3. Correlation between normalized weight and unit pixel magnitudes of optical flows with different parameters θ .
Electronics 09 02056 g003
Figure 4. The sensitivity of OFF2BD with various thresholds δ . (a) Represents the train-set; (b) represent the test-set.
Figure 4. The sensitivity of OFF2BD with various thresholds δ . (a) Represents the train-set; (b) represent the test-set.
Electronics 09 02056 g004
Figure 5. The sensitivity of OFF3WD with various thresholds ( α , β ) .
Figure 5. The sensitivity of OFF3WD with various thresholds ( α , β ) .
Electronics 09 02056 g005
Figure 6. Recognition accuracies of different algorithms.
Figure 6. Recognition accuracies of different algorithms.
Electronics 09 02056 g006
Figure 7. The confusion matrix of OFF2BD on SMIC and CASME II.
Figure 7. The confusion matrix of OFF2BD on SMIC and CASME II.
Electronics 09 02056 g007
Figure 8. The confusion matrix of OFF3WD on SMIC and CASME II.
Figure 8. The confusion matrix of OFF3WD on SMIC and CASME II.
Electronics 09 02056 g008
Table 1. Analysis of the state-of-the-art-review for micro-expression recognition.
Table 1. Analysis of the state-of-the-art-review for micro-expression recognition.
MethodsDescriptionAdvantage
STCLQP [25]Extracts the features of sign, magnitude and orientation components information and fused.For facial micro-expression appearance and motion feature extraction has a better effect.
TAOF [32]The motion intensity is calculated by accumulating the neighboring optical flow in the time interval, and the feature are further weighted locally.Enhanced the discrimination of the weighted features for micro-expression recognition.
BI-WOOF [34]Created a new feature extractor, which to encode essential expressiveness of the apex frame.Capable of representing the apexframe in a discriminative manner which emphasizes facial motion information at both bin and block levels.
FHOFO [37]Construct suitable angular histograms from optical flow vector orientations and using histogram fuzzification to encode the temporal pattern.Insensitive to change of illumination conditions among video clips and the extracted features are robust to the variation of expression intensities
RIMDS [38]Explored the shallower-architecture and lower-resolution input data, shrinking model and input complexities simultaneously for the composite-database.Achieved the superior performance compared to the micro-expression recognition approaches.
FMBH [39]Combing both the horizontal and the vertical components of the differential of optical flow as inspired from the motion boundary histograms (MBH).The unexpected motions, caused by residual mis-registration that appearsbetween images cropped from different frames, can be removed.
FDM [40]Characterize the movements of a micro-expression in different granularity.Provided an effective and intuitive understanding of human facial expressions.
STRCN [41]A novel deep recurrent convolutional networks based micro-expression recognition approach and capturing the spatiotemporal deformations of micro-expression sequence.Conquered the “limited and imbalanced training samples” problem and considered the spatiotemporal deformations.
Table 2. A summary of the different features of the SMIC and CASME II datasets.
Table 2. A summary of the different features of the SMIC and CASME II datasets.
FeatureSMIC-HSSMIC-VIS/NIRCASME II
subjects161635
samples16471247
Frames per sec.10025200
Posed/spontaneousSpontaneousSpontaneousSpontaneous
AU labelsnonoyes
Emotion class3(positive/negative/surprise)3(positive/negative/surprise)5(basic emotions)
TaggingEmotion categoryEmotion categoryEmotion/FACS
Table 3. The performance of different uniform LBP-TOP features on CASME II and SMIC.
Table 3. The performance of different uniform LBP-TOP features on CASME II and SMIC.
( R x , R y , R z ) CASME IISMIC
3 ×   3 5 ×   5 7 ×   7 3 ×   3 5 ×   5 7 ×   7
Acc (%)/F1-ScoreAcc (%)/F1-Score
(1,1,2)43.43/0.3446.88/0.3845.82/0.3849.42/0.4850.48/0.5045.58/0.45
(1,1,3)45.27/0.3647.04/0.3946.98/0.3948.62/0.4849.28/0.4947.85/0.48
(1,1,4)46.91/0.3947.65/0.4148.24/0.4247.57/0.4746.04/0.4543.25/0.42
(2,2,2)45.14/0.3645.32/0.3746.24/0.3943.61/0.4347.32/0.4649.54/0.51
(2,2,3)45.43/0.3344.92/0.3247.15/0.4047.76/0.4951.83/0.5144.38/0.47
(2,2,4)42.15/0.3246.31/0.3844.31/0.3745.17/0.4349.55/0.4743.47/0.45
(3,3,2)43.47/0.3543.28/0.3445.69/0.3746.04/0.4744.71/0.4247.85/0.49
(3,3,3)46.38/0.3746.93/0.3943.74/0.3546.41/0.4842.52/0.4146.12/0.48
(3,3,4)47.22/0.4145.82/0.3545.49/0.3645.79/0.4443.45/0.4544.91/0.44
Table 4. Comparison with different ME recognition methods.
Table 4. Comparison with different ME recognition methods.
MethodCASME IISMIC
Acc (%)F1-ScoreAcc (%)F1-Score
LBPTOP (reproduced)48.240.4251.830.51
LBPSIP [28]49.730.4747.420.46
STCLQP [25]58.390.5864.020.63
STCLBP-IP [24]59.51-57.93-
BI-WOOF [34]58.460.6061.290.61
FHOFO [37]60.520.6158.610.58
OFF2BD-LBPTOP50.460.4453.520.52
OFF2BD-LBPSIP52.770.5150.360.50
OFF2BD-STCLQP60.870.6264.840.63
OFF2BD-STCLBP-IP61.57-59.18-
OFF3WD-LBPTOP51.680.4654.350.54
OFF3WD-LBPSIP53.450.5151.570.52
OFF3WD-STCLQP61.280.6165.410.65
OFF3WD-STCLBP-IP60.53-61.83-
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Wu, J.; Xu, J.; Lin, D.; Tu, M. Optical Flow Filtering-Based Micro-Expression Recognition Method. Electronics 2020, 9, 2056. https://doi.org/10.3390/electronics9122056

AMA Style

Wu J, Xu J, Lin D, Tu M. Optical Flow Filtering-Based Micro-Expression Recognition Method. Electronics. 2020; 9(12):2056. https://doi.org/10.3390/electronics9122056

Chicago/Turabian Style

Wu, Junjie, Jianfeng Xu, Deyu Lin, and Min Tu. 2020. "Optical Flow Filtering-Based Micro-Expression Recognition Method" Electronics 9, no. 12: 2056. https://doi.org/10.3390/electronics9122056

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop