Next Article in Journal
Fully 3D Active Surface with Machine Learning for PET Image Segmentation
Next Article in Special Issue
Efficient Face Recognition System for Operating in Unconstrained Environments
Previous Article in Journal
An Automated Method for Quality Control in MRI Systems: Methods and Considerations
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Learning Descriptors Invariance through Equivalence Relations within Manifold: A New Approach to Expression Invariant 3D Face Recognition

by
Faisal R. Al-Osaimi
Al-Abediah Campus, Umm Al-Qura Univeristy, Taif Rd., Makkah 21955, Saudi Arabia
J. Imaging 2020, 6(11), 112; https://doi.org/10.3390/jimaging6110112
Submission received: 14 August 2020 / Revised: 18 October 2020 / Accepted: 19 October 2020 / Published: 22 October 2020
(This article belongs to the Special Issue 3D Human Understanding)

Abstract

:
This paper presents a unique approach for the dichotomy between useful and adverse variations of key-point descriptors, namely the identity and the expression variations in the descriptor (feature) space. The descriptors variations are learned from training examples. Based on labels of the training data, the equivalence relations among the descriptors are established. Both types of descriptor variations are represented by a graph embedded in the descriptor manifold. Invariant recognition is then conducted as a graph search problem. A heuristic graph search algorithm suitable for the recognition under this setup was devised. The proposed approach was tested on the FRGC v2.0, the Bosphorus and the 3D TEC datasets. It has shown to enhance the recognition performance, under expression variations, by considerable margins.

1. Introduction

3D face recognition has shown to achieve a considerable recognition accuracy and robustness, especially when compared to its 2D counterpart. There are vital applications of face recognition such as security, access control and human-machine interaction. The importance of its applications combined with the recent advances in the 3D digitization technologies have been the driving forces behind the interest in 3D face recognition among researchers. Despite reported advances in 3D face recognition in the recent years, the practical applications of 3D face recognition require even higher accuracies and robustness.
A particularly interesting recognition paradigm is concerned with the detection of key-points and then the extraction of descriptors from the local 3D surfaces around them. This paradigm inherently enjoys desirable properties such as the robustness to clutter and occlusions and the enabling of partial surface matching. While this paradigm has been very successful in the general object recognition [1,2,3], face recognition performance of the well-known approaches in this paradigm remained, until recently, below that of the state-of-the-art [4,5]. Recently, the rotation-invariant and adjustable integral kernel (RAIK) approach [6] which belongs to this paradigm has shown to highly perform in 3D face recognition when the matching is limited to the semi-rigid regions of the face (the nose and the forehead). The RAIK descriptors are discriminative, invariant to 3D rotations and completely representative of the underling surface. In the context of 3D face recognition, the descriptors also encode the expression variations. The 3D face variations pertaining to the identities of the individuals are essential for correct recognition. However, in practice these valuable variations mix with expression variations and rigid transformation, affecting the performance and robustness of face recognition.
When descriptors faithfully encode shape information of the local surfaces independently from the rigid transformations, it is imperative to assume that the local descriptors represent only the identity related shapes and the expression deformations of the local surfaces. One can consider expression deformations as displacements if descriptors, in the descriptor space, in an unknown and complex way. The differentiation between the two types of the descriptor displacements, namely the identity-induced displacements (IIDs) and the expression-induced displacements (EIDs), is learned and utilized to enhance recognition. The displacements between descriptors extracted from different scans of the same person and from the same location on the face (corresponding key-points) but under different facial expressions are basically EID. These displacements along with the displacements that set different individuals apart are extracted from training data and represented by a large graph embedded in the descriptor manifold for the utilization in the expression invariant face recognition. These displacements can be also perceived as one-to-one relations from the descriptor space to itself. In this terminology, the sub-graphs corresponding to the descriptors (the nodes) of the same individual under varying facial expressions represent equivalence relations (ERs) but the sub-graphs connecting the descriptors of different individuals represent identity relations (IRs).
In the rest of the paper, the term ensemble refers to the set of descriptors extracted from a 3D facial scan. The set of ensembles that are extracted from multiple 3D facial scans of a person is called a collection. Many training collections of descriptors are extracted from sets of 3D facial scans of many individuals. The 3D facial scans in each collection are under varying expressions, including the neutral expression. The descriptors of each ensemble are mapped (linked) to their corresponding descriptors in the other ensembles. This results in multiple sets of corresponded descriptors within each collection. Each one of them is then joined together by a simple spanning graph which represents the equivalence relations, ERs. Between the different collections, the corresponding equivalence graphs (ERs) are connected to each other, enabling the representation of the identity relations, IRs. One viable approach is to inter-connect the different equivalence graphs from the descriptors of a neutral ensemble to the descriptors of neutral ensembles in other collections (neutral to neutral connections). For invariant matching of a probe ensemble to a gallery ensemble, the descriptors are first corresponded. Then, dissimilarity measures between descriptors are found by searching the graph for the path connecting each corresponded descriptor pair such that the encountered IRs along the graph path have cost (as a vector quantity), the encountered ERs have zero-cost and the magnitude of the sum of all the encountered IRs is minimum. Finally, on the basis of the minimized IR quantities the dissimilarity measures are computed. See Figure 1 for a graphical illustration.

2. The Literature Review

The approaches to expression invariant 3D face recognition can be broadly categorized into two categories; the image space approaches and the feature space approaches. Dealing with 3D surfaces in the image space is more intuitive. It is not surprising to find early successful approaches in this category. The avoidance of the highly deformable regions of the face such as the mouth and the cheeks, e.g., [7,8,9,10], has shown to be considerably effective. Unfortunately, the considered regions of the face still undergo expression deformations which can adversely affect the recognition accuracy. Another class of methods in the same category is to deform the 3D facial surface in an attempt to remove or at least reduce the effects of the expression deformations. The approach proposed in [11], transforms the 3D facial surface by bending the surface while preserving the geodesic distances among its points into an invariant image. As that approach flattens the expressions, it also flattens some other facial surface geometries. Recently, several methods extract expression invariant measures from geodesic distances of the facial surface [12,13,14]. The annotated face model, which is an elastic deformable model [15], deforms a 3D face under one expression to a target 3D face under another expression [16]. The recognition is then performed on the fitted model. It is not clear how such model can differentiate between the expression deformations and the other unique geometries of the face. In another work [17], a low dimensional principal component analysis (PCA) subspace in which the expression deformations reside is found from the 3D shape residues of registered pairs of non-neutral and neutral scans, where each pair belong to the same subject. For invariant recognition, the expression deformations in the image space are separated from novel 3D residues based on their projection onto the subspace. These methods are better situated for global face matching. Therefore, their robustness and accuracy may be undermined under occlusions and wide pose variations. The work in [18], performs elastic invariant matching of local surface regions around a handful of manually picked facial anatomical landmarks. Nonetheless, reliable automatic localization of the facial landmarks remains an open problem.
In a wide range of approaches, the matching is performed based on extracted features [19,20,21,22,23,24]. Feature extraction reduces the dimensionality of the facial data. An independent set of features is preferable for the recognition. Some features may be less sensitive to the expressions than others, providing a limited level of expression invariance. Typical examples of feature extraction methods are the linear subspace methods such as the different variants of PCA [25,26,27,28,29], linear discriminant analysis (LDA) [26,28] and independent component analysis (ICA) [30,31,32].
In the feature space, there are different methods in the literature that attempt to find the space regions that are associated with certain identities or class labels. One widely used methods is to estimate a probability density function (PDF) over space regions. The estimation of a PDF in a multidimensional vector space faces practical challenges. The number of training data samples required to compute such PDF grows exponentially with the number of dimensions, referred to as “the curse of dimensionality” [33,34]. Strong assumptions about the PDF are often made, such as assuming normal distribution [35,36], to survive on the available training samples. The “kernel trick” method was used in several approaches feature space approaches, e.g., [37,38,39]. The kernel function replaces the dot product in their non-kernelized variants, e.g., the support vector machines k-SVMs [39] and k-PCA [40], they are often data driven. Its overall process can be considered as a nonlinear transformation of the feature space to induce more separable classes. While it provides an elegant approach for the non-linear separability, it does not explicitly address the expression variations. In contrast, the proposed work is explicit in handling expression variations and versatile in selecting the relevant subset of the driving data for each match.
The manifold methods [41,42,43,44,45] in most cases are concerned with the non-linear dimensionality reduction, where the feature distribution in the feature space may be locally of a much lower dimension than that of the manifold as a whole. The de facto example is a “Swiss roll” surface residing in a higher dimensional space. Typically, rather than using the Euclidean distances for matching, the geodesic distances on the manifold are used instead [46]. Sometimes the problem at consideration or its formulation guarantees that the feature data form a manifold of a specific low dimension and the availability of enough feature samples to recover the lower dimensional manifold. An example of such a problem is the manifold of a rotated template image [47], the manifold dimension in this case is the number of the degree of freedom and the data samples can be generated as needed. The expression and identity manifold is of a complex structure with several dimensions (may vary locally). This problem also lacks the availability of enough data samples to accurately recover the manifold. In contrast, the proposed does not attempt to recover the lower dimension of the manifold, unroll it, or extract geodesic distances. It perceives the manifold as a sparse distribution of data samples and the distances between certain points are shortened to zero.

3. The Proposed Approach

This section describes the steps of the proposed approach and discusses the concepts behind them in more details than previously provided.

3.1. Conceptual Analysis

Manifold has the notion of the Euclidean spaces (tangential spaces) locally at each of its points. Conventionally, a descriptor space is either treated as one Euclidean space or as a manifold. In both situations, the large distances in the descriptor space translate into large dissimilarity measures which is not plausible for the recognition under variations. In contrast, the proposed approach has the ability to bring and merge distant tangential manifold spaces with each other. Consequently, the proximity and the non-proximity among the manifold points can combine and contribute more meaningfully to the dissimilarity measure. This can be achieved through the establishment of equivalence relations between reference points (corresponded descriptors) in the descriptor manifold. Let Q = { D 1 , , D n } be a set of corresponded descriptors (an equivalence set) and T D i M denotes a tangential space at the i-th descriptor of the manifold, M. Each corresponded pair of descriptors, { D i , D j } , establishes an equivalence between all the tangential spaces in which D i and D j exist. The equivalency also extends to the manifold points around the descriptors. The same concept similarly applies for all the pair combinations of the descriptors in Q . This gives rise to the notion of the tangential space as quotient space, Q Q M .
The displacement vectors and distances based on which the dissimilarity measures are computed can be computed in the quotient space. Let x and y be two manifold points and the displacement between them d ( x , y ) is to be computed under the equivalence set Q . The equivalent images of x and y in all the tangential spaces at the different Q points are mapped to a reference tangential space (the quotient space) which can be the tangential space at any Q point. The mapping from the i-th tangential space to the r-th tangential space, E i r ( x ) , is provided by Equation (1) and similarly the mapping E j r ( x ) is provided by Equation (2). In the reference space there will be multiple images of x and y points and the displacement between them is defined as the displacement with the minimum magnitude (norm) between any equivalent image of x to any equivalent image of y , Equation (4), where · is the vector norm of the displacement vector and t = ( t 1 , t 2 , t 3 ) is the optimal triple of equivalent descriptors. A graphical illustration is provided in Figure 2.
E i r ( x ) = x + D r D i .
E j r ( y ) = y + D r D j .
t = arg min ( D i , D j , D r ) Q 3 E i r ( x ) E j r ( y ) .
d ( x , y ) = E t 1 t 3 ( x ) E t 2 t 3 ( y ) .
On the basis of the mapping relations, Equations (1) and (2), the differentiation between the expression variations and the identity variations can be achieved. The expression variation of a descriptor from one (the i-th) expression to another (the j-th) can be considered as the additive displacement vector Δ E i j = D j D i and then separated from the identity variation. Therefore, E i j ( · ) modifies (or displaces) the expression of the input descriptor from expression i to expression j.
In the next discussions, the term “identity” refers to the person identity for collections and ensembles and to the space point (location) for the unexpressed descriptors. Suppose that the identity of the descriptor x is p 1 and the identity of the descriptor y is p 2 . The descriptors x and y have both expression ( e x and e y ) and identity ( I ( x ) and I ( y ) ) components, Equations (5) and (6). The identity variation Δ I p 1 p 2 ( x , y ) is defined as the difference between the two identity components, Equation (7).
x = I ( x ) + e x .
y = I ( y ) + e y .
Δ I p 1 p 2 ( x , y ) = I ( y ) I ( x ) .
The identity variation Δ I p 1 p 2 ( x , y ) can be computed based on the displacement d ( x , y ) (Equation (4)) as shown in Equations (11) and (12).
d ( x , y ) = E i r ( x ) E j r ( y ) ,
= x y + D j D i ,
= I ( x ) + e x I ( y ) e y + D j D i , and
= Δ I p 1 p 2 ( x , y ) Δ E x i ( x ) + Δ E y j ( y ) .
When the expression of x is close (or ideally equals to) the expression D i and that of y equals to the expression D j then Δ I p 1 p 2 reduces to d ( x , y ) , Equation (12). This requirement is seamlessly achieved during the computation of d ( x , y ) according to Equation (4).
Δ I p 1 p 2 ( x , y ) = d ( x , y ) .
While the identity variations Δ I p 1 p 2 ( x , y ) is computed based on the descriptors x and y , Δ I p 1 p 2 ( . , . ) is independent of any particular descriptors as long they belong to the specified identities and at the corresponding key-points on the face. Therefore, the norm value of Δ I p 1 p 2 ( x , y ) or equally that of d ( x , y ) can be used in the computation of an invariant dissimilarity measure.
In general, the identity of a descriptor x can be changed from p 1 to the identity of any known descriptor, I ( . ) = p k , as in Equation (13).
I p 1 p 2 ( x ) = x + Δ I p 1 p k I ( I p 1 p k ( x ) ) = p k
Similarly, the expression of a descriptor x can be changed from e x to that of any known descriptor, E ( . ) = e k , as in Equation (14).
E e x e k ( x ) = x + Δ E e x e k E ( E e x e k ( x ) ) = e k
The identity and expression (changing) relations can be composed in any arbitrary sequence, e.g., that given in Equation (15).
x = E e 1 e 2 I p 2 p 3 E e x e 1 I p 1 p 2 ( x ) ,
= E e 1 e 2 E e x e 1 I p 2 p 3 I p 1 p 2 ( x ) , and
= E e x e 2 I p 1 p 3 ( x ) .
E ( x ) = e 2 .
I ( x ) = p 3 .
The identity and expression relation composition provides the proposed approach with the capacity to utilize the training data in invariant recognition. Let x and y be two novel descriptors respectively with identities p and g. To compute the invariant dissimilarity measure between them for example using just one equivalence relation, one need to search the training data for a pair of descriptors belonging to the same person c (an equivalence relation pair Q = { D 1 , D 2 } ) that minimizes Equation (22). Using the optimal Q, the identity variation Δ I p g is given equally by any of the displacements in Equation (24). Equations (20) and (21) basically find the image of x under the equivalence relation Q. Equation (21) is computable from the descriptors values but Equation (20) only illustrates its identity and expression composition. The dissimilarity measure Δ I p p is ideally zero (practically minimal) if both x and y have the same identity or non-zero (practically non-minimal) otherwise. Note that the descriptors form a graph path from x to y through D 1 and D 2 . More than one equivalence relation (belonging to different people) can be utilized in estimating Δ I p g , the details are provided in Section 3.4.
x = E D 2 D 1 I c p E D 1 D 2 E p D 1 I p c ( x )
= E D 1 D 2 ( x ) = x + D 2 D 1
Q = arg min Q | | x y | |
y = I p g E D 1 D 2 ( x )
Δ I p g = y E D 1 D 2 ( x ) = y D 2 + D 1 x

3.2. The Correspondence of the Descriptors

The correspondence of the descriptors is important for the proposed approach, since it enables the tracking of their identity and expression variations. For a pair of descriptor ensembles, N 1 and N 2 , the correspondence is defined as the mapping of each descriptor D i N 1 to one descriptor D j N 2 , such that each pair { D i , D j } correspond to a particular spatial location on the face. Let the correspondence of N 1 and N 2 be denoted by the set M 1 , 2 for which Equation (25) holds.
M 1 , 2 { { D i , D j } | D i N 1 D j N 2 } .
The correspondence is established based on a customized variant of the random sample consensus (RANSAC) algorithm [48]. The dissimilarity of the descriptors and their locations are both utilized. The descriptors in the two ensembles are initially matched against each other based on the dissimilarity measure in Equation (26). The summation is performed over all the descriptors. The descriptors in N 1 are corresponded to those in N 2 with the minimum dissimilarity measures.
d ( D i , D j ) = D i D j .
This will result in many correctly corresponded pairs but also will result in mis-correspondences. Let the key-point sets of the ensembles N 1 and N 2 be denoted respectively by K 1 and K 2 . Each key-point k in K 1 or K 2 is a vector of the 3D point coordinates, k = [ x , y , z ] . To correct the mis-correspondences, the 3D rigid transformations, both the rotation R and the translation t , that transform the key-points in K 1 to their corresponded ones in K 1 are first estimated using least square fitting. Then, the error of transformation (the Euclidean distance error), defined in Equation (28), is used to split the correspondences into inliers and outliers by comparison against a threshold value of the transformation error, e t h .
k i = R k i + t , k i K 1 .
e = k i k j , k j K 2 .
The rigid transformations are then recalculated base on the inlier correspondences and the process is iterated until the change in the norms of the new R and t diminishes. Using the converged values of R and t , the key-points in K 1 are transformed and then re-corresponded only to the K 2 key-points in their vicinity.

3.3. The Construction of the Embedded Graph

As previously mentioned, the graph is constructed from a training set of descriptors D organized in a set of ensembles N and a set collections C .
D = { D 1 , , D | D | } .
N = { N 1 , , N | N | } , where D N , N N D D .
C = { C 1 , , C | C | } , where N C , C C N N .
The ERs are extracted from the correspondences within the individual collections, between pairs of ensembles, while the IRs are extracted from those between pairs of collections.
Initially, sub-graphs representing the ERs are sequentially extracted from each collection C i in C . To achieve that, a list of ensemble pair combinations within the collection C i is generated and then each generated ensemble pair is corresponded as described. The use of all possible combinations may be overly redundant, especially for large number of ensembles within the collection. This is because the correspondence problem is transitive. The minimum number of such pairs required for connected sub-graphs is | C i | 1 , where | C i | is the number of ensembles in the collection. However, for robustness a sufficient level of redundancy is maintained, since the redundancy could mitigate possible errors in key-point localization and detection under different facial expressions.
A disciplined approach for the maintenance of the required redundancy is based on the reduction of the hope count between the ensembles, when representing the ensembles as graph nodes and the pair combinations as graph edges. First, a minimal connected graph, G i = ( V i , E i ) , is found, e.g., by connecting each ensemble to the next. Then, iteratively the graph diameter, the largest path between any two ensembles, is computed and an edge (an ensemble pair) is introduced between the two ensembles until the graph diameter falls below some predefined threshold. The correspondences are then computed for the ensemble pairs adjacent to every edge in E i of the graph G i and then combined in one correspondence set M C i for the collection C i , as in Equation (32).
M C i = e E i M j , k ( N j , N k ) , where { N j , N k } = e .
The pairs in the set M C i , when each is considered as a graph edge, form a new graph G C i with multiple connected sub-graphs G C i , s , where s = 1 , , S i .
G C i = ( V C i , E C i ) .
G C i , s = ( V C i , s , E C i , s ) , V C i , s V C i E C i , s E C i .
The number of the connected sub-graphs S i ideally should equal to the number of the descriptors per ensemble and the number of vertices of each subgraph | V C i , s | should equal to the number of the ensembles, | C i | , in the collection. However, this does not necessarily hold in practice. In fact, the number of the vertices can be less than | C i | for some sub-graphs, due to possible failure to detect some key-points, or it can be particularly much higher than | C i | , since more than one sub-graph can be joined together due to mis-correspondences. When the number of the vertices of any sub-graph is significantly low, it can be considered as an indication that the underling key-point is not repeatable and the sub-graph is discarded.
The larger connected sub-graphs are iteratively segmented (partitioned) based on the well-known spectral graph partitioning (clustering) algorithm [49,50]. The edges of the sub-graph to be partitioned are assigned the weight values shown in Equation (36). w e l is the weight of the edge e l after normalization (by division by the average non-normalized weight w ¯ ). The distance d between the adjacent descriptors to the edge is the same defined in Equation (26).
w e l = 1 d ( D j , D k ) , where { D j , D k } = e l , for l = 1 , , | E C i , s | .
w e l = w e l w ¯ , for l = 1 , , | E C i , s | .
The Laplacian matrix L C i , s of the weighted and connected sub-graph is then computed.
L C i , s = D C i , s A C i , s .
The diagonal matrix D C i , s is the degree matrix of the weighted graph, each diagonal element is the sum of the normalized weights of all the incident edges to the corresponding vertex of the graph. A C i , s is the adjacency matrix of the graph, each a i , j element is the normalized weight of the edge connecting the i-th vertex to the j-th vertex. The second smallest eigenvalue λ f of L C i , s is an indicator of how the graph is well-connected.
L C i , s x f = λ f x f .
The corresponding eigenvector to λ f is known as the Fiedler vector, x f . The Fielder vector elements has comparable values for strongly connected vertices. In the next step, the K-means clustering algorithm (with only two clusters) is applied to the elements of x f . Based on the resulting two clusters of vertices, the graph G C i , s partitioned. The described graph partitioning is iteratively applied until, the resulting graphs are strongly connected and roughly of the expected number of vertices, | C i | .
Finally, the vertices (the descriptors) of the resulting connected graphs are considered the equivalence sets Q C i , s , where s = 1 , , S i , of the collection C i . Each graph G C i , s is then simplified to the star graph T C i , s (the spanning graph). Every vertex descriptor in the equivalence set is connected to one neutral descriptor which is chosen as the nearest to the means of Q C i , s in case there are multiple neutral descriptors, called the bridging vertex (or descriptor), denoted by B i , s . Similarly, the equivalence sets and the equivalence star graphs are extracted for every collection. The equivalence graphs are then allowed to join corresponding ones in all other collections, from bridging to bridging vertices. These joining edges represent the IRs. It is possible not to define any particular bridging points and allow for IRs connections (edges) from all the equivalent vertices of one collection to all the vertices of the corresponding sets in all other collections. In this case, the spanning graph of the equivalence sets will be a line connecting them, rather than a star. However, the definition of the bridging points significantly simplifies the graph search problem (the matching), as will be discussed in the next subsection.

3.4. The Heuristic Graph Search

The graph search takes place during the matching of a probe ensemble N p to a gallery ensemble N g . The two ensembles are first corresponded. Theoretically, an invariant dissimilarity measure can be computed between N p and N g for an optimal graph path P * P connecting them. In the general case, possible candidate paths start with N p then zero or more collections are visited and finally terminate at N g . The paths should be simple, have no repeated edges or collections. At each visited collection two ensembles are visited (an entrance and an exit one). At the descriptor level, there are multiple paths (a path bundle) connecting the corresponded descriptors in parallel for each higher level path. The paths are eventually realized as sequences of descriptors. The entrance and exit ensembles may be fixed for the descriptor level path bundle. However, relaxing this requirement and letting the entrance and the exit ensembles to vary for the different descriptor level paths is beneficial when the variety of the training expressions in the collections are limited. Below are the definitions of the paths at the different levels.
P = { { N p , C I ( 1 ) , , C I ( | P | 2 ) , N g } | C I ( · ) C } . { { N p , N I ( 1 ) , E ( 1 ) , N I ( 1 ) , X ( 1 ) , , N I ( | P | 2 ) , E ( | P | 2 ) , N I ( | P | 2 ) , X ( | P | 2 ) , N g } | N I ( · ) , E ( . ) N N I ( · ) , X ( · ) N } . { { D p , i , D I ( 1 ) , E ( 1 ) , D I ( 1 ) , X ( 1 ) , , D I ( | P | 2 ) , E ( | P | 2 ) , D I ( | P | 2 ) , X ( | P | 2 ) , D g , j }
| D I ( · ) , E ( . ) D D I ( · ) , X ( . ) D { D p , i , D g , j } M p , g } = P ·
The multiple subscript indices of the path vertices uniquely point to the specific graph vertices and also indicate their identity I ( · ) , expression whether it is the entrance expression E ( · ) and the exit expression X ( · ) . Equation (41) shows how the invariant measure can be computed based on the graph paths.
s = min P P { x , y } M p , g E E ( | P | 2 ) X ( | P | 2 ) E E ( 2 ) X ( 2 ) E E ( 1 ) X ( 1 ) ( x ) y .
Existing optimal path searching algorithms such the Dijkstra and Bellman-Ford are not suitable for the solution of Equation (41). They deal scalar edge weights. In contrast, the edge weights in the tackled problem here are multidimensional vectors and the optimized quantity is the norm of their summation.
By considering only the bridging points, which were described earlier in Section 3.3, as the entrance and the exit vertices between collections, the graph density reduces and the maximum number of collections per path also reduces to two. This means that each considered path has at maximum three identity changes and two removable expressions. Apart from the constraint on the maximum path length, | P | , Equation (41) holds for this lighter version of the graph search problem.
The proposed heuristic graph search proceeds by assigning one collection to the probe ensemble and one collection to the gallery ensemble (possibly another one). These assignments are initially performed based on the vicinity (in the descriptor space) of the descriptors of the probe and the gallery ensembles to those in the assigned collections. A KD-tree of all the training descriptors was built during an off-line stage to enable an efficient search for the nearest neighbors. A table of the descriptors information containing their ensemble, collection and equivalence sets is associated with the KD-tree. The k nearest neighbors of each descriptor in the probe or the gallery ensemble vote for the different collections based on their associated information (labels). The collection that receives the highest number of votes is assigned to the ensemble. Next, the descriptors of N p and N g are assigned entrance and exit descriptors within the assigned collections. For this task, a separate KD-tree per collection was built (during an off-line stage), since smaller KD-tree are more efficient to search. Then, for each corresponded descriptor pair, { x , y } M p , g where x N p and y N g , the nearest three descriptors to x are considered as potential entrance descriptors to the collection assigned to N p . Similarly, the three nearest neighbors to y are found and considered as potential exit descriptors from the collection assigned to N g . Among the few combinations of potential entrance and exit descriptors, the one which yields the lowest value of the scalar function m ( x , y ) , defined in Equation (42), is assigned to x and y as respectively their entrance and exit descriptors.
m ( x , y ) = E E ( 2 ) X ( 2 ) E E ( 1 ) X ( 1 ) ( x ) y
= Δ I p g
At this point of time, only a good initial guess of the solution is found and the search for the optimal path (or measure) is not performed yet. Nonetheless, the most similar people (collections) and expressions are likely to be assigned to the probe and the gallery.
The optimization is carried out implicitly as nearest neighbor search. The descriptors of N p are first displaced as in Equation (44) which produces new images of descriptors, x i for i = 1 , , | N p | . The new descriptors are then used to re-assign the gallery a new training collection and new entrance and exit descriptors as described earlier. It is then followed by a similar displacement and re-assignment of the probe descriptors based on the images of the gallery descriptors, y i as in Equation (45). Each of the two steps implies the minimization of Equation (42).
x i = E E ( 2 ) X ( 2 ) E E ( 1 ) X ( 1 ) ( x ) = x + D I ( 1 ) , X ( 1 ) D I ( 1 ) , E ( 1 ) + D I ( 1 ) , X ( 2 ) D I ( 1 ) , E ( 2 )
y i = E X ( 2 ) E ( 2 ) E X ( 1 ) E ( 1 ) ( y ) = y + D I ( 1 ) , E ( 1 ) D I ( 1 ) , X ( 1 ) + D I ( 1 ) , E ( 2 ) D I ( 1 ) , X ( 2 )
This process is then iterated a few times until convergence. These steps are only committed when they result in further minimization of Equation (42).
The described graph search accounts for paths with one and two collections (as both the probe and the gallery can be assigned to the same collection). The direct path between the probe and the gallery should also be considered which is accounted for by the simple minimization in Equation (46).
m ( x , y ) = min { m ( x , y ) , x y } .

3.5. The Dissimilarity Measure between Ensembles

An overall dissimilarity measure, s, between any two ensembles can be computed from the dissimilarity measures between the corresponded descriptor pairs, i.e., the m values shown in Equation (46). The N descriptor measures with the least values are simply summed to produce an overall ensemble measure s, as in Equation (47). N is much less than the typical number of the corresponded descriptor pairs. This would avoid the measures with high values. For those ones the expression variations may not be effectively removed by the proposed approach. When computing the dissimilarity matrix, its entries are further normalized to range from zero to one for each probe.
s ( N p , N g ) = i N m i .

4. Experiments

A number of face recognition experiments were conducted on the FRGC v2.0 [51], the 3D TEC [52] and the Bosphorus [53] datasets. These datasets differ from each other in a number of aspects. First, the FRGC dataset has the largest number of individuals (466 people in the testing partition) among the three datasets. It has diverse facial expressions but about half of its facial scans are under neutral or near neutral expressions. On the other hand, the 3D TEC and the Bosphorus datasets can even pose a more significant challenge to the recognition under facial expression variations. In the case of the 3D TEC, the challenge mainly arises because the individuals are identical twins (107 twins/214 individuals). In the third and the fourth 3D TEC experiments, the probes and the galleries are under different facial expressions. In contrast, the probe and the gallery scans specified in the dataset for the first and the second 3D TEC experiments involve no expression variations. In the case of the Bosphorus dataset, there are many scans for only 105 different people. However, the facial expression challenge arises because the facial expressions are generally of a larger extent in comparison to the other two datasets.

4.1. RAIK Descriptor Based Experiments

As the proposed expression invariant approach requires a set of training data with many individuals and under different facial expressions including the neutral expression, the FRGC dataset is an appropriate choice for training the system. The FRGC dataset has a training partition. However, it has a limited number individuals and the individual sets of the training and the testing partitions are not mutually exclusive. For this reason, a significant part of the testing partition of the FRGC dataset was used for training the proposed system, all the facial scans of the first 300 individuals. The remaining scans of the testing partition was used for testing the proposed approach. A gallery of 166 neutral facial scans (one scan per subject) was formed. The remaining scans were split into a neutral and a non-neutral probe subsets. The trained system was then used to perform the tests on the 3D TEC and the Bosphorus datasets. In all the experiments including both the expression-invariant approach and the plain RAIK approach which was used for result comparison, the RAIK features were compressed using the principal component analysis (PCA), each to a vector of twenty PCA coefficients. The RAIK descriptor has two adjustable parameters α and β . They were respectively set to −0.15 and 0.0 for all the conducted experiments.
The recognition performance results of the experiments conducted on the FRGC dataset, Figure 3 and Figure 4, indicate that the proposed expression-invariant approach noticeably enhances the identification rates of the non-neutral probes at the first few ranks where the expression variations have more impact on the identification performance. While the first rank identification rate has increased from 97.69 % (for the plain RAIK approach) to 97.90 % (for the proposed approach based on the RAIK descriptors), the margin between the two rates has further increased at the second rank and peaked at the third rank where the identification rate has increased from 98.32 % to 98.95 % . It should be noted that the plain RAIK approach already achieves a very high identification performance because it limits the matching to the semi-rigid regions (the forehead and the nose) of the face. As more regions are considered, the identification rate margin between the proposed approach and the plain RAIK approach increases in favor of the proposed approach. This is because the identification rates of the plain RAIK approach declines more rapidly with the inclusion of the non-rigid regions of the face while proposed approach still declines but a slower pace. However, the performance of both approaches is optimal when only the semi-rigid regions of the face are considered. It could be concluded that the proposed approach contributes the reliability of the recognition in addition to the observed recognition performance enhancement. For the neutral experiment, the impact of the proposed approach is limited, which is expected as there are no expression variations and the identification rates for the neutral scans are already above 99.5 % . Some verification rate improvement was observed for the non-neutral experiment but it was not significant. It has increased from 98.11 to 98.31 % at 0.001 FAR.
The identification and verification rates of the first and the second experiments of the 3D TEC dataset were not significantly impacted by the proposed approach. The interpretation of this observation is that for these two experiments the probe and the gallery scans of the twins are under the same expressions. In contrast, the impact of the proposed approach on the third and the fourth experiments was more significant (Figure 5 and Figure 6). The proposed approach has increased the first rank identification rate of the third experiment from 85.51 to 89.25 % and from 86.45 to 89.25 % for the fourth experiment. It appears from the 3D TEC and the FRGC results that the recognition enhancement of the proposed approach becomes more significant when the expression variations are more challenging to the plain RAIK approach. For these two experiments, the verification rates were respectively 92.52 % and 91.12 % at 0.001 FAR for the proposed approach, in comparison to 88.79 % and 88.32 % at 0.001 FAR for the plain RAIK approach. The results of the Bosphorus dataset indicate that the proposed approach enhances the recognition performance for the probes under non-neutral facial expressions (Figure 7 and Figure 8). The identification rate had increased from 91.94 to 93.55 % and the verification rate had increased from 92.60 to 94.21 % at 0.001 FAR. There was a negligible degradation in the verification performance of the neutral expression scans. Nonetheless, both the proposed system and the plain RAIK system had achieved above 99.5 % verification at 0.001 FAR for the neutral scans.

4.2. Local 3D PCA Descriptor Based Experiments

The experiments in this subsection were conducted in a similar fashion to those performed using RAIK descriptor on the FRGC and the 3D TEC datasets. The RAIK descriptors were used instead of the local 3D PCA ones. Local 3D PCA descriptor is known to perform very well in 3D face recognition [4,54]. However, it is not rotation-invariant in contrast to RAIK. Therefore, the 3D faces were ensured to have a standard frontal pose using the principal components of the 3D face. The nose tip of the face was detected in the 3D face point-cloud and then the facial 3D points within 80 mm radius were cropped. They are then pose-corrected to their principal vertical and horizontal components and converted to range images. The local 3D PCA descriptors were extracted from 3D surface patches of 15 × 15 mm size. They are then translated along the z-coordinate so their central points have a depth value of zero. Finally, the dimensionality of the 3D local surfaces are reduced using standard PCA to 20D vectors.
3D face recognition using the proposed approach based on local PCA descriptors was performed and compared against the results of local PCA descriptors alone. The use of the proposed approach has shown to boost both the identification and verification rates of standard PCA when performed in similar settings. The improvement margins generally seem better than those observed for RAIK. However, the performance of the proposed approach based on RAIK is much better than the PCA one. This is probably because RAIK is rotation-invariant and more discriminative.
The proposed approach based on local PCA has shown to improve rank-1 identification rate from 92.76 to 94.6 % for the FRGC 3D faces under facial expressions, see Figure 9. The verification rate has also improved from 95.34 to 96.81 % at 0.001 FAR. The performance on 3D TEC dataset has also experienced considerable improvements, see Figure 10 and Figure 11. The first rank identification rates when using the proposed approach were 90.65 % , 90.65 % , 73.36 % and 73.83 % , respectively for the four standard 3D TEC experiments (experiment-I, II, III and IV). Without the application of the proposed approach, the rates were respectively 85.98 % , 86.92 % , 69.16 % and 68.22 % . The verification rates at 0.001 respectively before and after the application of the proposed approach were 87.85 % and 92.06 % for the first experiment but 88.79 % and 93.46 % for the second one. The rates have also improved for the third and fourth experiments respectively from 70.09 to 74.77 % and from 69.16 to 75.23 % at roughly 0.0011 FAR (the lowest possible one).
In comparison to the recent 3D face recognition literature, the proposed approach achieves comparable results to the state-of-art even if it does not generally outperform recent deep learning based advances, especially those based on holistic face recognition [55,56]. However, the RAIK descriptor in particular already achieves a high recognition performance on its own and the proposed approach has shown to further improve its performance. Some of the results of the proposed approach discussed earlier are in close proximity or even exceed those of deep learning. For example, identification rate ( 89.25 % ) of the proposed approach using RAIK for the third and fourth 3D TEC experiments exceed those of the highly performing “deep 3D face identification” method described in [55] (whose rates are respectively 81.3 % and 79.9 % ), see Table 1. Apart from recognition performance, conducting 3D face recognition based on local features may have advantages over holistic face recognition since it does not require the visibility of the whole face for matching.

5. Conclusions

The proposed research described a new manifold-based approach for learning facial expression invariance of the key-point descriptors. The descriptor variations induced by the facial expressions were handled with the equivalence relations within the descriptor manifold. Then, invariant dissimilarity measures (distances) were computed based on the equivalence relations. This approach has shown to improve the recognition performance especially when the facial scans being matched involve expression variations.

Funding

This research received no external funding.

Conflicts of Interest

The author declares no conflict of interest.

References

  1. Lowe, D.G. Object recognition from local scale-invariant features. In Proceedings of the Seventh IEEE International Conference on Computer Vision, Kerkyra, Greece, 20–27 September 1999; IEEE: New York, NY, USA, 1999; Volume 2, pp. 1150–1157. [Google Scholar]
  2. Belongie, S.; Malik, J.; Puzicha, J. Shape matching and object recognition using shape contexts. IEEE Trans. Pattern Anal. Mach. Intell. 2002, 24, 509–522. [Google Scholar] [CrossRef] [Green Version]
  3. Rublee, E.; Rabaud, V.; Konolige, K.; Bradski, G. ORB: An efficient alternative to SIFT or SURF. In Proceedings of the 2011 IEEE International Conference on Computer Vision (ICCV), Barcelona, Spain, 6–13 November 2011; pp. 2564–2571. [Google Scholar]
  4. Mian, A.S.; Bennamoun, M.; Owens, R. Keypoint detection and local feature matching for textured 3D face recognition. Int. J. Comput. Vis. 2008, 79, 1–12. [Google Scholar] [CrossRef]
  5. Al-Osaimi, F.; Bennamoun, M.; Mian, A. Interest-point based face recognition from range images. In Proceedings of the British Machine Vision Conference, Warwick, UK, 10–13 September 2007. [Google Scholar]
  6. Al-Osaimi, F.R. A novel multi-purpose matching representation of local 3D surfaces: A rotationally invariant, efficient, and highly discriminative approach with an adjustable sensitivity. IEEE Trans. Image Process. 2015, 25, 658–672. [Google Scholar] [CrossRef] [PubMed]
  7. Mian, A.S.; Bennamoun, M.; Owens, R. An efficient multimodal 2D-3D hybrid approach to automatic face recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2007, 29, 1927–1943. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  8. Chang, K.J.; Bowyer, K.W.; Flynn, P.J. Effects on facial expression in 3D face recognition. In Proceedings of the Defense and Security, International Society for Optics and Photonics, Orlando, FL, USA, 29–30 March 2005; pp. 132–143. [Google Scholar]
  9. Chang, K.I.; Bowyer, W.; Flynn, P.J. Multiple nose region matching for 3D face recognition under varying facial expression. IEEE Trans. Pattern Anal. Mach. Intell. 2006, 28, 1695–1700. [Google Scholar] [CrossRef] [PubMed]
  10. Faltemier, T.C.; Bowyer, K.W.; Flynn, P.J. A region ensemble for 3-D face recognition. IEEE Trans. Inf. Forensics Secur. 2008, 3, 62–73. [Google Scholar] [CrossRef]
  11. Bronstein, A.M.; Bronstein, M.M.; Kimmel, R. Expression-invariant 3D face recognition. In Proceedings of the Audio-and Video-Based Biometric Person Authentication, Guildford, UK, 9–11 June 2003; Springer: Berlin/Heidelberg, Germany, 2003; pp. 62–70. [Google Scholar]
  12. Drira, H.; Ben Amor, B.; Srivastava, A.; Daoudi, M.; Slama, R. 3D face recognition under expressions, occlusions, and pose variations. IEEE Trans. Pattern Anal. Mach. Intell. 2013, 35, 2270–2283. [Google Scholar] [CrossRef] [Green Version]
  13. Smeets, D.; Hermans, J.; Vandermeulen, D.; Suetens, P. Isometric deformation invariant 3D shape recognition. Pattern Recognit. 2012, 45, 2817–2831. [Google Scholar] [CrossRef]
  14. Berretti, S.; Del Bimbo, A.; Pala, P. 3D face recognition using isogeodesic stripes. IEEE Trans. Pattern Anal. Mach. Intell. 2010, 32, 2162–2177. [Google Scholar] [CrossRef] [Green Version]
  15. Metaxas, D.N.; Kakadiaris, I.A. Elastically adaptive deformable models. IEEE Trans. Pattern Anal. Mach. Intell. 2002, 24, 1310–1321. [Google Scholar] [CrossRef]
  16. Murtuza, M.N.; Lu, Y.; Karampatziakis, N.; Theoharis, T. Three-Dimensional Face Recognition in the Presence of Facial Expressions: An Annotated Deformable Model Approach. IEEE Trans. Pattern Anal. Mach. Intell. 2007, 29, 640–649. [Google Scholar]
  17. Al-Osaimi, F.; Bennamoun, M.; Mian, A. An expression deformation approach to non-rigid 3D face recognition. Int. J. Comput. Vis. 2009, 81, 302–316. [Google Scholar] [CrossRef] [Green Version]
  18. Maalej, A.; Amor, B.B.; Daoudi, M.; Srivastava, A.; Berretti, S. Shape analysis of local facial patches for 3D facial expression recognition. Pattern Recognit. 2011, 44, 1581–1589. [Google Scholar] [CrossRef] [Green Version]
  19. Gökberk, B.; İrfanoğlu, M.O.; Akarun, L. 3D shape-based face representation and feature extraction for face recognition. Image Vis. Comput. 2006, 24, 857–869. [Google Scholar] [CrossRef]
  20. Lu, X.; Jain, A.K. Multimodal Facial Feature Extraction for Automatic 3D Face Recognition; Tech. Rep. MSU-CSE-05-22; Department of Computer Science, Michigan State University: East Lansing, MI, USA, 2005. [Google Scholar]
  21. Xu, C.; Li, S.; Tan, T.; Quan, L. Automatic 3D face recognition from depth and intensity Gabor features. Pattern Recognit. 2009, 42, 1895–1905. [Google Scholar] [CrossRef]
  22. Tan, X.; Triggs, B. Fusing Gabor and LBP feature sets for kernel-based face recognition. In Analysis and Modeling of Faces and Gestures; Springer: Berlin/Heidelberg, Germany, 2007; pp. 235–249. [Google Scholar]
  23. Wang, Y.; Liu, J.; Tang, X. Robust 3D face recognition by local shape difference boosting. IEEE Trans. Pattern Anal. Mach. Intell. 2010, 32, 1858–1870. [Google Scholar] [CrossRef] [Green Version]
  24. Singh, C.; Mittal, N.; Walia, E. Face recognition using Zernike and complex Zernike moment features. Pattern Recognit. Image Anal. 2011, 21, 71–81. [Google Scholar] [CrossRef]
  25. Turk, M.; Pentland, A. Eigenfaces for recognition. J. Cogn. Neurosci. 1991, 3, 71–86. [Google Scholar] [CrossRef]
  26. Belhumeur, P.N.; Hespanha, J.P.; Kriegman, D. Eigenfaces vs. fisherfaces: Recognition using class specific linear projection. IEEE Trans. Pattern Anal. Mach. Intell. 1997, 19, 711–720. [Google Scholar] [CrossRef] [Green Version]
  27. Blanz, V.; Vetter, T. Face recognition based on fitting a 3D morphable model. IEEE Trans. Pattern Anal. Mach. Intell. 2003, 25, 1063–1074. [Google Scholar] [CrossRef] [Green Version]
  28. Wang, X.; Tang, X. A unified framework for subspace face recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2004, 26, 1222–1228. [Google Scholar] [CrossRef] [PubMed]
  29. Russ, T.; Boehnen, C.; Peters, T. 3d face recognition using 3d alignment for pca. In Proceedings of the 2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, New York, NY, USA, 17–22 June 2006; Volume 2, pp. 1391–1398. [Google Scholar]
  30. Draper, B.A.; Baek, K.; Bartlett, M.S.; Beveridge, J.R. Recognizing faces with PCA and ICA. Comput. Vis. Image Underst. 2003, 91, 115–137. [Google Scholar] [CrossRef]
  31. Harguess, J.; Aggarwal, J. A case for the average-half-face in 2D and 3D for face recognition. In Proceedings of the 2009 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops, Miami, FL, USA, 20–25 June 2009; pp. 7–12. [Google Scholar]
  32. Liu, C.; Wechsler, H. Independent component analysis of Gabor features for face recognition. IEEE Trans. Neural Netw. 2003, 14, 919–928. [Google Scholar] [PubMed] [Green Version]
  33. Verleysen, M.; François, D. The curse of dimensionality in data mining and time series prediction. In Computational Intelligence and Bioinspired Systems; Springer: Berlin/Heidelberg, Germany, 2005; pp. 758–770. [Google Scholar]
  34. Jain, A.; Zongker, D. Feature selection: Evaluation, application, and small sample performance. IEEE Trans. Pattern Anal. Mach. Intell. 1997, 19, 153–158. [Google Scholar] [CrossRef] [Green Version]
  35. Paalanen, P.; Kamarainen, J.K.; Ilonen, J.; Kälviäinen, H. Feature representation and discrimination based on Gaussian mixture model probability densities—Practices and algorithms. Pattern Recognit. 2006, 39, 1346–1358. [Google Scholar] [CrossRef]
  36. Raudys, S.J.; Jain, A.K. Small sample size effects in statistical pattern recognition: Recommendations for practitioners. IEEE Trans. Pattern Anal. Mach. Intell. 1991, 13, 252–264. [Google Scholar] [CrossRef]
  37. Schölkopf, B.; Smola, A.; Müller, K.R. Nonlinear component analysis as a kernel eigenvalue problem. Neural Comput. 1998, 10, 1299–1319. [Google Scholar] [CrossRef] [Green Version]
  38. Hotta, K. Robust face recognition under partial occlusion based on support vector machine with local Gaussian summation kernel. Image Vis. Comput. 2008, 26, 1490–1498. [Google Scholar] [CrossRef]
  39. Ksantini, R.; Boufama, B.S.; Ahmad, I.S. A new KSVM+ KFD model for improved classification and face recognition. J. Multimed. 2011, 6, 39–47. [Google Scholar]
  40. Yang, M.H. Kernel eigenfaces vs. kernel fisherfaces: Face recognition using kernel methods. In Proceedings of the 2013 10th IEEE International Conference and Workshops on Automatic Face and Gesture Recognition (FG), Shanghai, China, 22–26 April 2013; p. 0215. [Google Scholar]
  41. Balasubramanian, V.N.; Ye, J.; Panchanathan, S. Biased manifold embedding: A framework for person-independent head pose estimation. In Proceedings of the 2007 CVPR’07 IEEE Conference on Computer Vision and Pattern Recognition, Minneapolis, MN, USA, 17–22 June 2007; pp. 1–7. [Google Scholar]
  42. Chang, Y.; Hu, C.; Feris, R.; Turk, M. Manifold based analysis of facial expression. Image Vis. Comput. 2006, 24, 605–614. [Google Scholar] [CrossRef]
  43. Zhang, T.; Yang, J.; Zhao, D.; Ge, X. Linear local tangent space alignment and application to face recognition. Neurocomputing 2007, 70, 1547–1553. [Google Scholar] [CrossRef]
  44. Lin, T.; Zha, H.; Lee, S.U. Riemannian manifold learning for nonlinear dimensionality reduction. In Computer Vision—ECCV 2006; Springer: Berlin/Heidelberg, Germany, 2006; pp. 44–55. [Google Scholar]
  45. Pan, Y.; Ge, S.S.; Al Mamun, A. Weighted locally linear embedding for dimension reduction. Pattern Recognit. 2009, 42, 798–811. [Google Scholar] [CrossRef]
  46. Samir, C.; Srivastava, A.; Daoudi, M. Three-dimensional face recognition using shapes of facial curves. IEEE Trans. Pattern Anal. Mach. Intell. 2006, 28, 1858–1863. [Google Scholar] [CrossRef] [PubMed]
  47. Raytchev, B.; Yoda, I.; Sakaue, K. Head pose estimation by nonlinear manifold learning. In Proceedings of the ICPR 2004 17th International Conference on Pattern Recognition, Cambridge, UK, 26 August 2004; Volume 4, pp. 462–466. [Google Scholar]
  48. Fischler, M.A.; Bolles, R.C. Random sample consensus: A paradigm for model fitting with applications to image analysis and automated cartography. Commun. ACM 1981, 24, 381–395. [Google Scholar] [CrossRef]
  49. Fiedler, M. Algebraic connectivity of graphs. Czechoslov. Math. J. 1973, 23, 298–305. [Google Scholar] [CrossRef]
  50. Donath, W.E.; Hoffman, A.J. Lower bounds for the partitioning of graphs. IBM J. Res. Dev. 1973, 17, 420–425. [Google Scholar] [CrossRef]
  51. Phillips, P.J.; Flynn, P.J.; Scruggs, T.; Bowyer, K.W.; Chang, J.; Hoffman, K.; Marques, J.; Min, J.; Worek, W. Overview of the face recognition grand challenge. In Proceedings of the CVPR 2005. IEEE Computer Society Conference on Computer Vision and pattern Recognition, San Diego, CA, USA, 20–25 June 2005; Volume 1, pp. 947–954. [Google Scholar]
  52. Vijayan, V.; Bowyer, K.; Flynn, P. 3D twins and expression challenge. In Proceedings of the 2011 IEEE International Conference on Computer Vision Workshops (ICCV Workshops), Barcelona, Spain, 6–13 November 2011; pp. 2100–2105. [Google Scholar]
  53. Savran, A.; Alyüz, N.; Dibeklioğlu, H.; Çeliktutan, O.; Gökberk, B.; Sankur, B.; Akarun, L. Bosphorus database for 3D face analysis. In Proceedings of the European Workshop on Biometrics and Identity Management; Springer: Berlin/Heidelberg, Germany, 2008; pp. 47–56. [Google Scholar]
  54. Ouamane, A.; Chouchane, A.; Boutellaa, E.; Belahcene, M.; Bourennane, S.; Hadid, A. Efficient tensor-based 2d+ 3d face verification. IEEE Trans. Inf. Forensics Secur. 2017, 12, 2751–2762. [Google Scholar] [CrossRef]
  55. Kim, D.; Hernandez, M.; Choi, J.; Medioni, G. Deep 3D face identification. In Proceedings of the 2017 IEEE International Joint Conference on Biometrics (IJCB) IEEE, Denver, CO, USA, 1–4 October 2017; pp. 133–142. [Google Scholar]
  56. Tuan Tran, A.; Hassner, T.; Masi, I.; Medioni, G. Regressing robust and discriminative 3D morphable models with a very deep neural network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 5163–5172. [Google Scholar]
  57. Li, H.; Huang, D.; Morvan, J.M.; Wang, Y.; Chen, L. Towards 3D face recognition in the real: A registration-free approach using fine-grained matching of 3D keypoint descriptors. Int. J. Comput. Vis. 2015, 113, 128–142. [Google Scholar] [CrossRef]
Figure 1. An Overview illustration of the proposed approach (better seen in color).
Figure 1. An Overview illustration of the proposed approach (better seen in color).
Jimaging 06 00112 g001
Figure 2. In (a), two manifold points appearing in the tangential spaces at three equivalent manifold points (the square ones). It shows the displacement vector between the two points. In (b), the original two points and their images under the equivalency are shown with respect to the reference tangential space (the solid square). The invariant displacement vector under the equivalency is also shown.
Figure 2. In (a), two manifold points appearing in the tangential spaces at three equivalent manifold points (the square ones). It shows the displacement vector between the two points. In (b), the original two points and their images under the equivalency are shown with respect to the reference tangential space (the solid square). The invariant displacement vector under the equivalency is also shown.
Jimaging 06 00112 g002
Figure 3. The CMC curves of the non-neutral (left) and the neutral (right) subsets of the FRGC dataset which were spared for evaluation. The curves compare the performance when using the proposed learning approach based on the RAIK features to that when the RAIK features are used without learning. Both methods achieve a very high performance. However, a noticeable improvement is observed for the non-neutral scans in particular, especially for the first few ranks.
Figure 3. The CMC curves of the non-neutral (left) and the neutral (right) subsets of the FRGC dataset which were spared for evaluation. The curves compare the performance when using the proposed learning approach based on the RAIK features to that when the RAIK features are used without learning. Both methods achieve a very high performance. However, a noticeable improvement is observed for the non-neutral scans in particular, especially for the first few ranks.
Jimaging 06 00112 g003
Figure 4. The ROC curves for the non-neutral subset of the FRGC dataset which were spared for evaluation (RAIK descriptor).
Figure 4. The ROC curves for the non-neutral subset of the FRGC dataset which were spared for evaluation (RAIK descriptor).
Jimaging 06 00112 g004
Figure 5. The CMC curves of the four experiments of the 3D TEC dataset performed using RAIK descriptor. The first two experiments involve no expression variations. For these two experiments there is a limited effect of the learning approach on the identification performance. In contrast, the third and fourth experiments have expression variations among the twins facial scans, for which the proposed approach has shown to considerably improve the recognition performance.
Figure 5. The CMC curves of the four experiments of the 3D TEC dataset performed using RAIK descriptor. The first two experiments involve no expression variations. For these two experiments there is a limited effect of the learning approach on the identification performance. In contrast, the third and fourth experiments have expression variations among the twins facial scans, for which the proposed approach has shown to considerably improve the recognition performance.
Jimaging 06 00112 g005
Figure 6. The ROC curves of the four 3D TEC experiments based on RAIK.
Figure 6. The ROC curves of the four 3D TEC experiments based on RAIK.
Jimaging 06 00112 g006
Figure 7. The CMC curves of the Bosphorus dataset (RAIK descriptor).
Figure 7. The CMC curves of the Bosphorus dataset (RAIK descriptor).
Jimaging 06 00112 g007
Figure 8. The ROC curves of the Bosphorus dataset (RAIK descriptor).
Figure 8. The ROC curves of the Bosphorus dataset (RAIK descriptor).
Jimaging 06 00112 g008
Figure 9. The CMC and ROC curves of the FRGC facial scans under expression using local 3D PCA descriptor.
Figure 9. The CMC and ROC curves of the FRGC facial scans under expression using local 3D PCA descriptor.
Jimaging 06 00112 g009
Figure 10. The CMC curves of the four 3D TEC experiments based on local 3D PCA descriptor.
Figure 10. The CMC curves of the four 3D TEC experiments based on local 3D PCA descriptor.
Jimaging 06 00112 g010
Figure 11. The ROC curves of the four 3D TEC experiments based on local 3D PCA.
Figure 11. The ROC curves of the four 3D TEC experiments based on local 3D PCA.
Jimaging 06 00112 g011
Table 1. Rank-1 identification rate comparisons on the three datasets.
Table 1. Rank-1 identification rate comparisons on the three datasets.
ApproachFRGC V2.0Bosphorus3D TEC
Expression Neutral(Expression)Exp. IExp. IIExp. IIIExp. IV
Li et al. [57]96.3% (combined)98.82%NANANANA
Kim et al. [55]NANA99.2%94.8%94.8%81.3%79.9%
This paper97.90%99.55%93.55%94.39%92.99%89.25%89.25%
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Al-Osaimi, F.R. Learning Descriptors Invariance through Equivalence Relations within Manifold: A New Approach to Expression Invariant 3D Face Recognition. J. Imaging 2020, 6, 112. https://doi.org/10.3390/jimaging6110112

AMA Style

Al-Osaimi FR. Learning Descriptors Invariance through Equivalence Relations within Manifold: A New Approach to Expression Invariant 3D Face Recognition. Journal of Imaging. 2020; 6(11):112. https://doi.org/10.3390/jimaging6110112

Chicago/Turabian Style

Al-Osaimi, Faisal R. 2020. "Learning Descriptors Invariance through Equivalence Relations within Manifold: A New Approach to Expression Invariant 3D Face Recognition" Journal of Imaging 6, no. 11: 112. https://doi.org/10.3390/jimaging6110112

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop