Next Article in Journal
Accelerometer-Based Human Fall Detection Using Convolutional Neural Networks
Previous Article in Journal
Ultrasonic Detection Method for Grouted Defects in Grouted Splice Sleeve Connector Based on Wavelet Pack Energy
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Robust Face Recognition Based on a New Supervised Kernel Subspace Learning Method

by
Ali Khalili Mobarakeh
*,
Juan Antonio Cabrera Carrillo
and
Juan Jesús Castillo Aguilar
Department of Mechanical Engineering, University of Málaga, Doctor Ortiz Ramos s/n, 29071 Malaga, Spain
*
Author to whom correspondence should be addressed.
Sensors 2019, 19(7), 1643; https://doi.org/10.3390/s19071643
Submission received: 22 February 2019 / Revised: 1 April 2019 / Accepted: 2 April 2019 / Published: 6 April 2019
(This article belongs to the Section Optical Sensors)

Abstract

:
Face recognition is one of the most popular techniques to achieve the goal of figuring out the identity of a person. This study has been conducted to develop a new non-linear subspace learning method named “supervised kernel locality-based discriminant neighborhood embedding,” which performs data classification by learning an optimum embedded subspace from a principal high dimensional space. In this approach, not only nonlinear and complex variation of face images is effectively represented using nonlinear kernel mapping, but local structure information of data from the same class and discriminant information from distinct classes are also simultaneously preserved to further improve final classification performance. Moreover, in order to evaluate the robustness of the proposed method, it was compared with several well-known pattern recognition methods through comprehensive experiments with six publicly accessible datasets. Experiment results reveal that our method consistently outperforms its competitors, which demonstrates strong potential to be implemented in many real-world systems.

1. Introduction

In the field of face recognition, many different dimensional recognition approaches have been developed in recent times [1,2,3]. Dimensionality reduction is a main problem in numerous recognition techniques, caused by the great amount of data with high dimensions in many real-world utilizations [4,5,6]. In fact, dimensionality reduction techniques have been recommended by researchers to avoid “the cure of dimensionality,” so as to amend the computational efficiency of image recognition [7,8]. Generally, dimensionality reduction techniques can be classified into two main groups: i.e., linear and nonlinear. In linear methods, a significant low-dimensional subspace is intended to be discovered in the input data with high-dimensional space, where the embedded data in the input space has a linear structure [9,10,11,12]. Principle component analysis (PCA) is one of the most famous linear methods [13,14,15]. PCA aims to retain global geometric information for data representation through enhancing the trace of the feature covariance matrix [13,16,17]. Linear discriminant analysis (LDA) is a linear technique that seeks to find out the discriminant information for data classification by enhancing the ratio between inter-class and intra-class scatters [16,18]. Some of the limitations of both PCA and LDA are that they could suffer from the small sample size (SSS) [15] issue, and that they may fail to recognize many important data structures that are nonlinear [19,20].
Scholars have developed abundant practical nonlinear dimensionality reduction strategies [21] to address these problems. They can be classified into two types: manifold-learning-based and kernel-based techniques [22,23]. Manifold learning directly aims to discover principal nonlinear data with low-dimensional structures that are concealed in the input space. Isometric feature mapping (ISOMAP) [24,25] and local linear embedding (LLE) [26,27] are the most well-known manifold-learning-based techniques to find inherent low-dimensional embedding of data [28]. Based on some experiments carried out using these techniques, it is illustrated that these methods can discover meaningful embedded nonlinear data structures for face images adequately.
However, manifold-learning-based techniques could suffer from two issues in terms of pattern recognition [15]. The first one is called “overlearning of locality,” [29] since manifold learning keeps locality data structures but there is no straight connection with classification. Out-of-sample is another issue that shows why most manifold-learning-based techniques are not appropriate for image recognition tasks [30,31]. These techniques can yield an embedding directly from a training data set, but they are often not able to find the sample’s image in the embedding space when it is implemented to a new point. These problems cannot be overcome by the currently proposed manifold learning methods. Although a few supervised forms have been proposed, they still suffer from these problems [30,31,32] because they are all founded based on “locality” characterization. The local quantity is sufficient for one manifold modeling, but it does not work well for classification tasks in multi-manifold modeling [17].
In contrast to manifold-learning-based techniques and in order to indirectly represent observed patterns into possibly much larger dimensional feature vectors, kernel-based techniques have been proposed by applying a kernelized nonlinear representation method. In this approach, the nonlinear data structure can be more separable in the observation space and become linear in the feature space. The representative strategies include kernel Fisher discriminant (KFD) [23,33] and kernel principal component analysis (KPCA) [34,35]. Both have shown that they can be practical in many real-world functions, such as face recognition, to preserve the nonlinear data structure [36]. However, these kernel-based methods cannot directly consider the local data structure, which results in classification performance degradation.
Recently, the locality preserving projections (LPP) method has been proposed as a linear subspace learning method to address the out-of-sample problem [37]. LPP is an unsupervised linear subspace technique that has the remarkable advantage of being able to generate an explicit map. Similar to the one belonging to PCA and LDA, this map is linear and easy to compute, and is also effective for many face recognition tasks. Although LPP is designed based on “locality,” like most manifold learning methods, it still suffers from the “over learning of locality” problem, because there is no direct connection with classification in its algorithm. Therefore, on some occasions, it cannot be guaranteed to map an appropriate projection for classification purposes [37].
Subsequently, in order to address this issue and delve into more influential projections for classification tasks, the unsupervised discriminant projection (UDP) method was developed as a simple version of LPP [38]. UDP is considered a linear estimation of multi-manifold-based learning because it considers both the local and nonlocal scatters of data. In both LPP and UDP, the data class label information is not considered, which may degrade their pattern classification performance. Furthermore, the discriminant neighborhood embedding (DNE) method has been presented with the idea of using data class label information [39]. DNE can find a good embedding for classification, considering intra-class absorption and inter-class class expulsion. The main characteristic of DNE is called “discrimination,” meaning the ability to distinguish the same class from distinct classes. This specification of DNE can deal well with “out-of-sample” and “small training sample size” problems. Nevertheless, DNE cannot correctly preserve local information of data because it only concedes +1 and −1 to intra-class and inter-class neighbors [15]. Thus, much of the important geometrical structure information of data may be lost, and it might fail to find out the most significant sub-manifolds for pattern recognition. The locality-based discriminant neighborhood embedding method (LDNE) has recently been proposed to tackle the problems existing in LPP and DNE [15]. This method takes into account both the “locality” and the “discrimination” in a united modelling environment. However, many important non-linear data might be lost during the dimensionality reduction process, which dramatically influences classification accuracy.
According to the way dimensionality reduction algorithms “learn” from data to create predictions, they can be categorized into two different classes: supervised and unsupervised learning methods. Among these two, supervised machine learning is used more prevalently. In this case, the data scholar acts as a guide to instruct the algorithm regarding which results should be found by it [40]. The most well-known supervised algorithms include supervised LPP [39], local discriminant embedding (LDE) [41], neighborhood discriminant projection (NDP) [42], discriminant locality preserving projections (DLPP) [43], locally discriminating projection (LDP) [44], and geometry preserving projections (GPP) [45] It is clear that the aforementioned supervised techniques generally apply class label information in order to amend the dimensionality reduction. On the other hand, the unsupervised LPP-based algorithms generally aim to improve the locality preserving and discriminating capabilities to further enhance the final performance of the classification process. Graph-optimized locality preserving projections (GOLPP) [15,46], orthogonal locality preserving projection (OLPP) [47], and UDP [38] are some examples of the unsupervised LPP-based methods.
In this project, a new supervised subspace learning algorithm named “supervised kernel locality-based discriminant neighborhood embedding” (SKLDNE) is proposed. In this approach, not only can the nonlinear data structure be preserved by applying a kernelized nonlinear mapping method, but also both “locality” and “discrimination” of data in an integrated modeling environment are considered simultaneously. It should be noted that this technique is supervised through direct connection with classification in order to properly guide the procedure of dimensionality reduction. In order to have a reliable and powerful comparison, the efficiency of the proposed SKLDNE technique was compared with PCA, KPCA, LDA, UDP, LPP, DNE, and LDNE by a broad range of experiments with different publicly available face datasets, i.e., Yale face, Olivetti Research Laboratory (ORL) face, Head Pose, and Sheffield. Moreover, Finger Vein and Finger Knuckle databases were also applied to investigate the implementation of our algorithm in other types of databases rather than faces.
It is worthwhile to highlight several characteristics of the proposed approach here:
(1)
SKLDNE has been successfully designed to retain local geometric relations of the within-class samples, which are very important for image recognition. Generally, the categorization strength of methods with a linear learning algorithm is restricted. They fail to deal with complicated problems. Many effective nonlinear data features may be lost during the classification progress using linear techniques such as LDNE, LDA, DNE, and LPP. Therefore, applying a nonlinear method can effectively improve the classification performance.
(2)
This technique is a supervised learning method, as the data scholar acts as a guide to instruct the main algorithm whose conclusion should be found. SKLDNE considers class label information of neighbors in which there is a direct connection with classification, in order to enhance final recognition performance.
(3)
It benefits from the advantages of “locality” in LPP in which, due to the prior class-label information, geometric relations are preserved.
(4)
Not only can it build a compact submanifold by minimizing the distance between the same points in the same class, but it also expands the gaps among submanifolds of distinct classes simultaneously, which is called “discrimination.”
(5)
SKLDNE can resolve the SSS problem that is mostly faced by other aforementioned techniques such as PCA, LDA, UDP, and LPP, and the “overlearning of locality” problem in the manifold learning.
(6)
Due to its kernel weighting, it is very efficient in reducing the negative influence of the outliers on the projection directions, which effectively handles the drawbacks of linear models and makes it more robust to outliers.
The rest of this study is organized as follows: Section 2 categorizes LPP, DNE, and LDNE. Section 3 is devoted to describing the proposed SKLDNE method and the pertinent algorithm and mathematics. Section 4 focuses on the experiments and analyses carried out. Section 5 elucidates conclusions and future research.

2. Outline of LPP, DNE, and LDNE

In this section we will briefly review LPP, DNE, and LDNE, as our proposed method is designed to possess the best characteristics of these techniques in graph embedding.

2.1. Locality Preserving Projection

LPP is one of the successful linear algorithms used for dimensionality reduction that finds graph embedding of data sets. The manifold structure is directly modeled through creating the nearest-neighbor graph, disclosing vicinity relations of data points in order to preserve the local structure of the input data in the projection [48]. LPP works based on a linear approximation of a Laplacian Eigen map, which searches a transformation P in which input data X = [ x 1 , x 2 , …, x n ] with a high dimension is projected into subspace Z with a low-dimension, while the local structure is retained [49]. In order to calculate linear transformation T, the objective function should be minimized as follows [43]:
min P i , j = 1 n z i z j 2 H ( i , j )
where weight matrix H (called heat kernel) is obtained by the nearest-neighbor graph and z i = P T x i . If x i is among l nearest neighbor of x j or x j is among l nearest neighbor of x i , then
H ( i , j ) = e x i x j 2 t
where parameter t is an appropriate constant number. Otherwise, H ( i , j ) = 0 . On the other hand, when x i and x j are the nearest neighbors, weight matrix H could be clearly set as: H ( i , j ) = 1 . Otherwise, H ( i , j ) = 0 . The optimal transformation matrix can be calculated by converting the minimization problem into solving the generalized eigenvalue problem
X L X T P = λ X D X T P
where L = DH is the Laplacian matrix and D i i = j H ( i , j ) is a diagonal matrix.

2.2. Discriminant Neighborhood Embedding

Discriminant neighborhood embedding (DNE) is suggested based on an intuition of a dynamics theory. DNE works under the category of supervised learning, in which multi-class data points are pushed or pulled in a high dimensional space to modulate a favorable embedding of low dimensionality for classification [15]. Furthermore, DNE effectively avoids the complication of a singularity matrix, as there is no need to calculate the inverse matrix. Based on the main characteristics of DNE, it can present a good solution for the small sample size (SSS) and the out-of-sample problems [50]. Although this comprehensive technique is effective in pattern classification, it still cannot uphold the data’s geometrical structure information. The main steps of the DNE algorithm are as follows [39]:
(1)
The adjacent matrix H ¯ of graph G which refers to the underlying supervised manifold structure is as follows:
H ¯ i j = { 1 ,   x i k nn ( j )   or   x j k nn ( i )   and   ( c i c j ) + 1 ,   x i k nn ( j )   or   x j k nn ( i )   and   ( c i = c j ) 0 ,     otherwise
where ci is the class label of xi and knn (i) is the set of k nearest neighbors of xi. Note that each edge is weighed +1 or −1 respectively in order to determine the local intra-class attraction and inter-class repulsion between neighboring points.
(2)
The optimal transformation of matrix P can be defined as follows:
min i j z i z j 2 H ¯ i j
The minimization problem can be reduced to:
argmin tr ( P T X L ¯ X T P )
Subject to P T P = I , where L ¯ = D ¯ H ¯ and D i i ¯ = j H i j ¯ is a diagonal matrix.
Like LPP, parameter P (projection matrix) can be optimized by calculating the minimum eigenvalue solution to the generalized Eigen value problem as follows:
X L ¯ X T P = λ P
where P is constituted by the r eigenvectors corresponding to its first smallest negative eigenvalues of d, i.e., λ 1 λ 2 ≤…≤ λ d < 0 ≤ λ d + 1 .

2.3. Locality-Based Discriminant Neighborhood Embedding

In the multi-class classification assignment, N data points should be classified. The problem that arises here is finding a circumlocutory manifold embedded subspace. Based on the DNE, there are two classifications for the important characteristic of the manifold structure, namely inter-class expulsion, which is the interaction between pairs of neighbors from the same class, and intra-class absorption, meaning the interaction between pairs of neighbors from different classes [15]. These two classes can be defined as follows (Figure 1) [15]:
(1)
Intra-class absorption: the interaction between pairs of neighbors from the same class.
(2)
Inter-class expulsion: the interaction between pairs of neighbors from different classes.
It is possible to classify all data points based on absorption interaction or distracting behavior using these two classes. Therefore, in the subspace, neighbors from a similar class are absorbed, whereas neighbors from a distinct class become detachable. In order to formulate the method, first we consider that x i is a data point, N s ( x i ) is the intra-class neighbors of x i , N d ( x i ) denotes the inter-class neighbor of x i , and N ( x i ) represents all the neighbors of x i . Thus, to actualize this task with the purpose of understanding these two classes, edges between x i and inter-class neighbors and intra-class neighbors are indicated using various weights. Denoted weights are calculated using a kernel function based on the disparity between x i and its nearby neighbors. Diacritical adjacent graph D can be obtained by the ϵ -neighborhood. The diacritical adjacency weight matrix (DAWM) F is described as [15,20]:
F i j = { exp ( x i x j 2 t ) ,   x i N k s ( x i )   or   x i N k s ( x j ) + exp ( x i x j 2 t ) ,   x j N k d ( x i )   or   x i N k d ( x j ) 0 ,     otherwise
where t is the regulator, N k s ( x i ) is the intra-class neighbors of x i , and N k d ( x i ) is the inter-class neighbors of x i in a k-neighborhood. It is obvious that different samples lead to different classification results. In view of the fact that the individual feature space location of a sample indicates its conditions, a parameter has been defined to regulate adjacent weight between pairs of neighbors. This regulator can be formulated as:
t = 1 k Σ j = 1 k x i x j 2
To gain intra-class absorption and inter-class expulsion in the transformed space, it is recommended to apply a linear mapping method to project the intra-class absorption and inter-class abhorrence of input data points. As a result, the new low dimensional space can be defined as:
(1)
Intra-class absorption:
Φ ( P ) = Σ i j y i y j 2 W i j = Σ i j P T x i P T x j 2 W i j , x j N k ( x i )   or   x i N k ( x j )   and   ( c i = c j )
(2)
Inter-class expulsion:
Ψ ( P ) = Σ i j y i y j 2 W i j = Σ i j P T x i P T x j 2 W i j ,
x j N k ( x i )   or   x i N k   and   ( c i c j )
Finally, the difference between weighted distancse from each data point to the inter-class neighbors in N d ( x i ) and those from x i to the intra-class neighbors in N s ( x i ) in the mapped space must be calculated and, by maximizing this measurement, we can obtain the optimum result. This measurement can be referred to as a margin, which is calculated as follows:
Θ ( P ) = Ψ ( P ) Φ ( P )
Thus, if the original data points are close together, this margin can keep the projected data points as close as possible. However, we can prevent x i and x j from being mapped far apart if they are close by defending the retribution generation [15]:
Θ ( P ) = Σ i j P T x i P T x j 2 F i j

3. Supervised Kernel Locality-Based Discriminant Neighborhood Embedding

3.1. Main Idea

As already mentioned in Section 1, DNE cannot correctly preserve local information of data because it only concedes +1 to intra-class and −1 to inter-class neighbors, so it might fail to find out the most significant submanifold for pattern classification. In addition, LPP is designed based on “locality” since it has no direct connection with classification, and it still suffers from the “over learning of locality” problem [17]. Therefore, LDNE has been proposed to overcome the problems existing in LPP and DNE, considering both the “locality” and the “discrimination” in a unified modeling setting. However, it does not guarantee an appropriate projection for classification purposes, because many important non-linear data might be lost during its dimensionality reduction process. In some cases, LDNE also cannot distinguish inter-class and intra-class neighbors properly in order to conduct projection for all points, which can degrade the classification performance. In order to address these problems, we propose a new supervised subspace learning method named “supervised kernel locality-based discriminant neighborhood embedding” (SKLDNE). Combined with nonlinear data structure, locality and discrimination information, SKLDNE can yield an optimal subspace that best finds the indispensable submanifolds-based structure.
In our proposed SKLDNE, we first use nonlinear kernel mapping to represent the input data in implied feature space F. Afterwards, a linear transformation is searched to retain within-class geometric structures in the feature space. Hence, we can achieve a nonlinear subspace that is able to estimate the essential geometric structure of the face manifold. As a matter of fact, the proposed SKLDNE is modeled to take the nonlinear data in the feature space while important features of data including “locality” and “discrimination” are simultaneously preserved. In order to clearly elucidate the performance of our SKLDNE, we have compared it with several dimensionality reduction techniques including PCA, KPCA, LDA, UDP, LPP, DNE, and LDNE on six different publicly available datasets.

3.2. Mathematics

Suppose X = [ x 1 ,   x 2 , …, x n ] is a set of d-dimensional input samples. This input data is projected onto a higher dimensional feature space F via nonlinear mapping Ø: R n Ϝ. Then, manifold learning is then carried out on the projected samples ( X ) = [ ( x 1 ) , ( x 2 ) , , ( x n ) ] . Now assume that we have to find a projection transformation V in Ϝ. The optimization problem can be expressed as:
Maximize   i j n z i z j 2 F i j
Subject to V T V = I , where I denotes the identity matrix, z i = v T ( x i ) and z j = v T ( x j ) are the projection of ( x i ) and ( x j ) respectively onto V , and F i j represents the relationship between of x i and x j . The optimization problem can be kernelized as:
i j n z i z j 2 F i j = i j n v T ( x i ) v T ( x j ) 2 F i j
This equation can be rewritten from the square of norm in Equation (15) into the form of trace as follows:
i j n v T ( x i ) v T ( x j ) 2 F i j         = tr   { i j n ( v T ( x i ) v T ( x j ) ) ( v T ( x i ) v T ( x j ) ) T F i j }         = tr   { v T i j n ( 2 ( x i ) F i j ( x j ) T ( 2 ( x j ) F i j ( x i ) T   F i j ) V }
The linear transformation should lie in the span of ( x 1 ) , ( x 2 ) , , ( x n ) and α = [ α 1 , α 2 , ,   α n ] are the expansion coefficient vectors and
V = i = 1 n α i ( x i ) = ( X )   α
Substituting (17) into (16) we obtain:
U = 2 tr   { v T ( X ) ( D F ) ( X ) T V } = 2 tr   { v T ( X ) L   ( X ) T V } = 2 l = 1 m { v l T ( X ) L   ( X ) T V l }
where D i i = j F ( i , j ) is diagonal matrix and L = DF, L and D are the symmetric matrix and represent the number of eigenvalues. The optimization problem can be rewritten as:
Maximize   tr   { v T ( X ) L   ( X ) T V } = α T K L K α
Subject to α T K α = I , where K is a kernel matrix with k ( x i , x j ) = [   ( x i ) . ( x j ) ] , and a kernel in matrix form is:
K = ( X ) T ( X )
The corresponding generalized eigenvalue problem can be solved through calculating the maximum eigenvalues in ( X ) L   ( X ) T V = λ V , where the generalized eigenvector equivalent to the biggest eigenvalue is the main interest. Finally, we need to compute the dot product through a kernel and find its nearest neighbor in the embedding space. It should be noted that, although SKLDNE and other methods like DNE consider geometrical information of both intra-class and inter-class, the construction of their contiguous graphs and weights are done in very diverse ways.

4. Biometrics Application: Results and Analysis

In this part, to have a reliable and powerful comparison, the implementation of the suggested SKLDNE technique was compared with PCA, KPCA, LDA, UDP, LPP, DNE, and LDNE by broad comparative experiments on six different publicly available datasets, i.e., Sheffield, Yale face, ORL face, Head Pose, Finger Vein, and Finger Knuckle Databases [15]. For each dataset, depending on the number of data for each class, some samples were randomly chosen as training samples, whereas the rest of that class was chosen for testing. Furthermore, in the recognition phase [15], due to simplicity and in order to make the recognition outcome favorable, the nearest neighbor was utilized as a classifier. K-neighborhood variable k for computing the weight matrix is marked by Wk in the posterior explanations. In all the experiments, to achieve fair comparisons, Wk was selected as Wk = Tn − 1 (where Tn is training number samples of every class). Based on our experiments, all aforementioned methods achieved the optimal recognition rate with this value of Wk. It should be noted that there are two different kernels called Polynomial and Gaussian kernel. Based on our experimental results, the Polynomial has shown better performance than the Gaussian one. Therefore, only the polynomial kernel type was applied in this research.

4.1. Experimental Results with the Sheffield Face Database

The Sheffield Face Database (SH.F) [51] includes a total of 20 individuals, with 564 images in which each subject is illustrated, including poses ranging from profile to forehead views. The format of all images is PGM, image size is approximately 220 × 220 pixels in 256-bit grayscale. Figure 2 illustrates a one-subject sample with different poses from the Sheffield face multi view. In our experimentation, every image was resized to 112 × 92 pixels [52]. The maximal rate of recognition of each method and the related dimension implemented in the Sheffield database are illustrated in Table 1. The best performance result among other methods is assigned in bold-text. In addition, in all experiments due to the large number of implementations, it was decided to select some training and testing images that were more challenging for the classification task, in order to evaluate the performance of each technique in these critical areas. Thus, considering the small training sample size problem (SSS), first we selected a small number of training samples and then some large numbers to evaluate the performance of our SKLDNE method accurately. Tn = 1, 2, 3, 4, 5, 6, 7, 8, 15, 16, 17 were the training sample numbers selected from each class.
According to Table 1, three main conclusions can be drawn. First, SKLDNE significantly outperformed other methods (PCA, KPCA, UDP, LPP, LDA, DNE, and LDNE) over an extensive range of dimensionality for all different numbers of training and testing images, whether the training sample size was large or small. As can be observed, when the training sample number was small, SKLDNE clearly behaved more efficiently than all other recognition techniques, which proves the robustness of this method in the case of the small training sample size problem (SSS). Secondly, it can be seen that the recognition rates of all implementations were better when more training samples are used. Third, as shown in Figure 3, SKLDNE provided the best results at the lowest number of dimension values compared to its competitors. In addition, when dimensionality increased to about 30, the recognition accuracy of each method first increased rapidly and then became stable. The differences between the recognition rate of SKLDNE and other methods were emphasized when the training sample number was very small. However, for the larger training sample, the mentioned differences in the classification rate increased. For instance, in the training number of 1, 2, 3, SKLDNE achieved much better results than others. Moreover, for Tn = 17, SKLDNE reached 100% recognition rate, while the results were much lower for other methods. SKLDNE can effectively yield an optimal embedding subspace that finds substantial submanifolds-based structure with lower dimensionality. The within-class local structure, which is very important for face recognition, can be preserved simultaneously in a nonlinear kernel feature space. SKLDNE can solve the “out-of-sample” problem and the “overlearning of locality” problem in manifold learning, which other aforementioned methods often face. Therefore, based on these characteristics, it can be concluded that the recommended SKLDNE technique is a promising technique to be used for dimensionality reduction, with a very satisfactory performance in classification when dealing with high-dimensional data.

4.2. Experimental Results with the Yale Face Database

The Yale face database [53] consists of 165 images (grayscale) taken from 15 people [54] under diverse facial expressions and light conditions. The data include 11 images for every individual, with various facial expressions or environmental conditions, such as: normal, wearing glasses and no glasses, center-light, left and right light, happy, winking, sleepy, surprised, and sad. In the experimental results, every image was first cropped and then was resized to 32 × 32 pixels [55]. Figure 4a,b illustrates some sample images of one individual of the Yale database and corresponding cropped images.
PCA was used in all methods for feature extraction, and all methods include a PCA phase. The optimum rate of recognition of each technique and the equivalent dimension implemented in the Yale database are illustrated in Table 2. Furthermore, in all experiments, due to the large quantity of implementations, it was decided to select some training and testing images that were more challenging for the classification task to clarify the performance of the aforementioned dimensionality reduction techniques in these critical areas. Considering the small training sample size problem (SSS), we first selected a training number of 1 sample and then some larger numbers from 6 to 9 to evaluate the performance of our SKLDNE method in some common existing problems such as the SSS problem and the out-of-sample problem.
Table 2 shows that the SKLDNE method achieved the highest accuracy in 100% of implementations in the Yale Database. In Figure 5a–e, the comparative classification accuracies are plotted for each given Tn (training number) in each dataset through changing the dimensionality of the transformation matrix. As is shown, the recognition rate of each technique increased promptly till the dimensionality was almost 40, and then it stabilized. It can be observed in Table 2 that SKLDNE was implemented more efficiently than others among a wide variety of dimensionality in the Yale Face database. Meanwhile, the best implementation of SKLDNE was achieved at a smaller dimension value in most of training numbers for each data set compared to LDNE. Moreover, differences in the classification between SKLDNE and other methods are very clear, especially when training number was small, for instance, Tn = 1, 2. For training number 7, SKLDNE yielded an improvement of around 4.5% compared with DNE, LPP, LDA, and KPCA, and 6.6% in comparison with LDNE, UDP, and PCA respectively. For training number 9, both SKLDNE and LDNE gained 100% accuracy, while accuracies in other methods with the same training number were much lower. In order to explain the superiority of the proposed method, our SKLDNE first mapped the data in the kernel space to capture the substantial extracted data and then both geometrical and discriminant information of the data were taken, benefiting from a significant form of the affinity weight matrix to embed the graph. Although LPP, DNE, and LDNE outperforming PCA, KPCA, and UDP demonstrates that the discriminant and local data structure based methods are more suitable for face recognition, our SKLDNE had more nonlinear data representation, locality preservation, and discriminating power than other methods, and consequently achieved the best recognition accuracy. Therefore, based on the mentioned characteristics of SKLDNE, it can be concluded that our SKLDNE can overcome the “SSS,” “out-of-sample,” and “overlearning” problems.

4.3. Experimental Results with the ORL Database

The ORL face database [56] includes ten diverse grayscale images taken from 40 different individuals [57]. All 10 face images of each subject were captured at different times, with changes in the lighting, facial details (with glasses or no glasses) or facial expressions (smiling/not smiling, open/closed eyes,), against a dark homogeneous background, and with straight and frontal views. The size of all images is equal to 92 × 112 pixels in PGM format. The images were captured in 40 directories [58]. It should be noted that preprocessing was used and all original images were already cropped and resized. In this project, the size of 32 × 32 pixels was chosen for all ORL images. Figure 6 illustrates three different subjects (each with 10 images) from the ORL database.
In our experiments, the number of training samples Tn = 1, 4, 3, 4, 5, 6, 7, 8 were chosen from the dataset related to each subject to make the training sample set. The other numbers of images are applied as a testing sample set. As already mentioned, PCA was used in the classification phase in all methods. The maximal average accuracy (in percentage terms) and its corresponding dimension, followed by the alteration in the training sample sizes, are illustrated in Table 3. It can be observed in this yable and in Figure 7a–f that SKLDNE generally outperformed LDNE, whether the number of training sample size was small or not, in almost smaller numbers of dimensions. Moreover, as a supervised method, SKLDNE also significantly outperformed other techniques (KPCA, LPP, DNE, UDP, and LDA) regardless of the change in the training sample size. Compared to other techniques, SKLDNE performed better in small training sample size case. Furthermore, when the training number was equal to 8, SKLDNE had a zero error rate compared to that of PCA (4.1%), KPCA (3.5%), UDP (4%), LPP (3.5%), LDA (4%), DNE (3.75), and LDNE (3.5%). SKLDNE can simultaneously discover interclass and intraclass geometrical information, and have more nonlinear data representation, locality preservation, and discriminating power than other techniques. Therefore, SKLDNE does have merit over other techniques in terms of resolving classification problems in face recognition. This characteristic of SKLDNE in small sample size cases is indeed important to improve the recognition rate in practice, since face recognition is commonly a small sample size problem. Normally, a small number of images of each person is accessible in many real-world tasks.

4.4. Experimental Results with the Head Pose Database

The Head Pose database (Figure 8) [59] includes 2790 images of faces taken from 15 individuals with pan and tilt variation angles from −90 to +90 degrees. For each individual, 2 sets with 93 different positions were captured [60].
As can be observed in Table 4, SKLDNE performed better than the other seven methods, regardless of the variation in the training sample size. The maximal recognition rate of SKLDNE when Tn = 130 was up to 99.28%, while for other methods it was much lower. This reveals that, when the given training sample size for each class gets larger, SKLDNE can obtain much better results than other methods. Two more points can also be outlined. First, our supervised method with kernel weighting can notably enhance the class classification performance, but applying the kernel trick has no significant influence on PCA performance. Second, SKLDNE achieves optimal recognition rates at almost smaller number of dimensions as the recognition rate of SKLDNE retains the best results as the dimension varies from 14 to 30. Compared to the other techniques, SKLDNE preserves the more discriminative and local features of face images. It also preserves more local geometric relations of the within-class samples by nonlinear kernel mapping. It should be noted that linear methods such as LPP, LDA, LDNE, DNE, and UDP often fail to deliver good classification performance when face images are subject to complex nonlinear changes such as expression, lighting, pose and so on. Figure 9 indicates that the recognition implementations of all methods first sharply increase while the projected dimensions are added, and then, after obtaining the optimum, they tend to become stable.

4.5. Experimental Results with the Finger Vein Database.

The finger vein database used in this project was collected from 51 individuals (male and female) who were aged between 21 and 56 [61]. 10 images were captured from each subject. Four fingers were used for capturing, including right and left middle finger and right and left index finger. There are 204 different fingers in the database, and the data consist of 2040 images in total, in which each finger image originally had a dimension of 480 × 160 pixels. In our implementations, each image was resized to 32 × 32. The captured images from one person can be seen in Figure 10.
In this section, the performance of each technique followed by a maximal recognition rate was evaluated through changing the dimensionality of transformation matrix. The number of training samples Tn = 1, 2, 3, 4, 5, 6, 7 was chosen from the image gallery to make the training sample set, and the rest of images were selected for testing [62]. Furthermore, the PCA classifier (nearest neighbor NN using the Euclidean distance) was applied in the recognition phase, and the best recognition rates of each technique are indicated in Table 5.
Based on the experiment results shown in Figure 11, our SKLDNE gained the best recognition rate among all different training numbers, which proves that it has an encouraging performance compared to other advanced methods. Regarding the small training sample size case, the SKLDNE method still showed that it performed significantly better than other techniques, as the maximal recognition accuracy rate of SKLDNE in training number 2 was almost 5% more than PCA, KPCA, UDP, LPP, and DNE, and 1.2% more than LDNE. Our SKLDNE was always able to represent its optimal embedding space with a lower value of dimensions in comparison with the other seven techniques. For example, SKLDNE for Tn = 7 achieved 100% recognition accuracy in the smallest value of projected dimensions (26). This conveys that our approach is more effective, due to its significant characteristic that it can not only represent nonlinear and complex variations in of images, but can also model both locality of LPP and discrimination of DNE simultaneously. This fact demonstrates the good performance of our proposed method.

4.6. Experimental Results with the Finger Knuckle Print Database

This database is organized by the Polytechnic University of Hong Kong [63], and is freely available online. Based on the database description, finger knuckle print (FKP) images were collected from 165 individual volunteers (males and females) [64]. The samples were collected in two distinct sessions, and in each one 6 images were captured from 4 fingers (including left and right index finger and the left and right middle finger). Therefore, 7920 finger images in total were taken from 660 various fingers. In our experiments, the original image of the database is cropped and is then resized to 32 × 32 pixels. Figure 12 illustrates a cropped sample of the FKP database.
Table 6 shows that the SKLDNE recognition performance was significantly more efficient than other techniques, regardless of the variation in the training sample size, in the Finger Knuckle database. The recognition rate of SKLDNE when Tn = 11 was equal to 100%, while for the LDNE it was equal to 97.2%. Another point that is worth mentioning on the comparison of recognition performance in SKLDNE and other methods is related to the small training sample size case, as SKLDNE had the best performance in this respect. Again, all best recognition performances of SKLDNE were almost achieved at smaller values of dimension on every Tn per data set. It can also be observed that, when the given training sample size of each class became larger, SKLDNE achieved much better results than other techniques. For example, for Tn = 6, the accuracy of SKLDNE was around 8% and 21% more than LDNE and PCA, respectively. In order to well explain the superiority of SKLDNE, it should be noted that it can preserve more effective nonlinear features and more geometrical and discriminant information so, it can tackle the small sample size, the out-of-sample, and the “overlearning of locality” problems. Hence, it can be concluded that the SKLDNE approach is a promising supervised technique with a satisfactory performance of classification in order to be applied in the Finger Knuckle database. The recognition rates in comparison with the variation of dimensions for each Tn are shown in Figure 13a–i.

4.7. Classification Performance and Computational Cost

In this section, by means of a new group of experiments, we evaluated SKLDNE performance through changing k-neighborhood variation Wk (from 1 to 30 with a scale of 2). The number of training samples selected for each database were Tn = 4, 5, 6, 7, 8, 15, 16, 17 in Sheffield, Tn = 2, 6, 7, 8, 9 in Yale, Tn = 5, 6, 7, 8, 9 in ORL, Tn = 120, 130, 140, 150, 160 in Head Pose, Tn = 5, 6, 7, 8, 9 in Finger Vein, and Tn = 7, 8, 9, 10, 11 in Finger Knuckle databases. It should be noted that the rest of each class in each dataset was used for testing. The training number samples were randomly selected. The maximum recognition rates of SKLDNE in comparison with Wk for different numbers of training samples are indicated in Figure 14a–f.
It can be seen that the classification performance became better as the number of training samples increased. The classification performance of SKLDNE improved first with a rise of Wk until almost Wk = 9 in the Sheffield and Head Pose, and then it decreased dramatically. In the Yale, ORL, Finger Vein, and Finger Knuckle datasets, it can easily be observed that the recognition performance of SKLDNE enhanced rapidly when Wk varied from 1 to 7 in Yale, 1 to 10 in ORL, 1 to 9 in Finger Vein, from 1 to 6 in Finger Knuckle, and then it decreases when Wk became larger, since large values of k-neighborhood variable Wk have an effect on creating the weight adjacent matrix. It has already been proven that the k-neighborhood selected for data point might contain more outliers belonging to other classes at a large number of Wk when a dataset includes many classes with a small number of samples for each class. Thus, the constructed adjacent weight matrix does not have sufficient discrimination for image recognition [15].
The experiment for analyzing computational cost was carried out on an Intel(R) Core i5-4200U CPU, 2.3 GHz, 10 GB RAM machine using MATLAB (R2016, Natick, MA, USA). The computational costs of the different classification methods using the Yale data base are listed in Table 7.
From the results shown in Table 7, it can be observed that the proposed SKLDNE was faster than its main competitors, such as LDNE, DNE, and LPP. The processing times of PCA, KPCA, UDP, and LDA were lower. However, the recognition rate results illustrate that these methods were much less accurate than the SKLDNE method.
Now, to have a more reliable comparison, we briefly compare our recognition results of the proposed method with previously published works, including a deep learning method named Deep Belief Networks (DBNs) [65,66] with a traditional multilayer perceptron model (MLP) [66] in the used facial expression databases, i.e., the JAFFE Database [67]. As a deep learning method, DBNs have an unsupervised feature learning ability. The JAFFE database includes 10 individuals (Japanese women) with 7 different expressions, and has around 3 or 4 images for each expression. There are 213 images in total in this database. Each image has a resolution pixel of 256 × 256. In detail, we divided all image samples into 10 parts, 90% of which were applied for training, and the remaining were applied for testing. Table 8 illustrates the recognition performance comparisons in the JAFFE database when dealing with three different image resolutions of 16 × 16, 32 × 32, and 64 × 64. We can clearly see that the proposed SKLDNE method achieved the best recognition performance (100% in all cases), in comparison with other previously reported results [66], which are much lower. This is attributed to the main characteristics of SKLDNE, which effectively represents more nonlinear data structure and has more locality and discrimination information preserving power. The results again show the robustness of SKLDNE for facial expression recognition.

5. Conclusions and Future Research

In this study, the performance of several well-known pattern recognition strategies was analyzed to clarify which techniques are best suited to be applied in face recognition. We also analyzed the weakness and robustness of each technique. As already mentioned, DNE cannot correctly preserve local information of data because it only assigns +1 to intra-class and −1 to inter-class neighbors, so it might fail to discover the most significant submanifolds for pattern classification. LPP is designed based on “locality” since it has no direct connection with classification, and it still suffers from the “over learning of locality” problem. LDNE has been proposed to overcome the problems existing in LPP and DNE; however, it does not guarantee an appropriate projection for classification purposes because many important non-linear data might be lost during its dimensionality reduction process. In addition, in some cases, LDNE cannot distinguish inter-class and intra-class neighbors well either in order to conduct projection for all points. This can degrade the classification performance. In order to address these problems, we have proposed a new supervised subspace learning algorithm named “supervised kernel locality-based discriminant neighborhood embedding” (SKLDNE). Combined with nonlinear data structures, locality, and discrimination information, SKLDNE can yield an optimal subspace that best finds the indispensable submanifolds-based structure. Six publicly available datasets, i.e., Yale face, ORL face, Sheffield, Head Pose, Finger Vein and Finger Knuckle, were used to illustrate the significance of the proposed technique. Experimental results reveal that SKLDNE outperforms and demonstrates potential to be implemented in real world systems compared to other advanced dimensionality reduction methods by obtaining the highest recognition rates in all experiments. Representing complex nonlinear variations makes SKLDNE more powerful and more intuitive than LDNE and other aforementioned techniques in terms of classification. According to the results, SKLDNE could also resolve the “small training sample size” problem, and it had the best performance compared to others at smaller numbers of projected dimensions in each number of training samples per data set. Moreover, when the given training sample size for each class grew larger, SKLDNE also achieved much better results than other techniques. The overlearning of locality problem and the out-of-sample problem in manifold learning can be avoided by applying our developed classifier. As a future plan, we will modify this classifier in order to make it directly applicable for two-dimensional data to effectively reduce the computational costs.

Author Contributions

Conceptualization, Methodology and Software, A.K.M.; Data and Formal Analysis, A.K.M. and J.J.C.A.; Resources, A.K.M., J.A.C.C. and J.J.C.A.; Writing—Original Draft Preparation, A.K.M.; Writing—Review & Editing, A.K.M., J.A.C.C. and J.J.C.A.; Visualization, A.K.M.; Supervision, J.A.C.C. and J.J.C.A.; Project Administration, J.A.C.C. and J.J.C.A.; Funding Acquisition, J.A.C.C. and J.J.C.A.

Funding

The APC was funded by the University of Malaga.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Yang, M.-H.; Ahuja, N.; Kriegman, D. Face recognition using kernel eigenfaces. In Proceedings of the Image Processing, Vancouver, BC, Canada, 10–13 September 2000. [Google Scholar]
  2. Belhumeur, P.N.; Hespanha, J.P.; Kriegman, D.J. Eigenfaces vs. Fisherfaces: Recognition Using Class Specific Linear Projection; Yale University: New Haven, CT, USA, 1997. [Google Scholar]
  3. Yanagawa, Y.; Sakuragi, M.; Minato, Y. Face Identification Device. U.S. Patent No. 7,853,052, 14 December 2010. [Google Scholar]
  4. Yang, G.; Xi, X.; Yin, Y. Finger vein recognition based on a personalized best bit map. Sensors 2012, 12, 1738–1757. [Google Scholar] [CrossRef]
  5. Rosdi, B.A.; Shing, C.W.; Suandi, S.A. Finger vein recognition using local line binary pattern. Sensors 2011, 11, 11357–11371. [Google Scholar] [CrossRef]
  6. Mobarakeh, A.K.; Rizi, S.M.; Khaniabadi, S.M.; Bagheri, M.A.; Nazari, S. Applying Weighted K-nearest centroid neighbor as classifier to improve the finger vein recognition performance. In Proceedings of the 2012 IEEE International Conference on Control System, Computing and Engineering (ICCSCE), Penang, Malaysia, 23–25 November 2012. [Google Scholar]
  7. Turk, M.A.; Pentland, A.P. Face recognition using eigenfaces. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Maui, HI, USA, 3–6 June 1991. [Google Scholar]
  8. Zheng, Z.; Yang, F.; Tan, W.; Jia, J.; Yang, J. Gabor feature-based face recognition using supervised locality preserving projection. Signal Process. 2007, 87, 2473–2483. [Google Scholar] [CrossRef]
  9. Jain, A.K.; Duin, R.P.; Mao, J. Statistical pattern recognition: A review. IEEE Trans. Pattern Anal. Mach. Intell. 2000, 22, 4–37. [Google Scholar] [CrossRef]
  10. Webb, A.R. Statistical Pattern Recognition; John Wiley & Sons: Hoboken, NJ, USA, 2003. [Google Scholar]
  11. Liu, Q.; Hunag, R. Face recognition using kernel-based fisher discriminant analysis. In Proceedings of the Fifth IEEE International Conference on Automatic Face and Gesture Recognition, Washington, DC, USA, 21 May 2002. [Google Scholar]
  12. Schouten, B.; Jacobs, B. Biometrics and their use in e-passports. Image Vis. Comput. 2009, 27, 305–312. [Google Scholar] [CrossRef]
  13. Mobarakeh, A.K. Finger Vein Recognition Using Local Mean Based K-Nearest Centroid Neighbor Classifier. Adv. Mater. Res. 2013, 628, 427–432. [Google Scholar] [CrossRef]
  14. Damavandinejadmonfared, S.; Mobarakeh, A. Evaluate and determine the most appropriate method to identify finger vein. Procedia Eng. 2012, 41, 516–521. [Google Scholar] [CrossRef]
  15. Gou, J.; Yi, Z. Locality-based discriminant neighborhood embedding. Comput. J. 2013, 56, 1063–1082. [Google Scholar] [CrossRef]
  16. Damavandinejadmonfared, S.; Mobarakeh, A. Finger vein recognition using PCA-based methods. World Acad. Sci. Eng. Technol. 2012, 66. [Google Scholar] [CrossRef]
  17. Niu, B. Statistical Pattern Recognition: Locality Preserving Embeddings and Ensemble of Rules. Ph.D. Thesis, Hong Kong Polytechnic University, Hong Kong, China, 2008. [Google Scholar]
  18. Yu, H.; Yang, J. A direct LDA algorithm for high-dimensional data—With application to face recognition. Pattern Recognit. 2001, 34, 2067–2070. [Google Scholar] [CrossRef]
  19. Yang, J.; Yang, J.-Y. Why can LDA be performed in PCA transformed space? Pattern Recognit. 2003, 36, 563–566. [Google Scholar] [CrossRef]
  20. Mobarakeh, A.K.; Carrillo, J.A.C.; Agui-lar, J.J.C.; Khaniabadi, S.M.; Nazari, S. Face Recognition using a Newly Developed Linear Subspace Learning Method. J. Telecommun. Electron. Comput. Eng. (JTEC) 2017, 9, 141–144. [Google Scholar]
  21. Pratihar, D.K. Non-Linear Dimensionality Reduction Techniques. In Encyclopedia of Data Warehousing and Mining, 2nd ed.; IGI Global, Dilip Kumar Pratihar (Indian Institute of Technology): Kharagpur, India, 2009; pp. 1416–1424. [Google Scholar]
  22. Shawe-Taylor, J.; Cristianini, N. Kernel Methods for Pattern Analysis; Cambridge University Press: Cambridge, UK, 2004. [Google Scholar]
  23. Schölkopf, B.; Smola, A.; Müller, K.-R. Kernel principal component analysis. In International Conference on Artificial Neural Networks; Springer: Berlin/Heidelberg, Germany, 1997. [Google Scholar]
  24. Blackburn, J.; Ribeiro, E. Human motion recognition using isomap and dynamic time warping. In Human Motion–Understanding, Modeling, Capture and Animation; Springer: Berlin/Heidelberg, Germany, 2007; pp. 285–298. [Google Scholar]
  25. Ning, X.; Li, W. BULDP: Biomimetic Uncorrelated Locality Discriminant Projection for Feature Extraction in Face Recognition. IEEE Trans. Image Process. 2018, 27, 2575–2586. [Google Scholar] [CrossRef]
  26. Roweis, S.T.; Saul, L.K. Nonlinear dimensionality reduction by locally linear embedding. Science 2000, 290, 2323–2326. [Google Scholar] [CrossRef]
  27. Belkin, M.; Niyogi, P. Laplacian eigenmaps for dimensionality reduction and data representation. Neural Comput. 2003, 15, 1373–1396. [Google Scholar] [CrossRef]
  28. Tang, L.; Lei, Y. Dimensionality reduction based on minimax risk criterion for face recognition. In Proceedings of the 2010 International Joint Conference on Neural Networks (IJCNN), Barcelona, Spain, 18–23 July 2010. [Google Scholar]
  29. Huang, P.; Chen, C. Feature extraction using local structure preserving discriminant analysis. Neurocomputing 2014, 140, 104–113. [Google Scholar] [CrossRef]
  30. Kouropteva, O.; Okun, O.; Pietikäinen, M. Supervised locally linear embedding algorithm for pattern recognition. In Proceedings of the Iberian Conference on Pattern Recognition and Image Analysis, Mallorca, Spain, 4–6 June 2003. [Google Scholar]
  31. De Ridder, D.; Loog, M.; Reinders, M.J. Local fisher embedding. In Proceedings of the 17th International Conference on Pattern Recognition, Cambridge, UK, 23–26 August 2004. [Google Scholar]
  32. Vlassis, N.; Motomura, Y.; Kröse, B. Supervised dimension reduction of intrinsically low-dimensional data. Neural Comput. 2002, 14, 191–215. [Google Scholar] [CrossRef]
  33. Schölkopf, B.; Smola, A.; Müller, K.-R. Nonlinear component analysis as a kernel eigenvalue problem. Neural Comput. 1998, 10, 1299–1319. [Google Scholar] [CrossRef]
  34. Baudat, G.; Anouar, F. Generalized discriminant analysis using a kernel approach. Neural Comput. 2000, 12, 2385–2404. [Google Scholar] [CrossRef]
  35. Yang, J.; Frangi, A. KPCA plus LDA: A complete kernel Fisher discriminant framework for feature extraction and recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2005, 27, 230–244. [Google Scholar] [CrossRef]
  36. Li, S.Z. Encyclopedia of Biometrics: I-Z; Springer Science & Business Media: Berlin/Heidelberg, Germany, 2009; Volume 2. [Google Scholar]
  37. Lu, J.; Tan, Y.-P. Regularized locality preserving projections and its extensions for face recognition. IEEE Trans. Syst. Man Cybern. Part B (Cybern.) 2010, 40, 958–963. [Google Scholar]
  38. Deng, W. Comments on “Globally Maximizing, Locally Minimizing: Unsupervised Discriminant Projection with Application to Face and Palm Biometrics”. IEEE Trans. Pattern Anal. Mach. Intell. 2008, 30, 1503–1504. [Google Scholar] [CrossRef]
  39. Zhang, W.; Xu, X. Discriminant neighborhood embedding for classification. Pattern Recognit. 2006, 39, 2240–2243. [Google Scholar] [CrossRef]
  40. Kotsiantis, S.B.; Zaharakis, I.; Pintelas, P. Supervised machine learning: A review of classification techniques. Emerg. Artif. Intel. Appl. Comput. Eng. 2007, 160, 3–24. [Google Scholar]
  41. Chen, H.-T.; Chang, H.-W.; Liu, T.-L. Local discriminant embedding and its variants. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, San Diego, CA, USA, 20–25 June 2005. [Google Scholar]
  42. You, Q.; Zheng, N. Neighborhood discriminant projection for face recognition. In Proceedings of the 18th International Conference on. Pattern Recognition, Hong Kong, China, 20–24 August 2006. [Google Scholar]
  43. Lu, G.-F.; Lin, Z.; Jin, Z. Face recognition using discriminant locality preserving projections based on maximum margin criterion. Pattern Recognit. 2010, 43, 3572–3579. [Google Scholar] [CrossRef]
  44. Kishore, K.; Kumar, K.H. Hybrid face recognition with locally discriminating projection. In Proceedings of the International Conference on Signal Acquisition and Processing, Bangalore, India, 9–10 February 2010. [Google Scholar]
  45. Wang, T.; Xia, T.; Hu, X.-M. Geometry preserving projections algorithm for predicting membrane protein types. J. Theor. Biol. 2010, 262, 208–213. [Google Scholar] [CrossRef]
  46. Zhang, L.; Qiao, L.; Chen, S. Graph-optimized locality preserving projections. Pattern Recognit. 2010, 43, 1993–2002. [Google Scholar] [CrossRef]
  47. Shao, J.-D.; Rong, G.; Lee, J.M. Generalized orthogonal locality preserving projections for nonlinear fault detection and diagnosis. Chemom. Intell. Lab. Syst. 2009, 96, 75–83. [Google Scholar] [CrossRef]
  48. Wang, H.; Chen, S.; Hu, Z.; Zheng, W. Locality-preserved maximum information projection. IEEE Trans. Neural Netw. 2008, 19, 571–585. [Google Scholar] [CrossRef]
  49. Cheng, J.; Liu, Q. Supervised kernel locality preserving projections for face recognition. Neurocomputing 2005, 67, 443–449. [Google Scholar] [CrossRef]
  50. Wan, M.; Lai, Z.; Jin, Z. Two-Dimensional Local Graph Embedding Analysis (2DLGEA) for Face Recognition. In Proceedings of the Chinese Conference on Pattern Recognition, Beijing, China, 22–24 October 2008. [Google Scholar]
  51. The UMIST Face Database. Available online: http://www.sheffield.ac.uk/eee/research/iel/research/face (accessed on 5 April 2019).
  52. Li, W.; Ruan, Q.; Wan, J. Semi-supervised dimensionality reduction using estimated class membership probabilities. J. Electron. Imaging 2012, 21, 043010. [Google Scholar] [CrossRef]
  53. The Yale Face Database. Available online: http://vision.ucsd.edu/datasets/yale_face_dataset_original/yalefaces.zip (accessed on 5 April 2019).
  54. Chen, C.; Shi, J.; Yu, Y. Locally supervised discriminant analysis in Kernel space. In Proceedings of the 2009 International Conference on Computational Intelligence and Security, Beijing, China, 11–14 December 2009. [Google Scholar]
  55. Zhao, D.-L.; Liu, C.-Q.; Zhang, Y.-H. Discriminant feature extraction using dual-objective optimization model. Pattern Recognit. Lett. 2006, 27, 929–936. [Google Scholar] [CrossRef]
  56. The ORL Face Database. Available online: http://www.cl.cam.ac.uk/research/dtg/attarchive/facedatabase.html (accessed on 5 April 2019).
  57. Zhang, T.; Fang, B. Topology preserving non-negative matrix factorization for face recognition. IEEE Trans. Image Process. 2008, 17, 574–584. [Google Scholar] [CrossRef]
  58. Wan, H.; Guo, G.; Wang, H.; Wei, X. A new linear discriminant analysis method to address the over-reducing problem. In Proceedings of the International Conference on Pattern Recognition and Machine Intelligence, Warsaw, Poland, 30 June–3 July 2015. [Google Scholar]
  59. The ORL Face Database. Available online: http://www-prima.inrialpes.fr/perso/Gourier/Faces/HPDatabase-old.html (accessed on 5 April 2019).
  60. Damavandinejadmonfared, S.; Al-Arashi, W.H.; Suandi, S.A. Pose invariant face recognition for video surveillance system using kernel principle component analysis. In Proceedings of the Fourth International Conference on Digital Image Processing (ICDIP 2012), International Society for Optics and Photonics, Kuala Lumpur, Malaysia, 8 June 2012. [Google Scholar]
  61. Mahri, N.; Suandi, S.A.S.; Rosdi, B.A. Finger vein recognition algorithm using phase only correlation. In Proceedings of the 2010 International Workshop on Emerging Techniques and Challenges for Hand-Based Biometrics (ETCHB), Istanbul, Turkey, 22 August 2010. [Google Scholar]
  62. Wan, M.; Lai, Z.; Jin, Z. Feature extraction using two-dimensional local graph embedding based on maximum margin criterion. Appl. Math. Comput. 2011, 217, 9659–9668. [Google Scholar] [CrossRef]
  63. Zhang, T.; Li, X. Multimodal biometrics using geometry preserving projections. Pattern Recognit. 2008, 41, 805–813. [Google Scholar] [CrossRef]
  64. Zhang, L.; Li, H. Encoding local image patterns using Riesz transforms: With applications to palmprint and finger-knuckle-print recognition. Image Vis. Comput. 2012, 30, 1043–1051. [Google Scholar] [CrossRef]
  65. Hinton, G.E.; Osindero, S.; Teh, Y.-W. A fast learning algorithm for deep belief nets. Neural Comput. 2006, 18, 1527–1554. [Google Scholar] [CrossRef]
  66. Zhao, X.; Shi, X.; Zhang, S. Facial expression recognition via deep learning. IETE Tech. Rev. 2015, 32, 347–355. [Google Scholar] [CrossRef]
  67. Lyons, M.; Akamatsu, S.; Kamachi, M.; Gyoba, J. Coding facial expressions with gabor wavelets. In Proceedings of the Third IEEE International Conference on Automatic Face and Gesture Recognition, Nara, Japan, 14–16 April 1998. [Google Scholar]
Figure 1. The interactions by attraction and repulsion for the points between different classes.
Figure 1. The interactions by attraction and repulsion for the points between different classes.
Sensors 19 01643 g001
Figure 2. A sample of pre-cropped face images in the Sheffield Face database [51].
Figure 2. A sample of pre-cropped face images in the Sheffield Face database [51].
Sensors 19 01643 g002
Figure 3. (ak) The comparative recognition results changing the dimensionality of the transformation matrix for each given training number Tn on each data (Sheffield Face database).
Figure 3. (ak) The comparative recognition results changing the dimensionality of the transformation matrix for each given training number Tn on each data (Sheffield Face database).
Sensors 19 01643 g003aSensors 19 01643 g003b
Figure 4. (a) A subset of the original Yale database. (b) A subset of cropped images [53].
Figure 4. (a) A subset of the original Yale database. (b) A subset of cropped images [53].
Sensors 19 01643 g004
Figure 5. (ae). The comparative recognition results changing the dimensionality of the transformation matrix for each given training number Tn on each dataset (Yale Face database).
Figure 5. (ae). The comparative recognition results changing the dimensionality of the transformation matrix for each given training number Tn on each dataset (Yale Face database).
Sensors 19 01643 g005aSensors 19 01643 g005b
Figure 6. Example of six different subjects (each with 4 images) from the ORL database [56].
Figure 6. Example of six different subjects (each with 4 images) from the ORL database [56].
Sensors 19 01643 g006
Figure 7. (af) The comparative recognition results changing the dimensionality of the transforming matrix for each given training number Tn on each data (ORL database).
Figure 7. (af) The comparative recognition results changing the dimensionality of the transforming matrix for each given training number Tn on each data (ORL database).
Sensors 19 01643 g007
Figure 8. A subset of some images of one subject from the Head Pose database [59].
Figure 8. A subset of some images of one subject from the Head Pose database [59].
Sensors 19 01643 g008
Figure 9. (af) Comparative recognition results changing the dimensionality of the transformation matrix for each given training number Tn in each dataset (Head Pose database).
Figure 9. (af) Comparative recognition results changing the dimensionality of the transformation matrix for each given training number Tn in each dataset (Head Pose database).
Sensors 19 01643 g009
Figure 10. Example of captured images of one person in the Finger Vein database [61].
Figure 10. Example of captured images of one person in the Finger Vein database [61].
Sensors 19 01643 g010
Figure 11. (ag). Comparative recognition results changing the dimensionality of the transformation matrix for each given training number Tn (Finger Vein database).
Figure 11. (ag). Comparative recognition results changing the dimensionality of the transformation matrix for each given training number Tn (Finger Vein database).
Sensors 19 01643 g011aSensors 19 01643 g011b
Figure 12. A cropped sample of the finger knuckle print (FKP) database [63].
Figure 12. A cropped sample of the finger knuckle print (FKP) database [63].
Sensors 19 01643 g012
Figure 13. (ai). Comparative recognition results changing the dimensionality of the transformation matrix for each given training number Tn (Finger Knuckle database).
Figure 13. (ai). Comparative recognition results changing the dimensionality of the transformation matrix for each given training number Tn (Finger Knuckle database).
Sensors 19 01643 g013aSensors 19 01643 g013b
Figure 14. (af). Maximum recognition rate of SKLDNE versus Wk for different number of training samples on Sheffield, Yale, ORL, Head Pose, Finger Vein and Finger Knuckle databases.
Figure 14. (af). Maximum recognition rate of SKLDNE versus Wk for different number of training samples on Sheffield, Yale, ORL, Head Pose, Finger Vein and Finger Knuckle databases.
Sensors 19 01643 g014aSensors 19 01643 g014b
Table 1. Maximum recognition accuracies (in percentage terms) of supervised kernel locality-based discriminant neighborhood embedding (SKLDNE) and other methods for the different numbers of training and testing images on the Sheffield Face and corresponding dimensions (shown in parentheses).
Table 1. Maximum recognition accuracies (in percentage terms) of supervised kernel locality-based discriminant neighborhood embedding (SKLDNE) and other methods for the different numbers of training and testing images on the Sheffield Face and corresponding dimensions (shown in parentheses).
DatabaseSheffield Face
Tn12345678151617
PCA45.147.9449.684949.6451.1552.554.5485.59392
(18)(18)(18)(18)(30)(26)(30)(26)(18)(28)(10)
KPCA45.248.250.3150.3350.751.9254.1555.987.59392
(18)(22)(30)(38)(42)(58)(50)(69)(38)(30)(14)
UDP45.3348.148.4351.6650.7151.955.4157.27879592.5
(10)(22)(30)(30)(45)(66)(54)(50)(90)(86)(14)
LPP45.5548.252.8150.6656.0752.354.1655.99093.3395
(14)(26)(12)(34)(38)(54)(46)(62)(74)(42)(34)
LDA45.250.2948.4358.6656.0759.236061.3692.59397.5
(18)(26)(14)(34)(22)(6)(42)(50)(22)(30)(18)
DNE45.248.2350.3151.3351.7851.954.5856.3687.593.3392.5
(18)(34)(14)(74)(58)(42)(74)(78)(38)(30)(20)
LDNE45.2750.5856.8758.666573.0776.6676.819096.197.5
(22)(9)(14)(14)(14)(34)(18)(14)(22)(14)(18)
SKLDNE46.3852.9459.0662.6669.6478.4683.7580.4593.7598.33100
(10)(10)(14)(10)(14)(30)(10)(10)(34)(10)(10)
Table 2. Maximum recognition accuracies (in percentage terms) of SKLDNE and other methods for the different number of training and testing images on Yale Face database and corresponding dimensions (shown in parentheses).
Table 2. Maximum recognition accuracies (in percentage terms) of SKLDNE and other methods for the different number of training and testing images on Yale Face database and corresponding dimensions (shown in parentheses).
DatabaseYale Face
Tn16789
PCA51.6681.6688.8886.693
(29)(22)(26)(26)(10)
KPCA5083.39186.6693.3
(10)(30)(30)(90)(10)
UDP49.1681.6688.89092.9
(25)(50)(54)(28)(18)
LPP518591.196.6693.3
(22)(26)(30)(34)(18)
LDA5081.6691.193.393.3
(22)(18)(22)(98)(50)
DNE51.6683.3919093
(30)(30)(30)(66)(10)
LDNE6083.388.8893.33100
(19)(50)(57)(48)(42)
SKLDNE60.838595.5596.66100
(22)(38)(62)(46)(26)
Table 3. Maximum recognition accuracies (in percentage terms) of SKLDNE and other methods for the different number of training and testing images in the ORL Face database and corresponding dimensions (shown in parentheses).
Table 3. Maximum recognition accuracies (in percentage terms) of SKLDNE and other methods for the different number of training and testing images in the ORL Face database and corresponding dimensions (shown in parentheses).
DatabaseORL Face
Tn145678
PCA78.7585.4187.595.6295.8395.9
(30)(30)(26)(20)(10)(10)
KPCA81.56878996.296.6696.25
(46)(54)(66)(34)(34)(20)
UDP8086.6689.594.7596.696
(54)(90)(98)(38)(18)(14)
LPP80.6287.5909595.897.5
(58)(86)(94)(34)(30)(62)
LDA80.9387.919096.2595.8396
(54)(34)(38)(22)(34)(18)
DNE81.5687.088996.296.6696.25
(46)(54)(66)(34)(34)(10)
LDNE85929295.69596.5
(26)(62)(78)(30)(24)(54)
SKLDNE85.9393.339496.8797.5100
(38)(61)(50)(66)(22)(18)
Table 4. Maximum recognition accuracies (in percentage terms) of SKLDNE and other methods for the different number of training and testing images on Head Pose database and corresponding dimensions (shown in parentheses).
Table 4. Maximum recognition accuracies (in percentage terms) of SKLDNE and other methods for the different number of training and testing images on Head Pose database and corresponding dimensions (shown in parentheses).
DatabaseHead Pose
Tn307090110120130
PCA66.2163.6950.1679.7383.682.67
(30)(26)(26)(26)(74)(26)
KPCA66.3864.3159.3784.8686.2285.71
(30)(70)(30)(90)(22)(62)
UDP65.8364.3958.7585.5287.2185.71
(6)(22)(34)(98)(94)(78)
LPP68.0164.459.785.3988.3687.85
(30)(82)(26)(90)(62)(50)
LDA68.2165.146086.5787.0488.57
(6)(18)(22)(98)(98)(46)
DNE66.2864.558.584.6386.2285.71
(30)(70)(30)(90)(74)(62)
LDNE69.766.25996.99898.02
(20)(19)(18)(24)(18)(18)
SKLDNE70.767.0660.8398.9499.0199.28
(30)(18)(22)(22)(18)(14)
Table 5. Maximum recognition accuracies (in percentage terms) of SKLDNE and other methods for the different number of training and testing images on Finger Vein database and corresponding dimensions (shown in parentheses).
Table 5. Maximum recognition accuracies (in percentage terms) of SKLDNE and other methods for the different number of training and testing images on Finger Vein database and corresponding dimensions (shown in parentheses).
DatabaseFinger Vein
Tn1234567
PCA79.6689.3794.8596.396.419999.54
(30)(34)(34)(26)(26)(26)(27)
KPCA79.3388.7594.5796.3396.499.299.5
(30)(30)(30)(30)(26)(26)(27)
UDP8090.7595.8596.8979999.5
(30)(30)(30)(34)(30)(34)(30)
LPP79.6689.594.8596.8396.69999.56
(30)(22)(34)(26)(22)(34)(34)
LDA79.6691.7596.7197.1698.299.599.55
(30)(26)(30)(30)(30)(22)(29)
DNE8190.1295.429796.899.15100
(90)(90)(62)(86)(66)(54)(86)
LDNE80.7594.1796.8597.596.499.2599.2
(86)(38)(62)(76)(74)(66)(42)
SKLDNE81.2295.3897.7198.599.299.75100
(54)(26)(26)(34)(34)(26)(26)
Table 6. Maximum recognition accuracies (in percentage terms) of SKLDNE and other methods for the different number of training and testing images on Finger Knuckle database and corresponding dimensions (shown in parentheses).
Table 6. Maximum recognition accuracies (in percentage terms) of SKLDNE and other methods for the different number of training and testing images on Finger Knuckle database and corresponding dimensions (shown in parentheses).
DatabaseFinger Knuckle
Tn14567891011
PCA50.1867.568.2859.6675.292949397.5
(30)(78)(94)(30)(26)(82)(90)(38)(40)
KPCA52.961.8763.4259.6680.287.7589.669397.15
(50)(60)(39)(37)(30)(25)(26)(27)(20)
UDP56.1865.2567.1463.8882.29293.396.597
(62)(98)(90)(90)(98)(98)(90)(70)(35)
LPP55717267.582.292.594.29698
(62)(94)(94)(98)(94)(98)(98)(74)(86)
LDA5372.1272.1468.8384.892.794.339798
(62)(94)(98)(98)(90)(74)(90)(82)(26)
DNE53.8167.568.365.3380.492949397
(98)(88)(94)(94)(94)(82)(90)(38)(20)
LDNE53.976.757872.7084.692.7594.6693.397.2
(86)(87)(86)(94)(76)(34)(58)(46)(26)
SKLDNE56.3679.3781.4280.1690.69898.6699100
(22)(86)(98)(90)(66)(66)(26)(22)(18)
Table 7. The computational costs of the different classification methods using the Yale data base.
Table 7. The computational costs of the different classification methods using the Yale data base.
MethodSKLDNELDNEDNEKPCALDALPPUDPPCA
Time (s)0.350.40.370.0120.040.360.060.02
Table 8. Maximum recognition accuracies (in percentage terms) of SKLDNE and other methods on the JAFFE database.
Table 8. Maximum recognition accuracies (in percentage terms) of SKLDNE and other methods on the JAFFE database.
Method16 × 1632 × 3264 × 64
MLP64.7684.7686.19
DBNs + MLP88.5789.0590.95
SKLDNE100100100

Share and Cite

MDPI and ACS Style

Khalili Mobarakeh, A.; Cabrera Carrillo, J.A.; Castillo Aguilar, J.J. Robust Face Recognition Based on a New Supervised Kernel Subspace Learning Method. Sensors 2019, 19, 1643. https://doi.org/10.3390/s19071643

AMA Style

Khalili Mobarakeh A, Cabrera Carrillo JA, Castillo Aguilar JJ. Robust Face Recognition Based on a New Supervised Kernel Subspace Learning Method. Sensors. 2019; 19(7):1643. https://doi.org/10.3390/s19071643

Chicago/Turabian Style

Khalili Mobarakeh, Ali, Juan Antonio Cabrera Carrillo, and Juan Jesús Castillo Aguilar. 2019. "Robust Face Recognition Based on a New Supervised Kernel Subspace Learning Method" Sensors 19, no. 7: 1643. https://doi.org/10.3390/s19071643

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop