Next Article in Journal
Efficient Architectures for Full Hardware Scrypt-Based Block Hashing System
Previous Article in Journal
A Novel Deep Learning Model Compression Algorithm
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Facial Identity Verification Robust to Pose Variations and Low Image Resolution: Image Comparison Based on Anatomical Facial Landmarks

Division of Electronics and Display Engineering, Hoseo University, 20 Hoseo-ro 79, Babang-eub, Asan 31499, Chungcheongnam-do, Korea
Electronics 2022, 11(7), 1067; https://doi.org/10.3390/electronics11071067
Submission received: 4 February 2022 / Revised: 1 March 2022 / Accepted: 17 March 2022 / Published: 28 March 2022
(This article belongs to the Special Issue Face Recognition: Latest Trends and Future Perspectives)

Abstract

:
Face comparison/face mapping is one of the promising methods in face biometrics which needs relatively little effort compared with face identification. Various factors may be used to verify whether two faces are of the same person, among which facial landmarks are one of the most objective indicators due to the same anatomical definition for every face. This study identified major landmarks from 2D and 3D facial images of the same Korean individuals and calculated the distance between the reciprocal landmarks of two images to examine their acceptable range for identifying an individual to obtain standard values from diverse facial angles and image resolutions. Given that reference images obtained in the real-world could be from various angles and resolutions, this study created a 3D face model from multiple 2D images of different angles, and oriented the 3D model to the angle of the reference image to calculate the distance between reciprocal landmarks. In addition, we used the super-resolution method of artificial intelligence to address the inaccurate assessments that low-quality videos can yield. A portion of the process was automated for speed and convenience of face analysis. We conclude that the results of this study could provide a standard for future studies regarding face-to-face analysis to determine if different images are of the same person.

1. Introduction

The face is one of the most important physical features of an individual that externally expresses identity via its own distinctive characteristics. Therefore, the face and face images are critical data in the identity verification process. Currently, with increasing utilization of security and surveillance systems, acquired images are commonly used in public safety and image forgery detection [1,2,3,4]. Therefore, numerous studies have attempted to develop methods to improve the accuracy of identity verification using facial images.
The methods for identity verification using facial images can be categorized into two major approaches. The first is called automated facial recognition (AFR), which uses computer algorithms to compare a database of facial images with the target facial image and identify the most similar database images. Another method is to compare and analyze facial images of the person side-by-side. The second approach is called facial comparison/mapping; this method can be relatively easy for people to use as compared with AFR, which requires a large image database and considerable development cost [5].
In a forensic context, landmark-based face comparison methods can be useful. A method that uses likelihood ratios (LRs) to assess the grouping of facial images based on the morphometric indices was presented in [6]. The landmark values, such as averages and SDs for the reference image, were estimated, and it calculated the probability of these values from the probability distributions for the suspect as well as the population to acquire the likelihood ratio. In addition, using the facial ratio from inter-landmark distances, it performed intra- and inter-sample comparisons using the mean absolute value, Euclidean distance and cosine distance between ratios [7]. The statistics of the two frontal-faces were examined to determine the calculation that best identifies a detectable correlation difference between faces belonging to the same person. In addition, a series of area triplets was constructed for geometric invariance including the area ratio and angle, then used as feature vectors in face identification by the detected landmarks [8].
Some more specific ways of verifying identity are photoanthropometry, morphological analysis, and facial superimposition [9,10,11,12]. When using these methods, many standards can be used in the analysis, such as facial landmarks, the distance between landmarks, the ratio and angle of the landmarks, and distinctive facial structures [13,14,15]. Facial landmarks have the same anatomical definition for every face. Therefore, such landmarks could be used as an objective standard if one could place reciprocal landmarks on test and comparison faces.
The human face has about 80 to 90 nodal points. Facial verification systems measure important aspects, such as the distance between the eyes, the width of the nose, and the length of the chin [16,17]. In addition, the reciprocal landmarks in the frontal or profile face are used directly or defined newly by extending the points [7,18,19,20]. As already demonstrated by many studies, deep facial features for face verification perform much better than the methods that use facial landmarks [18,21,22]. However, the facial landmarks are not only visually easier to explain than deep facial features, but also objective indicators that define the appearance of a face. Therefore, the face landmarks can be useful when forensic cases need to be discussed in court [23,24,25]. The face mapping/comparison is a task of finding out if two people are the same person, similar to inferring lookalike. Applications such as Google’s ‘Art & Culture’ app and Microsoft’s ‘CelebLike’ are famous for their ability to show the faces of celebrities who resemble the user. Finding a similar face is related to the face verification that selects one ID that is most likely to match the input in that it infers the face image that is most perceptually similar to the input [21,26,27]. Humans can identify two people as different identities, even if their faces are very similar [27]. To estimate the similarity between two face images, many studies have utilized the distance between low-level feature descriptors, such as SIFT (Scale-Invariant Feature Transform), facial shape, facial features, or facial texture [8,28,29,30,31]. Recent studies have explored improving facial verification performance by combining these traditional features and deep-learning technologies [32,33,34].
However, there can exist differences between identifiable features based on the image resolution or angle of the face. To handle the pose problem, a promising method is to use 3D facial information because the restored 3D face model is available to render target facial poses. It is an ill-posed problem to generate 3D information from a single image [35], therefore, prior-model based methods have been proposed, which require 3D scan data of a large number of people to reproduce various face shapes of the input [36,37,38]. In addition, a 3D geometry can be directly estimated from images using the SfM (Structure from Motion)—methods for estimating camera movements from image sequences with various viewpoints [39,40]. In addition, if there are several photos of the object illuminated in different directions, the depth can be estimated more accurately by combining this information. These photometric stereo techniques have been traditionally applied for 3D reconstruction [41,42]. Recently, deep-learning-based super-resolution studies have brought great progress in image quality improvement. There are many methods for super resolution that are based on deep-learning technology, such as super resolution using deep convolutional network (SRCNN), coupled deep convolutional auto-encoder (CDCA), and very deep convolutional network-based super resolution (VDSR) [43,44,45].
Theoretically, the distance between reciprocal landmarks would be close to zero if the facial pose of two images obtained from an individual were aligned perfectly. However, it is impossible for the calculated distance to be zero due to the bias that occurs when a professional analyst identifies landmarks or due to inaccuracies that occur when images of different resolutions are compared. Nevertheless, if one could minimize such potential errors and accurately locate landmarks on both images, then the distance between reciprocal landmarks would approach zero if the images being compared were of the same individual. Therefore, it might be possible to use the distance between reciprocal landmarks of images of the same individual as a standard for identity verification.
In general, large-scale datasets, such as MS-Celeb-1M [46] with 10 million face images, are required for face identification, and complex algorithms and deep-learning architectures need to be designed to process them. Our study has the novelty in that it provides effective landmarks for face identification by analyzing traditional facial landmarks that represent the structure of a face, as a way to determine whether it is the same person by comparing two input images. This can provide visual clues for identification in that it is simpler and contains more interpretable morphometric features than the deep-learning features. In addition, the images of suspects acquired from security cameras contain variations in facial pose and expression, low-image resolution, and occlusion of the face by hair or accessories. These factors can degrade the comparison performance of experts such as investigators [47,48].
This study aimed to resolve issues caused by different poses and image resolutions. First, to solve the problems caused by pose variation between reference and comparison images, we created a 3D face model of a subject via multiple images acquired from multiple cameras with various angles, then rotated the model pose to match that of the comparison image. Second, the comparison image obtained from CCTV (Closed-Circuit Television) generally has low resolution, which can degrade the performance of the face analyst who must manually locate facial landmarks. Thus, image enhancement could improve the accuracy of the location of landmarks. In this paper, a super-resolution method via deep learning is applied to solve this problem. In addition, to improve speed, convenience, and accuracy of face analysis, we utilized a machine learning-based facial landmark detector, which is widely used in computer vision. Further, we provide an index for identity verification based on facial landmarks and an associated threshold from which to determine whether different images have the same identity.
The contributions of this study can be summarized as follows:
  • We present a landmark-based face mapping method which can represent the morphology of the face and easier provide visual and interpretable cues than deep features.
  • We provide the landmark indices and associated thresholds by which to determine whether input images have the same identity.
  • To cope with the images of low resolution and various poses, this study extracts more accurate facial landmarks from the input faces through restoring the 3D model to correct the poses and improving the image quality.

2. Materials and Methods

The overall flow of the identity verification method described in this paper, which uses the distance between reciprocal landmarks of a reference image and comparison image, is as follows.
First, to estimate a 3D geometry of the comparison face, we used photogrammetric range imaging technique from a sequence of facial images. The facial pose of the reconstructed 3D face model is rotated into the same pose as the comparison image obtained by security cameras such as CCTV. In addition, super-resolution technology is applied to the comparison image with low resolution to accurately extract facial landmarks. Next, the facial landmarks within each image are detected by an elastic model-based facial landmark detector; a professional facial analyst may adjust the detected landmarks for increased accuracy. Based on these facial landmarks, size normalization is performed to generate two images of same size and enable comparison of the locations of landmarks. The size of the facial image is adjusted by setting the interpupillary distance (IPD) to 100 pixels for a frontal presentation of the image. For other views, the distance between the nasion (midline depth of the nasal root) and gnathion (lowest median landmark on the lower border of the chin) are set to 100 pixels. Finally, we analyze the distribution of distances for the same individual and different individuals by calculating the distance between reciprocal landmarks of normalized facial images. The whole process is described in Figure 1.
To perform an identity verification experiment, we first constructed a Korean face dataset that consisted of over 2500 images (the K-Face DB [49]. https://github.com/k-face/k-face_2019, accessed on 21 December 2019). The dataset was collected from 50 Koreans (25 males and 25 females) who ranged in age from 20 to 50 years. By using a multiple-camera system with various angles and a light controller, over 50 images of each person were captured. Specifically, 40 images were collected as comparison images with 5 different facial angles, 2 different lighting conditions under 400 lux and 200 lux, and 4 different resolutions which include the ones higher than WQXGA (Wide Quad Extended Graphics Array) and WVGA (Wide Video Graphics Array) and lower than WVGA. The obtained image resolutions are 2592 × 1728, 864 × 576, 346 × 230, 173 × 115, respectively. Here, we assume the image with higher resolution than WQXGA and WVGA as high-resolution and mid-resolution image, respectively, and the rest as low-resolution image. Additionally, 12–14 images from each camera angle under uniform lighting were acquired to generate the 3D face models. The 3D face reconstruction process extracted 2D feature points from images, calculated their correspondences, extracted a mesh, and restored the facial textures based on the SfM technique [39]. We applied GPU (Graphics Processing Unit) acceleration such that the total execution time was less than one minute. Generally, there exists a problem of fixing the model’s scale and rotation coordinate system when restoring 3D data from photos. To address this issue, we post-processed all models to have physical dimensions and coordinate system based on measured distances and locations of some of the cameras of the multi-camera system. With these steps, we created a dataset that consisted of a 3D face model and 40 2D facial images per person, which we used to conduct the distance-calculation experiment for identity verification. Before collecting data, we informed the research participants about the general purpose of the research and allowed them to sign a consent form if they wished to participate. An example of a constructed dataset, which included obtained images and reconstructed 3D face models, is shown in Figure 2. The resolutions of images collected by real-world security cameras vary. If an image is low-definition, there are limitations in accurately determining facial landmarks. To address this issue, we utilized a VDSR network [45] which has simple and good performance by learning the mapping between low-resolution and high-resolution images based on residual learning. The overall structure is that it passes an input image through several convolutional layers and acquires an enhanced image as a result. To train the architecture, the CelebA and BERC databases were used [50,51]. Figure 3 shows an example of the result of super resolution based on deep learning.
Next, the facial landmarks are detected in the reference image and comparison image via the above steps. For the automatic landmark detection process, we utilized the wild-feature detector [52] for feature detection tasks. This assumes that the face shape is a tree structure, and uses a part-based model for face detection, pose estimation, and facial feature detection. In the current approach, the landmark detector extracts the facial region, facial pose angle, and facial landmarks concurrently, using 3D face structure and an elastic model. In particular, the facial landmarks considered in this paper are shown in Table 1 and Table 2. These landmarks were chosen with reference to previous studies of identity verification, and include 4 positions on eyebrows to utilize this part of the face in identity verification [2,7,20,53] as described in Figure 4.
To train the facial landmark detector, the Multi-PIE database was used [54]. The Multi-PIE database is a large-scale facial database that consists of different poses, illumination conditions, and facial expressions of 337 of a variety of races, with a male-to-female ratio of 7 to 3. Figure 5 is a sample from the Multi-PIE database.
After detecting the 22 facial landmarks and 13 facial landmarks for the frontal pose and side pose, respectively, professional facial analysts manually adjusted the detected facial landmarks. Before calculating the distance between reciprocal landmarks, image size normalization was performed. This procedure normalized the face sizes within a comparison pair of images with different face sizes. Different images of the same individual could differ in size because of the characteristics of the respective cameras, the distance between camera and subject, and so on. Here, we set the IPD to 100 pixels for frontal poses and the distance between the nasion and gnathion to 100 pixels for other poses. Figure 6 describes the examples of facial region size differences for two images of the same person and Figure 7 illustrates the example of size normalization, respectively.

3. Results

Using the materials and data described above, we experimentally assessed identity verification in 3 ways. First, we calculated the distance between each reciprocal landmarks. The nasion and gnathion were excluded when calculating the distances because they were used to normalize image size. Thus, 22 distances for frontal poses and 11 distances for side poses were calculated. Euclidean Distance (ED) was used as the distance measure, according to Equation (1).
D i s t = i = 1 n ( ( x i n x j n ) 2 + ( y i n y j n ) 2 ) n
The dissimilarity of each landmark position was subject to pixel error. IPD is generally 63.36 mm [55]. Thus, a 1-pixel error was equivalent to 0.63 mm of distance because we set the IPD to 100 pixels. Accordingly, we analyzed the pixel error distribution for the same individual and for two different individuals to obtain standard values for identity verification. Before the experiment, we compared the average distance between reciprocal landmarks of each image resolution. Figure 8 shows the result of such comparison among high/mid-resolution, low-resolution, and enhanced low-resolution based on the super-resolution method. As mentioned in Section 1, low-image resolution can degrade the comparison performance of investigators. Thus, low-resolution images cause a problem of making the EDs larger, even if two images are identical. As shown in Figure 8, the EDs of low-resolution images (red bar) are more than double the distance of EDs of high/mid-resolution images (blue bar). By applying the super-resolution to low-resolution image (green bar), we could reduce the EDs for the same identity. Based on this result, we constructed a comparison image dataset which consists of high/mid-resolution images and enhanced low-resolution images obtained via super resolution. By using the constructed comparison dataset, we measured the EDs between each reciprocal landmarks according to 5 facial pose angles. The results are shown in Figure 8 and Table 3. In Figure 8, the X-axis, Y-axis, blue bars, and red bars refer to the landmark indexes defined in Table 1, calculated ED, distances for different images of the same individual, and distance for images of different individuals, respectively.
As shown in Figure 9 and Table 3, the distance between reciprocal landmarks tended to decrease as the pose angle approached that of a frontal pose. In particular, the landmarks near the eyes for the frontal pose had small values due to the size normalization based on the center of the two eyes; this decreased the distance between landmarks for different individuals. Thus, it was difficult to distinguish individuals using the landmarks near the eyes for frontal poses. For example, the right exocanthion (index number 16) had an average pixel ED of 2.96 for the same identity and 3.05 for different identities. If these distances are converted into mm, the values are 1.86 mm and 1.92 mm, respectively. Thus, if the facial analyst even slightly mis-positioned the landmark, two images obtained from different persons might be diagnosed as the same identity. In the case of landmarks, except the eye area, the distance between landmarks for the same identity was about half as small as that for different persons, and the average EDs of different identity near the ear area was almost three times larger than that of the same identity. Among the landmarks in the ear area, we could confirm that both otobasion inferius (index number 19, 21) are the most discriminable based on the average distance between same and different individual in case of frontal facial pose. For side poses, the ED between landmarks near the eye and ear were significantly different for the same identity versus different identities, which validates the importance of landmarks near the eye and ear for identity verification.
Based on the mean and standard deviation of ED of the landmark, we measured the d value to select the reliability of the facial landmark [56]. The d measurement is statistically used to calculate the distance between two distributions based on both mean and standard deviation. The equation of d is as follows.
d = μ i m p o s t e r μ g e n u i n e σ g e n u i n e 2 + σ i m p o s t e r 2 2
In Equation (2), μ g e n u i n e and μ i m p o s t e r denote the mean of ED for same identity and different identity, respectively. In addition, σ -genuine and σ -imposter are the standard deviations of them. As the distance between the two distributions increases, the value of d also becomes greater. Based on this, it is possible to select the reliable landmarks that can distinguish the identity of the same and different individuals. As shown in Table 3, we can confirm that the landmarks in the eye and ear region have the reliability to verify the same identity compared to other landmarks.
In addition, we calculated the confidence interval of the population mean to be 95% confidence. In case of this experiment, the number of samples for the same identity was 400, so the confidence interval of the population of right otobasion inferius (index number 19) ranged from 5.261 to 5.938. From this result, we set the maximum value of the confidence interval as standard value for identity verification. Thus, we are confident that the probability of two images having the same identity is high if the ED from otobasion inferius is smaller than 5.938. The threshold for identity verification of all the landmarks are shown in the rightmost column in Table 3.
In the second experiment, we divided the face into 4 regions (eye, nose, lip, and ear) and calculated EDs using the landmarks in each region. This experiment enabled us to determine which facial regions were more effective for identity verification. The facial landmarks and index numbers in each region are shown in Table 4.
In this experiment, the landmarks used for size normalization were excluded, as in the previous experiment. From the result, the d in the eye region of the frontal face has a very low value as in the previous experiment, which means that it could not be utilized for same identity verification. Based on the d , the eye region, except the frontal pose and ear region, are the important facial parts to verify same identity which have more reliability than other regions. The results are shown in Table 5.
In the last experiment, we calculated the ED based on entire facial landmarks, as shown in Table 6. The d between two distributions of the overall pose angle were similar or higher than that of the previous experiment. In other words, as the use of overall landmarks, the probability of distinction between the same person and different person would increase, which means that it can be used more reliably in identity verification. All of the standard values for identity verification in each facial pose were 2.879, 3.002, 1.406, 2.307, 2.507, respectively. All proposed methods are described in Algorithm 1.
Algorithm 1 Proposed Method for Selecting Reliable Landmark Indices.
Input: Multi-view references images, comparison image
Output: ED and landmark indices
 1:
Repeat:
 2:
Apply SfM [41] on multi-view images, generate a 3D face mesh
 3:
Rotate the 3D mesh to a comparison facial pose and acquire rendered image
 4:
Conduct VDSR [47] on the comparison image, acquire high-definition image
 5:
Conduct wild-feature detector [54] on the rendered and comparison image, obtain landmark points
 6:
Select the landmark indices for frontal and profile face images following [2,20,21,55]
 7:
Normalize the images by IPD and distance between the nasion and gnathion, both 100 pixels
 8:
Calculate distances between the selected landmarks following Equation (1).
 9:
Calculate d measure following Equation (2)
 10:
Continue Until: Paired (genuine and imposter) images ended
 11:
Analyze the distribution for the same and different individuals
 12:
Select reliable landmark indices

4. Discussion and Conclusions

In this paper, by creating a 3D face model from multiple images taken from multiple angles and reorienting it to match the angle of a comparison image, we resolved the issues associated with comparisons of different face angles and poses. Inaccuracies caused by low-resolution images typical of security cameras frequently occurred. To resolve this problem, we applied the super-resolution method based on deep learning, which converted the input into a high definition image, and proved the effect of super resolution by comparing the average distance with that from a low-resolution image. In addition, we applied an automatic facial landmark detector to improve speed, convenience, and accuracy for face analysts who typically must manually indicate facial landmarks on each image.
Finally, we provided a standard ED value and more reliable facial landmarks for identity verification. By using the d measurement between two distributions, which are from same identity and different identity, we could confirm the reliable landmarks according to each facial pose. As mentioned in the experimental results, the landmarks near the eyes and ears had a higher distinguish ability than others. The most reliable landmarks were otobasion inferius and otobasin superius in overall facial pose, and exocanthion, endocanthion, inner eyebrow, outer eyebrow were also reliability ones in the case of side pose. In addition, we proved that eyes and ears are significant facial components for identity verification by analyzing the distributions from landmark grouping comparison experiment. In addition, the standard threshold for identity verification was determined with 95% confidence by statistical analysis based on mean and standard deviation.
In future work, we plan to study standard measures by applying the proposed methods from various face datasets and, further, identity verification based on image comparisons that are robust to different facial expressions and occlusion of facial areas by hair or accessories. We also aim to improve on the reliability of our identity verification method by analyzing the relationships between lines and angles that define landmarks.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The author declares no conflict of interest.

References

  1. Bashbaghi, S.; Granger, E.; Sabourin, R.; Parchami, M. Deep learning architectures for face recognition in video surveillance. arXiv 2019, arXiv:1802.09990. [Google Scholar]
  2. Lee, W.J.; Kim, D.M.; Lee, U.Y.; Cho, J.H.; Kim, M.S.; Hong, J.H.; Hwang, Y.I. A preliminary study of the reliability of anatomical facial landmarks used in facial comparison. J. Forensic Sci. 2018, 64, 519–527. [Google Scholar] [CrossRef] [PubMed]
  3. Johnston, R.A.; Bindemann, M. Introduction to forensic face matching. Appl. Cogn. Psychol. 2013, 27, 697–699. [Google Scholar] [CrossRef]
  4. Li, Y.; Meng, J.; Luo, Y.; Huang, X.; Qi, G.; Zhu, Z. Deep convolutional neural network for real and fake face discrimination. In Proceedings of the Chinese Intelligent Systems Conference, Shenzhen, China, 24–25 October 2020; pp. 590–598. [Google Scholar] [CrossRef]
  5. FISWG. Guidelines for fa7ial Comparison Methods. Available online: www.fiswg.org (accessed on 8 August 2018).
  6. Rajesh, V.; Navdha, B.; Arnav, B.; Kewal, K. Towards facial recognition using likelihood ratio approach to facial landmark indices from images. Forensic Sci. Int. Rep. 2022, 5, 100254. [Google Scholar]
  7. Kleinberg, K.F.; Siebert, J.P. A study of quantitative comparisons of photographs and video images based on landmark derived feature vectors. Forensic Sci. Int. 2012, 219, 248–258. [Google Scholar] [CrossRef] [PubMed]
  8. Juhong, A.; Pintavirooj, C. Face recognition based on facial landmark detection. In Proceedings of the 10th Biomedical Engineering International Conference, Hokkaido, Japan, 31 August–2 September 2017; pp. 1–4. [Google Scholar] [CrossRef]
  9. Houlton, T.M.R.; Steyn, M. Finding Makhubu: A morphological forensic facial comparison. Forensic Sci. Int. 2018, 285, 77–82. [Google Scholar] [CrossRef] [PubMed]
  10. Stavrianos, C.; Zouloumis, L.; Papadopoulos, C.; Emmanouil, J.; Petalotis, N.; Tsakmalis, P. Facial mapping: Review of current methods. Res. J. Med. Sci. 2012, 6, 13–20. [Google Scholar] [CrossRef]
  11. Moreton, R.; Morley, J. Investigation into the use of photoanthropometry in facial image comparison. Forensic Sci. Int. 2011, 212, 231–237. [Google Scholar] [CrossRef]
  12. Kleinberg, K.F.; Vanezis, P.; Burighton, A.M. Failure of anthropometry as a facial identification technique using high-quality photographs. J. Forensic Sci. 2007, 52, 779–783. [Google Scholar] [CrossRef]
  13. Atsuchi, M.; Tsuji, A.; Usumoto, Y.; Yoshino, M.; Ikeda, N. Assessment of some problematic factors in facial image identification using a 2D/3D superimposition technique. Leg. Med. 2013, 15, 244–248. [Google Scholar] [CrossRef]
  14. Porighter, G.; Doran, G. An anatomical and photographic technique for forensic facial identification. Forensic Sci. Int. 2000, 114, 97–105. [Google Scholar] [CrossRef]
  15. Catterick, T. Facial measurements as an aid to recognition. Forensic Sci. Int. 1992, 56, 23–27. [Google Scholar] [CrossRef]
  16. NEC Corporation FR White Paper. Available online: https://www.nec.com/en/global/solutions/safety/pdf/NEC-FR_white-paper.pdf (accessed on 20 August 2021).
  17. Thorat, S.B.; Nayak, S.K.; Dandale, J.P. Facial recognition technology: An analysis with scope in India. Int. J. Comput. Netw. Inf. Secur. 2010, 8, 325–330. [Google Scholar]
  18. Giuseppe, A.; Fabrizio, F.; Claudio, G.; Claudio, V. A Comparison of face verification with facial landmarks and deep features. In Proceedings of the 2018 the Tenth International Conference on Advances in Multimedia, Athens, Greece, 22–26 April 2018; pp. 1–6. [Google Scholar]
  19. Bakshi, S.; Kumari, S.; Raman, R.; Sa, P.K. Evaluation of periocular over face biometric: A case study. Procedia Eng. 2012, 38, 1628–1633. [Google Scholar] [CrossRef] [Green Version]
  20. Bottino, A.; Cumani, S. A fast and robust method for the identification of face landmarks in profile images. WSEAS Trans. Comput. 2008, 7, 1250–1259. [Google Scholar]
  21. Parkhi, O.M.; Vedaldi, A.; Zisserman, A. Deep face recognition. In Proceedings of the 2015 British Machine Vision Conference, Swansea, UK, 7–10 September 2015; pp. 41.1–41.12. [Google Scholar] [CrossRef] [Green Version]
  22. Taigman, Y.; Yang, M.; Ranzato, M.A.; Wolf, L. DeepFace: Closing the gap to human-level performance in face verification. In Proceedings of the 2014 IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 1701–1708. [Google Scholar] [CrossRef]
  23. Arbab-Zavar, B.; Wei, X.; Bustard, J.D.; Nixon, M.S.; Li, C.-T. On forensic use of biometrics. In Handbook of Digital Forensics of Multimedia Data and Devices Publishing House; Wiley-IEEE Press: Hoboken, NJ, USA, 2007; pp. 207–304. [Google Scholar]
  24. Khan, M.A.; Anand, S.J. Suspect identification using local facial attributed by fusing facial landmarks on the forensic sketch. In Proceedings of the 2020 International Conference on Contemporary Computing and Applications (IC3A), Lucknow, India, 5–7 February 2020; pp. 18–186. [Google Scholar]
  25. Vezzetti, E.; Marcolin, F.; Tornincasa, S.; Moos, S.; Violante, M.G.; Dagnes, N.; Monno, G.; Uva, A.E.; Fiorentino, M. Facial landmarks for forensic skull-based 3d face reconstruction: A literature review. In Proceedings of the 2016 Augmented Reality, Virtual Reality, and Computer Graphics, Otranto, Italy, 24–27 June 2016; pp. 172–180. [Google Scholar]
  26. Zhu, Z.; Luo, Y.; Chen, S.; Qi, G.; Mazur, N.; Zhong, C.; Li, Q. Camera style transformation with preserved self-similarity and domain-dissimilarity in unsupervised person re-identification. J. Vis. Commun. Image Represent. 2021, 80, 103303. [Google Scholar] [CrossRef]
  27. Sadovnik, A.; Gharbi, W.; Vu, T.; Gallagher, A. Finding your lookalike: Measuring face similarity rather than face identity. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, Salt Lake City, UT, USA, 18–22 June 2018; pp. 2345–2353. [Google Scholar] [CrossRef] [Green Version]
  28. Cao, Q.; Ying, Y.; Li, P. Similarity metric learning for face recognition. In Proceedings of the 2013 IEEE International Conference on Computer Vision, Sydney, Australia, 1–8 December 2013; pp. 2408–2415. [Google Scholar] [CrossRef]
  29. Tredoux, C. A direct measure of facial similarity and its relation to human similarity perceptions. J. Exp. Psychol. Appl. 2002, 8, 180–193. [Google Scholar] [CrossRef]
  30. Abudarham, N.; Yovel, G. Same critical features are used for identification of familiarized and unfamiliar faces. Vis. Res. 2019, 157, 105–111. [Google Scholar] [CrossRef]
  31. Pierrard, J.-S.; Vetter, T. Skin detail analysis for face recognition. In Proceedings of the 2007 IEEE Conference on Computer Vision and Pattern Recognition, Minneapolis, MN, USA, 17–22 June 2007; pp. 1–8. [Google Scholar] [CrossRef] [Green Version]
  32. Yang, J.; Bulat, A.; Tzimiropoulos, G. FAN-Face: A simple orthogonal improvement to deep face recognition. In Proceedings of the 2020 Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; pp. 12621–12628. [Google Scholar] [CrossRef]
  33. Macarulla Rodriguez, A.; Geradts, Z.; Worring, M. Likelihood ratios for deep neural networks in face comparison. J. Forensic Sci. 2020, 65, 1169–1183. [Google Scholar] [CrossRef]
  34. Songsri, K.; Zafeiriou, S. Complement face forensic detection and localization with facial landmarks. arXiv 2019, arXiv:1910.05455. [Google Scholar]
  35. Kemelmacher-Shlizerman, I.; Seitz, S.M. Face reconstruction in the wild. In Proceedings of the IEEE International Conference on Computer Vision, Barcelona, Spain, 6–13 November 2011; pp. 1746–1753. [Google Scholar]
  36. Blanz, V.; Vetter, T. A morphable model for the synthesis of 3D faces. In Proceedings of the SIGGRAPH ’99: 26th Annual Conference on Computer Graphics and Interactive Techniques, Los Angeles, CA, USA, 8–13 August 1999; pp. 187–194. [Google Scholar]
  37. Paysan, P.; Knothe, R.; Amberg, B.; Romdhani, S.; Vetter, T. A 3D face model for pose and illumination invariant face recognition. In Proceedings of the IEEE Conference on Advanced Video and Signal Based Surveillance, Genova, Italy, 2–4 September 2009; pp. 296–301. [Google Scholar]
  38. Vlasic, D.; Brand, M.; Pfister, H.; Popović, J. Face transfer with multilinear models. ACM Trans. Graph. 2005, 24, 426–433. [Google Scholar] [CrossRef]
  39. Snavely, N.; Seitz, S.M.; Szeliski, R. Photo tourism: Exploring photo collections in 3D. ACM Trans. Graph. 2006, 25, 835–848. [Google Scholar] [CrossRef]
  40. Debevec, P.E.; Taylor, C.J.; Malik, J. Modeling and rendering architecture from photographs: A hybrid geometry-and image-based approach. In Proceedings of the 1996 SIGGRAPH, New Orleans, LA, USA, 4–9 August 1996; pp. 11–20. [Google Scholar] [CrossRef] [Green Version]
  41. Kemelmacher-Shlizerman, I.; Basri, R. 3D Face Reconstruction from a Single Image Using a Single Reference Face Shape. IEEE Trans. Pattern Anal. Mach. Intell. 2011, 33, 394–405. [Google Scholar] [CrossRef] [PubMed]
  42. Roth, J.; Tong, Y.; Liu, X. Unconstrained 3D face reconstruction. In Proceedings of the Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 2606–2615. [Google Scholar] [CrossRef]
  43. Dong, C.; Loy, C.C.; He, K.; Tang, X. Image super-resolution using deep convolutional networks. IEEE T. Pattern Anal. 2016, 38, 295–307. [Google Scholar] [CrossRef] [Green Version]
  44. Sharma, M.; Chaudhury, S.; Lall, B. Deep learning based frameworks for image super-resolution and noise-resilient super-resolution. In Proceedings of the 2017 IEEE International Joint Conference on Neural Networks, Anchorage, AK, USA, 14–19 May 2017; pp. 744–751. [Google Scholar] [CrossRef]
  45. Kim, J.; Lee, J.K.; Lee, K.M. Accurate image super-resolution sing Very Deep Convolutional Networks. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 1646–1654. [Google Scholar] [CrossRef] [Green Version]
  46. Guo, Y.; Zhang, L.; Hu, Y.; He, X.; Gao, J. MS-Celeb-1M: A dataset and benchmark for large-scale face recognition. In Proceedings of the 14th European Conference on Computer Vision, Amsterdam, The Netherlands, 8–16 October 2016; pp. 87–102. [Google Scholar] [CrossRef] [Green Version]
  47. Ogawa, Y.; Wada, B.; Taniguchi, K.; Miyasaka, S.; Imaizumi, K. Photo anthropometric variations in Japanese facial features: Establishment of large-sample standard reference data for personal identification using a three-dimensional capture system. Forensic. Sci. Int. 2015, 257, e1–e9. [Google Scholar] [CrossRef]
  48. Yoshino, M.; Noguchi, K.; Atsuchi, M.; Kubota, S.; Imaizumi, K.; Thomas, C.D.; Clement, J.G. Individual identification of disguised faces by morphometrical matching. Forensic Sci. Int. 2002, 127, 97–103. [Google Scholar] [CrossRef]
  49. Choi, Y.; Park., H.; Nam., G.P.; Kim, H.; Choi, H.; Cho, J.; Kim, I.-J. K-FACE: A large-Scale KIST face database in consideration with unconstrained environments. arXiv 2021, arXiv:2103.02211. [Google Scholar]
  50. Liu, Z.; Luo, P.; Wang, X.; Tang, X. Deep learning face attributes in the wild. In Proceedings of the 2015 IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; pp. 3730–3738. [Google Scholar] [CrossRef] [Green Version]
  51. Choi, S.H.; Lee, Y.J.; Lee, S.J.; Park, K.R.; Kim, J. Age estimation using a hierarchical classifier based on global and local facial features. Pattern Recogn. 2011, 44, 1262–1281. [Google Scholar] [CrossRef]
  52. Zhu, X.; Ramanan, D. Face detection, pose estimation, and landmark localization in the wild. In Proceedings of the 2012 IEEE Conference on Computer Vision and Pattern Recognition, Providence, RI, USA, 16–21 June 2012; pp. 2879–2886. [Google Scholar] [CrossRef]
  53. Wu, J.; Tse, R.; Heike, C.L.; Shapiro, L.G. Learning to compute the plane of symmetry for human faces. In Proceedings of the 2nd ACM Conference on Bioinformatics, Computational Biology and Biomedicine, Chicago, IL, USA, 1–3 August 2011; pp. 471–474. [Google Scholar] [CrossRef] [Green Version]
  54. Gross, R.; Matthews, I.; Cohn, J.F.; Kanada, T.; Baker, S. Multi-PIE. In Proceedings of the 2008 8th IEEE International Conference on Automatic Face Gesture Recognition, Amsterdam, Netherlands, 17–19 September 2008; pp. 1–8. [Google Scholar] [CrossRef]
  55. Gordon, C.C.; Blackwell, C.L.; Bradtmiller, B.; Parham, J.L.; Barrientos, P.; Paquette, S.P.; Corner, B.D.; Carosn, J.M.; Venezia, J.C.; Rockwell, B.M.; et al. 2012 Anthropometric Survey of U.S. Army Personnel: Methods and Summary Statistics; Technical Report NATICK/15-007; U.S. Army Natick Soldier Research, Development and Engineering Center: Natick, MA, USA, 2014. [Google Scholar]
  56. Thomas, W. Elementary Signal Detection Theory; Oxford University Press: New York, NY, USA, 2001. [Google Scholar]
Figure 1. Overall framework of the proposed method.
Figure 1. Overall framework of the proposed method.
Electronics 11 01067 g001
Figure 2. Example of dataset construction: (a) obtained images; (b) 3D face model based on 3D image reconstruction.
Figure 2. Example of dataset construction: (a) obtained images; (b) 3D face model based on 3D image reconstruction.
Electronics 11 01067 g002
Figure 3. Example of the result of super resolution: (a) low-resolution input images; (b) result of super resolution.
Figure 3. Example of the result of super resolution: (a) low-resolution input images; (b) result of super resolution.
Electronics 11 01067 g003
Figure 4. The location of facial landmarks: (a) frontal pose; (b) side pose.
Figure 4. The location of facial landmarks: (a) frontal pose; (b) side pose.
Electronics 11 01067 g004
Figure 5. Example face images from the Multi-PIE database [54].
Figure 5. Example face images from the Multi-PIE database [54].
Electronics 11 01067 g005
Figure 6. Example of facial region size differences for two images of the same person obtained using different cameras.
Figure 6. Example of facial region size differences for two images of the same person obtained using different cameras.
Electronics 11 01067 g006
Figure 7. Example of size normalization: (a) frontal pose; (b) side pose.
Figure 7. Example of size normalization: (a) frontal pose; (b) side pose.
Electronics 11 01067 g007
Figure 8. The comparison of average ED of same identity between high/mid-resolution, low-resolution, and enhanced low-resolution based on super-resolution method: (a) Left 90°; (b) Left 45°; (c) Front; (d) Right 45°; (e) Right 90°.
Figure 8. The comparison of average ED of same identity between high/mid-resolution, low-resolution, and enhanced low-resolution based on super-resolution method: (a) Left 90°; (b) Left 45°; (c) Front; (d) Right 45°; (e) Right 90°.
Electronics 11 01067 g008
Figure 9. Average ED between each reciprocal landmark with different facial poses: (a) Left 90°; (b) Left 45°; (c) Front; (d) Right 45°; (e) Right 90°.
Figure 9. Average ED between each reciprocal landmark with different facial poses: (a) Left 90°; (b) Left 45°; (c) Front; (d) Right 45°; (e) Right 90°.
Electronics 11 01067 g009
Table 1. Frontal pose facial landmark index numbers and definitions [2,7,20,53].
Table 1. Frontal pose facial landmark index numbers and definitions [2,7,20,53].
LandmarkAbbreviationDefinition
IndexTitle
Midline
1GnathiongnThe lowest median landmark on the lower border of the chin
2Labiale inferiusliThe midpoint of the lower vermilion line
5Labiale superiuslsThe midpoint of the upper vermilion line
6SubnasalesnThe point at which the columella merges with the upper lip in the midsagittal plane
9PronasaleprnThe most protruded point of the nasal tip
10NasionnThe midline depth of the nasal root
Bilateral
3,4CheilionchThe point located at each labial commissure
7,8AlarealThe most lateral point of the alar contour
11,15EndocanthionenThe point at the inner commissure of the palpebral fissure
12,16ExocanthionexThe point at the outer commissure of the palpebral fissure
13,17Outer EyebrowoeThe point at the outer eyebrow
14,18Inner EyebrowieThe point at the inner eyebrow
19,21Otobasion inferiusobiThe lowest point of attachment of the external ear to the head
20,22Otobasion superiusobsThe highest point of attachment of the external ear to the head
Table 2. Side pose facial landmark index numbers and definition [2,7,20,53].
Table 2. Side pose facial landmark index numbers and definition [2,7,20,53].
LandmarkAbbreviationDefinition
IndexTitle
Midline
1GnathiongnThe lowest median landmark on the lower border of the chin
2Labiale inferiusliThe midpoint of the lower vermilion line
3CheilionchThe point located at each labial commissure
4Labiale superiuslsThe midpoint of the upper vermilion line
5SubnasalesnThe point at which the columella merges with the upper lip in the midsagittal plane
6PronasaleprnThe most protruded point of the nasal tip
7NasionnThe midline depth of the nasal root
8ExocanthionexThe point at the outer commissure of the palpebral fissure
9EndocanthionenThe point at the inner commissure of the palpebral fissure
10Inner EyebrowieThe point at the inner eyebrow
11Outer EyebrowoeThe point at the outer eyebrow
12Otobasion inferiusobiThe lowest point of attachment of the external ear to the head
13Otobasion superiusobsThe highest point of attachment of the external ear to the head
Table 3. Mean and Standard deviation of ED between each reciprocal landmark.
Table 3. Mean and Standard deviation of ED between each reciprocal landmark.
PoseLandmarkSame IdentityDifferent Identity d Standard Value for Identity Verification
IndexMeanStd.MeanStd.
Left 90°22.4881.5295.0992.7341.1792.638
32.4131.4855.2692.7011.3102.559
42.0961.4074.8112.6401.2832.234
52.3952.1344.8082.6151.0112.605
62.0741.4214.6462.6011.2272.213
811.0308.58232.07818.4041.46611.871
910.0627.81031.25517.8621.53710.828
109.2417.04431.50817.6841.6549.931
1111.9929.03734.77919.9641.47112.877
1213.43810.39335.32220.4611.34914.456
1312.74510.17534.77619.9211.39313.742
Left 45°22.1221.5693.7582.0490.8972.275
32.5651.8334.9782.5581.0842.745
42.1481.3963.8672.2550.9162.285
52.6791.6454.0032.1690.6882.840
62.9232.3514.3392.4270.5933.153
810.9208.98932.34520.3191.36411.801
910.0468.22531.04419.5731.39910.852
109.8287.76531.76919.7261.46410.589
1113.21510.59035.14321.6641.28614.253
1213.50110.51633.15320.8101.19214.531
1313.18310.61133.14220.8091.20814.223
Front17.1584.32513.9778.8710.9777.582
25.0102.92610.4076.2781.1025.297
35.7393.37510.8166.5190.9786.070
45.3163.0009.7465.7580.9655.610
54.6123.0409.2855.9850.9844.910
63.7592.3757.5344.5341.0433.991
74.2882.3607.7314.4390.9684.520
84.4752.2857.7574.2190.9674.699
95.0562.8847.6414.5650.6775.338
104.1842.7165.112.8720.3354.451
112.7721.8783.0481.7250.1532.956
122.7681.8743.0481.7250.1552.952
1310.9557.25311.9966.9080.14711.666
145.9392.9708.4974.4550.6766.230
152.9691.8403.0491.7410.0453.149
162.9611.8303.0491.7400.0493.141
1710.5866.52511.5936.5420.15411.225
185.3332.9588.6404.4470.8765.623
195.6003.45316.8099.2721.6025.938
206.6334.27915.3288.4431.2997.053
215.0903.15217.8439.8121.7505.399
225.6794.13514.7808.1621.4076.084
Right 45°22.2171.3563.7952.0840.8982.349
32.6551.9165.4142.8281.1422.843
42.1471.3234.1682.1651.1272.276
52.5001.4714.1092.2260.8532.644
63.1572.2134.6632.4450.6463.374
88.2457.02029.18117.3381.5838.933
98.7966.70429.50617.8851.5339.453
108.7685.92929.00417.3631.5609.349
1110.4526.19929.31216.6751.49911.059
128.1897.21629.47117.1681.6168.897
138.8917.13529.12816.5191.5909.590
Right 90°22.3131.4465.1512.7281.3002.455
32.4641.6125.6212.9501.3282.622
42.5311.8405.0712.7801.0772.712
52.6582.2054.9242.6250.9352.874
62.3791.7374.4032.4540.9522.549
89.5877.26634.80321.4691.57310.299
910.5637.85534.61921.6091.48011.333
1010.0107.38034.02621.6041.48810.734
1110.1986.68034.73320.6731.59710.853
129.7207.38335.08421.0221.61010.444
139.3636.95635.03320.3151.69110.045
Table 4. Facial landmarks and index numbers of each facial region.
Table 4. Facial landmarks and index numbers of each facial region.
Facial RegionFrontal PoseSide Pose
LandmarksLandmark IndexLandmarksLandmark Index
EyeEndocanthion,
Exocanthion, Outer
Eyebrow, Inner
Eyebrow
11,12,13,14,
15,16,17,18
Exocanthion,
Endocanthion, Inner
Eyebrow, Outer
Eyebrow
4
NoseSubnasale, Alare,
Pronasale, Nasion
6,7,8,9,10Subnasale, Pronasale,
Nasion
5,6,7
LipLabiale inferius,
Labiale superius,
Cheilion
2,3,4,5Labiale inferius,
Labiale superius,
Cheilion
2,3,4
EarOtobasion inferius,
Otobasion superius
19,20,21,22Otobasion inferius,
Otobasion superius
12,13
Table 5. Mean and Std. of EDs between reciprocal landmarks in each region.
Table 5. Mean and Std. of EDs between reciprocal landmarks in each region.
Facial RegionPose AngleSame IdentityDifferent Identity d Standard Value
MeanStd.MeanStd. for Identity Verification
EyeLeft 90°5.4313.91716.3169.0751.5575.815
Left 45°5.7324.22816.4459.9421.4026.146
Front2.4451.0582.9070.9960.4502.549
Right 45°4.6943.02314.7578.4351.5884.990
Right 90°5.1433.51217.38910.4801.5675.487
NoseLeft 90°1.0990.7972.2881.1151.2271.177
Left 45°1.3880.8552.0500.9210.7451.472
Front2.1170.7863.431.4631.1222.194
Right 45°1.4100.7722.1500.9400.8601.486
Right 90°1.2340.8742.2611.0811.0461.319
LipLeft 90°1.4210.8253.0331.3261.5091.492
Left 45°1.3900.8202.5921.0131.3041.470
Front2.7471.2405.2722.6541.2192.869
Right 45°1.4320.7712.7411.0801.3951.508
Right 90°1.4960.7983.1741.3771.4911.575
EarLeft 90°9.3387.15824.98213.9271.41310.040
Left 45°9.5027.36923.58414.4801.22610.224
Front3.1221.4678.5283.6291.9533.266
Right 45°6.1724.90720.88511.6171.6506.653
Right 90°6.8874.87024.95614.3301.6887.364
Table 6. Mean and Std. of EDs between reciprocal landmarks in each region.
Table 6. Mean and Std. of EDs between reciprocal landmarks in each region.
Pose AngleSame IdentityDifferent Identity d Standard Value for
MeanStd.MeanStd. Identity Verification
Left 90°2.6941.8917.6394.0101.5772.879
Left 45°2.8062.0027.4884.3631.3793.002
Front1.3640.4222.4480.7321.8141.406
Right 45°2.1741.3556.7383.5701.6902.307
Right 90°2.3591.5177.9664.4481.6872.507
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Hong, Y.-J. Facial Identity Verification Robust to Pose Variations and Low Image Resolution: Image Comparison Based on Anatomical Facial Landmarks. Electronics 2022, 11, 1067. https://doi.org/10.3390/electronics11071067

AMA Style

Hong Y-J. Facial Identity Verification Robust to Pose Variations and Low Image Resolution: Image Comparison Based on Anatomical Facial Landmarks. Electronics. 2022; 11(7):1067. https://doi.org/10.3390/electronics11071067

Chicago/Turabian Style

Hong, Yu-Jin. 2022. "Facial Identity Verification Robust to Pose Variations and Low Image Resolution: Image Comparison Based on Anatomical Facial Landmarks" Electronics 11, no. 7: 1067. https://doi.org/10.3390/electronics11071067

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop