Next Article in Journal
ISLS: An Illumination-Aware Sauce-Packet Leakage Segmentation Method
Previous Article in Journal
Using Wearable Digital Devices to Screen Children for Mental Health Conditions: Ethical Promises and Challenges
Previous Article in Special Issue
Predicting Facial Attractiveness from Colour Cues: A New Analytic Framework
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Multitask Learning Strategy with Pseudo-Labeling: Face Recognition, Facial Landmark Detection, and Head Pose Estimation

1
School of Electrical and Electronic Engineering, Yonsei University, 50 Yonsei-ro, Seodaemun-gu, Seoul 03722, Republic of Korea
2
School of Computer Science and Engineering, Kunsan National University, 558 Daehak-ro, Gunsan 54150, Jeollabuk-do, Republic of Korea
3
Taejae JeonMX Division, Samsung Electronics Co., Ltd., Suwon 16677, Gyeonggi-do, Republic of Korea
*
Author to whom correspondence should be addressed.
Sensors 2024, 24(10), 3212; https://doi.org/10.3390/s24103212
Submission received: 9 April 2024 / Revised: 8 May 2024 / Accepted: 14 May 2024 / Published: 18 May 2024
(This article belongs to the Special Issue Deep Learning Based Face Recognition and Feature Extraction)

Abstract

:
Most facial analysis methods perform well in standardized testing but not in real-world testing. The main reason is that training models cannot easily learn various human features and background noise, especially for facial landmark detection and head pose estimation tasks with limited and noisy training datasets. To alleviate the gap between standardized and real-world testing, we propose a pseudo-labeling technique using a face recognition dataset consisting of various people and background noise. The use of our pseudo-labeled training dataset can help to overcome the lack of diversity among the people in the dataset. Our integrated framework is constructed using complementary multitask learning methods to extract robust features for each task. Furthermore, introducing pseudo-labeling and multitask learning improves the face recognition performance by enabling the learning of pose-invariant features. Our method achieves state-of-the-art (SOTA) or near-SOTA performance on the AFLW2000-3D and BIWI datasets for facial landmark detection and head pose estimation, with competitive face verification performance on the IJB-C test dataset for face recognition. We demonstrate this through a novel testing methodology that categorizes cases as soft, medium, and hard based on the pose values of IJB-C. The proposed method achieves stable performance even when the dataset lacks diverse face identifications.

1. Introduction

Face analysis has improved considerably in various tasks, including detection [1,2], recognition [3,4,5], and estimation [6,7], with several models achieving remarkable accuracy and efficiency. This progress is attributed to the development of various network structures and the availability of a large number of face datasets for training. Although sufficient datasets are available for major tasks such as face recognition and face detection, substantial gaps exist in data availability for tasks requiring sensitive labeling, such as facial landmark detection and head pose estimation. This difficulty is especially noticeable in datasets with few face identifications (IDs) or a limited total amount of data, making it challenging to create robust and generalized models for the above-mentioned tasks.
Various methods have been explored to address the issue of the scarcity of face IDs within datasets. Typical single-task learning methods have been developed to overcome this challenge by finding commonalities between datasets or using information from other tasks. In tasks such as heterogeneous face recognition [8,9,10], style embedding features are learned to extract only common features from datasets in different domains with the same face class, or real faces are generated in different domains via generative models and then face recognition is performed to determine the identity. In studies of facial expression or stress recognition [11,12], small changes in facial landmarks have been used to specify the region to focus on, resulting in a more accurate estimate. Efforts towards domain adaptation have been made to address the challenge of data imbalance, which has been a significant problem regarding the effective functioning of facial landmark detection and head pose estimation in real-world scenarios [13,14,15,16,17,18]. Multitask learning has also been used to solve the problem of the lack of face IDs. A multitask cascaded convolutional neural network (MTCNN) [19] combines face detection and facial landmark detection to provide more accurate facial detection and alignment. To improve the facial recognition performance, some studies [20,21] have implemented multitask learning using facial attributes such as gender, hair color, and facial expressions.
However, these approaches assume label availability for multiple tasks within a dataset. Consequently, they inherently suffer from challenges as a fundamental solution to the lack of face IDs and overall volume in datasets. Alternative approaches exist, such as improving the data quantity through research in pseudo-labeling methods [22,23,24,25,26]. Inspired by multitask learning methods such as the all-in-one CNN [27] and SwinFace [21], as well as pseudo-labeling methods [22,23], such as those proposed by Pan et al., we hypothesize that the use of datasets that have numerous images with a variety of face IDs would be feasible for a task with a small number of face IDs. When combining datasets with different numbers of face IDs, we posit that the performance of precise pseudo-labeling can enable robust and generalized usability for any integration task in real-world scenarios. Figure 1 shows the overall pseudo-labeling and multitask learning framework. We present pseudo-labeling and multitask learning methods for face recognition, head pose estimation, and facial landmark detection to validate our assumptions.
To achieve precise pseudo-labeling, we prepared three pretrained networks: one pretrained multitask network for facial landmark detection and head pose estimation trained on the 300W-LP dataset and two pretrained networks for face recognition trained on the VGGFace2 [28] and MS1MV3 [29] datasets, respectively. First, we designed our pseudo-labeling framework by referring to 300W-LP [22], a previous labeling framework for facial landmarks and head pose estimation. In 300W-LP [22], the authors used iterative fitting to find the optimal labels because, when relabeling the 300W [30] dataset to 300W-LP, it is not possible to label manually. In contrast, in this study, instead of using iterative fitting, we designed a sampling and fitting algorithm for various views of the image and found the optimal label. In our pseudo-labeling framework, to reduce the bias of the existing pretrained networks and for more robust feature learning, the landmark pseudo-label was supplemented using an alignment method of eye landmarks based on eye blink detection. Furthermore, given that most facial analysis datasets contain images of celebrities, we used two pretrained networks for redundant validation through pseudo-facial identification. Finally, we produced a pseudo-labeled training dataset (PLTD) for multitask learning.
In the following sections, we describe how to perform multitask learning using the pseudo-labeled training dataset and propose two new loss functions for learning that complement each other without converging to a single task. By using our proposed pseudo-labeling and multitask learning frameworks, we show that the robustness and generalization of our model is increased as the number of face IDs and background noise in the dataset increases, and we can achieve similar or better performance than other state-of-the-art (SOTA) methods in each task. Our contributions can be summarized as follows.
  • We applied our pseudo-labeling framework to tasks lacking a face ID, such as facial landmark detection and head pose estimation. We also used an inherently large face ID face recognition dataset.
  • Within our pseudo-labeling framework, we prepared a pseudo-labeled training dataset by assigning entirely new class labels. This dataset contained 8.3 M images annotated with 93 K facial class labels, 68 landmarks, and Euler angles using three pretrained networks.
  • In our pseudo-labeling framework, we used an eye blink detection network to align the eye landmarks to make them more robust, which the pretrained network could not achieve. We then employed a selective fitting algorithm to generate high-quality pseudo-labels and validated them using pretrained networks to prevent duplicated face IDs.
  • In our multitask learning framework, we designed a loss that generated a synergy between tasks for pose-invariant face recognition and pose-guided facial landmark detection. In addition, we developed a new face recognition evaluation method for pose-invariant evaluation on the IJB–C [31] dataset.
  • Finally, we constructed a novel pseudo-labeling and multitask learning framework and demonstrated its SOTA or near-SOTA performance, thus illustrating its applicability to not only the three tasks but also the entire field of facial analysis.

2. Related Works

2.1. Single-Task Learning

2.1.1. Face Recognition

Investigations into face recognition technology using deep learning techniques have mainly focused on refining the datasets and loss functions. The high accuracy in face recognition tasks is largely attributed to the quality of the associated datasets, which have grown immensely in size. Datasets such as CASIA-WebFace [32], VGGFace2 [28], MS-Celeb-1M [33], Megaface [34], and Webface [35] have between thousands and up to 100,000 human classes, and datasets with 4 million human classes have been created recently.
As these large, high-quality datasets become available, the design of practical loss functions for training on these datasets has become crucial. In deep-learning-based methods, the traditional SoftMax loss is based on the Euclidean distance between the embedding features. However, newer methods, such as L-Softmax [36] and A-Softmax [37], focus on angular distances that allow more effective decision boundaries for unseen human classes. Further advances, such as CosFace [38], introduce additional distances in the angular space to distinguish human classes more effectively. ArcFace [4] improves this by adding a margin to the angular distance, thereby further expanding the decision boundary. However, as the datasets have grown, these advanced loss functions have become computationally intensive, particularly in fully connected layers. Partial FC [5] addresses this issue by remembering the center of each human class’s embedding features and employing efficient sampling through multi-GPU setups. It considers only the positives and negatives of each human class, thereby reducing the computational cost of training.

2.1.2. Head Pose Estimation

Head pose estimation is primarily used as a lightweight indicator to determine the user status in areas such as virtual reality/augmented reality (AR) for human–computer interaction, autonomous driving systems, and similar applications.
Algorithms such as POSIT [39] and Perspective-n-Points [40] were used to deduce rotation matrices from landmarks and convert them into pose values. The construction of datasets was challenging because it was necessary to label the Euler angles accurately. Further, a variety of parameters, such as the distance between the camera and subject, the focal length, and other camera parameters, had to be kept consistent, making the process of obtaining the data difficult.
However, owing to the creation of the CMU-MultiPIE [41], BIWI Kinect Head Pose Database [42], and CMU-Panoptic [43] datasets, which are labeled with Euler angles for various poses, deep-learning-based head pose estimation methods such as FSA-Net [44], TriNet [45], and RankPose [46] have shown substantial performance improvements. These datasets have facilitated progress in head pose estimation and facial landmark detection. This task also offers a supplementary tool for the learning of pose-invariant features in face recognition and facial landmark detection, and it employs projection algorithms such as Perspective-n-Points to facilitate mutual estimation.

2.1.3. Facial Landmark Detection

At present, facial landmark detection is being used in various applications, as it involves locating key characteristic points on the face and using them to create 3D face models or produce virtual facial movements by mapping them to the blend shapes of 3D avatars.
Among deep-learning-based methods [6,7,22,47,48,49,50,51], TCDCN [47] represents an early 2D facial landmark detection model using deep learning; it demonstrates superior performance compared with previous nonparametric methods, even with a simplistic CNN layer structure. In contrast, FAN [48] adopts the stacked hourglass network structure to enable 2D and 3D facial landmark detection. Further, the Wing loss [49] has been proposed as a new loss function that focuses on small errors in the loss function to achieve refined learning in landmark detection.

2.2. Multitask Learning

In deep learning, multitask learning is often used to improve the performance of a primary task. The MTCNN [19] face detector is a good example of multitask learning in face analysis. It emphasizes the accuracy of face identification in images by simultaneously learning to distinguish between faces and backgrounds and locating five key points to facilitate the accurate determination of the bounding box coordinates. The primary goal of multitask learning is to increase the efficiency of all constituent tasks; nevertheless, the parameters must be carefully assigned to match the network architecture and learning methodology to avoid one-sided biases and achieve a balanced performance improvement.
As an example of good parameter assignment tailored to specific tasks, the all-in-one CNN [27] describes a universal multitask learning model that integrates tasks corresponding to the main areas of face analysis, namely face recognition, face detection, head pose estimation, and facial landmark detection. This method specifies face recognition as the primary network branch from which the subnetwork branches are constructed and subsequently trained for additional tasks. SwinFace [21] introduces a multitask learning approach that achieves SOTA performance while proficiently managing approximately 42 tasks, including face recognition, facial expression, age estimation, and facial attribute estimation; the results demonstrate that this learning method generalizes the network performance for various facial regions of interest (ROI) required in different tasks.
As facial landmark detection and head pose estimation are inter-projectable, a dataset called 300W-LP [22] was created to combine these two tasks into a single dataset, using a face model such as the 3D Morphable Model (3DMM) [52] to enable sophisticated facial landmark detection and head pose estimation. FAN [48], 3DDFA [22,53,54], and SynergyNet [7] leverage 300W-LP [22] to present multitask learning methods that extend beyond facial landmarks and head pose estimation to include 3D face alignment. In particular, SynergyNet [7] focuses on predicting the complete 3D face geometry, including 3D face alignment from 3D face modeling and face orientation estimation using a proposed multitask, multimodal, multi-representation landmark refinement network. This study focuses on 3D landmarks and important face attributes by using their embedded information to guide the learning of the 3D face geometry. SynergyNet demonstrates robustness under various conditions and exhibits SOTA performance in facial landmark detection and head pose estimation.

3. Proposed Method

3.1. Pseudo-Labeling

In this section, we introduce our pseudo-labeling framework for the integration of three tasks: face recognition (FR), head pose estimation (HPE), and facial landmark detection (FLD).

3.1.1. Dataset Selection Strategy

Table 1 lists the commonly used datasets for each task. We focus on the ratios of the datasets, the number of images per ID, and the total amount of data. We select three datasets as the pseudo-labeled training dataset: two FR datasets, VGGFace2 [28] and MS1MV3 [29], and 300W-LP [22], which is the labeled dataset for HPE and FLD.
BIWI [42] and CMU-Panoptic [43] are mainly used as HPE datasets, and 300 W [30], AFLW [55], COFW [56], and WFLW [57] are mainly used as FLD datasets. However, these HPE and FLD datasets face challenges in pseudo-labeling each other because of an absolute lack of face IDs, insufficient data, or a large domain gap between the datasets. For example, HPE datasets have face images in large poses, whereas FLD datasets have no face image information in large poses, which likely indicates mislabeling. However, HPE and FLD are sufficiently related to allow for mutual perspective projection, and the 300W-LP [22] dataset was developed in another study to contain labels for HPE and FLD tasks. The 300W-LP dataset was created by applying the 3DDFA [22] fitting algorithm through the 3D Face Morphable Model [52] from 300 W; it includes 2D and 3D landmarks and the yaw, pitch, and roll of the pose labels in large pose variants. However, although this dataset was expanded from 3837 images to 61,225 images, it still has issues, such as a lack of face IDs, an inability to detect eye blinks, and the distortion of most images in three dimensions, leading to a lack of diversity in the background domain. To resolve these issues, we attempted to integrate the data from this dataset with the abundant face IDs from the FR datasets and train a multitask network using 300W-LP to attempt pseudo-labeling.
When selecting an FR dataset for pseudo-labeling, as shown in Figure 2, we primarily considered datasets with a balance among the number of images per ID and the total number of IDs. We chose VGGFace2 [28] as it allows sufficient background diversity. However, when training facial recognition networks using VGGFace2, we observed that the small number of face IDs caused a noticeable drop in verification performance in the challenge test dataset, IJB-C [31]. To compensate for this, we considered adding other FR datasets, such as MS1MV3 [29], MegaFace2 [34], and WebFace260M [35], which have several IDs. MegaFace2 was not considered owing to its small number of images per ID, and the training of WebFace260M was computationally expensive. Consequently, we used MS1MV3 for noise and alignment. Finally, we selected two FR datasets, VGGFace2 and MS1MV3. To assign new IDs to the integrated dataset, we prepared two individual networks on each dataset by using the ResNet [58] architecture.

3.1.2. Pseudo-Labeling Landmark and Pose

We devised a pseudo-labeling framework to create an integrated training dataset for multitask learning, as shown in Figure 3. Our pseudo-labeling framework consists of two main steps. Initially, the two FR datasets are subjected to pseudo-labeled landmarks and pose labels. We prepared a pretrained model by using the 300W-LP [22] dataset on the ResNet50 [58] model and concurrently employed Google’s MediaPipe [59] landmark detection framework to minimize bias in our network. One of the issues described earlier is the notable influence of the face detection bounding box area used during training. It affects the generalization of tasks dealing with insufficient face IDs to real-world applications. Therefore, an algorithm that uses multiple detection ROIs for a single image is essential in determining accurate landmark and pose values. Thus, we implemented an algorithm to extract quality pseudo-labels by generating multiple input images from a single image. By applying various mixtures of crops and padding from our data augmentation set, we enabled our pretrained network and MediaPipe [59] to extract landmark and pose values from diverse perspectives. Algorithm 1 describes the sampling and fitting algorithm intended for pseudo-labeling.
Algorithm 1 Sampling and Selective Fitting Algorithm for Pseudo-Labeling.
Input: number of total samples N, image for sampling I, set of augmentations A = { a 1 , . . . , a N } , pretrained network with 300W-LP [22] dataset n e t , Euclidean distance threshold t d .
Output: Pseudo 68 landmarks and pose values.
 1:
make a new list D
 2:
for  n = 1 to N do
 3:
    augmented image I a n ( I )
 4:
     L 68 , P 3 n e t ( I )
 5:
     I e y e l , I e y e r Crop ( I , L 68 [ 37 : 48 ] )
 6:
     σ l , σ r n e t e y e ( I e y e l , I e y e r )
 7:
     L 68 [ 37 : 48 ] AdjustUpperEyes ( L 68 [ 37 : 48 ] , σ l , σ r )
 8:
     L 468 m e d MediaPipe ( I )
 9:
     L 68 m e d IndexMatching ( L 468 m e d )
10:
    if  L 68 is valid landmark then
11:
        append L 68 , P 3 to D
12:
    end if
13:
    if  L 68 m e d is valid landmark then
14:
        append L 68 m e d , P 3 to D
15:
    end if
16:
end for
17:
while Remaining D > 3  do
18:
     L 68 m e a n MeanLandmark ( D )
19:
     i d x CalcMostFarLandmark ( D , L 68 m e a n )
20:
    delete i d x in D list
21:
end while
22:
if Remaining D 3  then
23:
     L 68 m e a n MeanLandmark ( D )
24:
     a s s i g n true
25:
    for  L 68 s a m p l e in D  do
26:
        if Distance ( L 68 s a m p l e , L 68 m e a n ) > t d  then
27:
            a s s i g n false
28:
        end if
29:
    end for
30:
    if  a s s i g n  then
31:
        assign L 68 m e a n , P 3 m e a n pseudo label of I
32:
    end if
33:
end if
In particular, in Algorithm 1, we create multiple image inputs by applying the augmentation set A to each image I in the FR datasets. These inputs are then processed by our pretrained network n e t and trained using 300W-LP [22] and MediaPipe to extract the landmark vector L and pose vector P , respectively. Representing the landmark vector and the pose vector in Algorithm 1, L k denotes the vector containing the number of k landmarks. However, 300W-LP, which we used as the training dataset, rarely includes cases with eye blinks. Consequently, our pretrained model struggles with closed, obstructed, or narrowly opened eyes because it has not been trained using such samples. In contrast, the FR datasets that we chose include a diverse range of cases, and labeling without adjusting for the landmarks corresponding to the eyes would make the resulting pseudo-labeled dataset unable to account for challenging cases such as closed eyes.
Therefore, to incorporate a different level of complexity for eye landmarks compared with 300W-LP [22], we used a CNN-based eye blink detection network, as proposed by Hu et al. [60]. We pretrained an eye blink detection network using the ResNet18 architecture [58] based on the labeling information of cropped eyes and the binary class of opened or closed eyes as refined by Song et al. [61], using the BioID [62], AR [63], and CAS-PEAL [64] datasets.
As shown in Figure 4, we applied the trained eye blink detection network, n e t e y e , in Algorithm 1. This network was applied to the eye regions identified by L 68 [ 37 : 48 ] , corresponding to the 37th to 48th indices of the landmark vector. It extracted σ l and σ r values ranging from 0 to 1, representing the probability of blinks in the left and right eyes, respectively. In the AdjustUpperEyes method in Algorithm 1, these σ l and σ r values were then multiplied by the difference values, h l 1 , h l 2 , h r 1 , and h r 2 , representing the distances between the upper and lower eyes of each face, to incorporate the weight of the eye part of the landmark. Subsequently, the validity of the extracted landmarks was assessed. Landmarks that were abnormally small compared with the width and height of the input image and those that showed the x-axis coordinates of the landmarks clustered in the center of the image were considered invalid. Only samples validated through this process were incorporated into the list of landmark vectors D .
Meanwhile, as MediaPipe [59] is not trained with 300W-LP [22], we did not apply AdjustUpperEyes. Unlike typical landmark detection, MediaPipe outputs 468 landmarks. Therefore, to find the indices of these 468 landmarks that are most similar to those of the 68 landmarks, we used the 300W-LP dataset for the pretraining network. We compared the results from our pretrained network with MediaPipe’s results to extract the 68 landmark indices most analogous to 300W-LP from the 468 landmarks provided by MediaPipe, as illustrated in Figure 5. Similarly, after confirming that the landmark vector extracted via MediaPipe was valid, we added it to the list of landmark vectors D .
For selective fitting, we calculated the average landmark vector from the list of landmark vectors D , excluded the landmark vector with the farthest Euclidean distance from the average landmark vector from the list D , and repeated this step until the length of the list D was three or fewer. These were accepted as pseudo-labels for the image I only when the mean Euclidean distance between L 68 m e a n and each of the remaining samples was lower than the Euclidean distance threshold, t d . Consequently, the final image would store three or fewer samples’ average landmark and average pose values. The distance threshold t d was established based on the average performance of the AFLW2000-3D [22] test, with a normalized mean square error of 0.05, in the 3D FLD tasks.

3.1.3. Pseudo-Labeling Face Identification

We have described how to label pseudo-landmarks and pseudo-pose labels. Next, we describe a framework for the distribution of face IDs to integrate the dataset in the following Algorithm 2. First, most FR datasets are composed of data on celebrities and other famous individuals. Consequently, some duplicate IDs exist between the FR datasets. To filter out duplicates between the VGGFace2 [28] and MS1MV3 [29] datasets, we calculated the similarity threshold values t M S and t V G G while pretraining the face recognition task with each dataset. During the pretraining process, we used the LFW [65] verification test, commonly employed to evaluate the performance of face recognition tasks, to measure the optimal similarity threshold that maximizes the verification performance. This threshold was then used as the similarity threshold for the detection of duplicates in Algorithm 2. Both angular and Euclidean distances can be used to measure similarity; however, the same distance metric must be employed in the loss function used during pretraining. Since we used the large margin cosine loss [38] function in our proposed method, we employed the angular distance for the cosine similarity in detecting data duplicates. For multitask learning, we filtered out duplicate IDs between VGGFace2 [28] and MS1MV3 [29] and added new IDs to 300W-LP [22] for the pseudo-labeled training dataset.
Algorithm 2 Identifying Duplicate IDs to Assign New IDs for the Integrated Dataset.
Input: Pretrained network using VGGFace2 [28] n e t V G G , pretrained network using MS1MV3 [29] n e t M S , VGGFace2 as D V G G , MS1MV3 as D M S , number of IDs in D V G G l V G G , number of IDs in D M S l M S , 300W-LP [22] as D 300 W , similarity threshold t V G G for n e t V G G , and similarity threshold t M S for n e t M S .
Output: Our pseudo-labeled training dataset D T with self-curated and unduplicated IDs.
 1:
E V G G Empty list of length l V G G + l M S
 2:
E M S Empty list of length l V G G + l M S
 3:
for each ID i V G G in D V G G  do
 4:
     S I set all images in i V G G
 5:
     E V G G [ i V G G ] Mean ( n e t V G G ( S I ) )
 6:
     E M S [ i V G G ] Mean ( n e t M S ( S I ) )
 7:
end for
 8:
for each ID i M S in D M S  do
 9:
     S I set all images in i M S
10:
     E V G G [ l V G G + i M S ] Mean ( n e t V G G ( S I ) )
11:
     E M S [ l V G G + i M S ] Mean ( n e t M S ( S I ) )
12:
end for
13:
Sim V G G ComputeSimilarityList ( E V G G , t V G G )
14:
Sim M S ComputeSimilarityList ( E M S , t M S )
15:
U FindDuplicates ( Sim V G G , Sim M S )
16:
D T MergeDatasets ( D V G G , D M S , U )
17:
D 300 W SemiIdentify ( D 300 W )
18:
U repeat step 1 to 15 using D T , D 300 W
19:
D T MergeDatasets ( D T , D 300 W , U )
As we integrated VGGFace2 [28] and MS1MV3 [29] as D T , we needed to assign new IDs to 300W-LP [22], which does not have any existing IDs. In particular, 300W-LP is a dataset that distorts each image of 300W [30], which holds approximately 3800 images, into 10 to 20 images, thus augmenting the dataset. Consequently, we assumed that each of the approximately 3800 images possessed a unique ID. We assigned arbitrary IDs to 61,225 images as D 300 W and checked for duplicate IDs at D T . We reviewed 92 internal duplicate IDs U and directed them to an integrated dataset of VGGFace2 and MS1MV3. We confirmed the absence of duplicate IDs and prepared the pseudo-labeled training dataset labeled with 8,271,075 images and 93,094 IDs, along with pseudo-landmark and pseudo-pose values. Figure 6 illustrates the constructed training dataset.

3.2. Multitask Learning

In this section, we discuss how to train the three tasks using the integrated training dataset discussed in the previous section. In particular, because the pseudo-labeled training dataset used pseudo-labeling rather than exact labeling, outliers inevitably existed, and we needed to find a learning technique to reduce these imperfections. The multitask learning of different facial analysis tasks also requires coordination on the optimal face regions required by each task. For example, for facial landmark detection, optimal learning results are achieved when the face region includes a significant portion of the head, chin, and neck. In contrast, face recognition can achieve good results when the background is minimized and major facial features such as the eyes, nose, and mouth are correctly aligned in a cropped face area. To address this performance trade-off based on the face area, we considered the required bounding box area of the face to determine the maximum area that could be used without degrading the performance of the task, and we actively used data augmentation methods to cut out arbitrary face areas between the maximum and minimum areas and use them for training.

3.2.1. Network Architecture

To show that our proposed learning method is feasible, we chose ResNet [58], a network structure commonly used in face-recognition-based networks, as a baseline and used the ResNet50 structure for our pretrained model to make the comparison as fair as possible. Furthermore, we used the modified IResNet50 [66], which is used in recent face-recognition-based methods such as ArcFace [4] and Partial FC [5], as the backbone of the multitask learning network. IResNet has no critical differences from ResNet, but we chose and modified the IResNet structure because it allowed us to appropriately modify the details of the normalization and activation layers required by each task in face recognition, facial landmark detection, and head pose estimation. While tuning the network architecture for each task, we also considered maximizing the shared parameters to minimize the inference time loss. To do this, we referenced several face analysis tasks [7,44,67,68] and finally split the shared and independent parameters at a 3:1 ratio for each task in the baseline network. The modified IResNet50 consists of four layers, where the first three layers are shared layers and the last layer is divided into sublayers customized for each task, as shown in Figure 7.

3.2.2. Multitask Loss Functions

To ensure balanced learning between tasks and avoid convergence to a specific task, weights must be assigned to each task, as given in Equation (1):
L t o t a l = L F R + λ 1 L F L D + λ 2 L H P E ,
where L F R , L F L D , and L H P E denote the loss functions of face recognition, facial landmark detection, and head pose estimation, respectively. We set the base criterion at 1 for L F R and applied the learning weights λ 1 and λ 2 to L F L D and L H P E , respectively. These weights were adjusted based on the sum of the values represented by the final output of each task. We fine-tuned each task’s contribution to the overall multitask framework by choosing λ 1 to be 100 and λ 2 to be 10 to provide an appropriate learning ratio based on the final convergence of each task’s loss function. We created our multitask loss function, described in Equation (1), by computing the total loss value L t o t a l by summing all weighted losses from each task.
The loss function for the head pose estimation task is given in Equation (2):
L H P E = 1 N P i = 1 N P | P P R E D i P G T i | ,
where N P is the number of elements (=3), representing yaw, pitch, and roll, and P is the pose vector including these three Euler angle values. It is calculated by computing the mean absolute error between the predicted pose vector P P R E D and the ground truth pose vector P G T . The head pose tends to be less influenced by the face region than the other two tasks and achieves convergence in the shortest time during training. Therefore, we designed the losses for the remaining two tasks to complement each other by assigning learning weights through the pose values to the other two tasks.
We prepared two major loss functions for the loss function of facial landmark detection L F L D , as presented in Equation (3):
L F L D = L w i n g + λ P L M S E ,
where one is the Wing loss L w i n g [49] and the other is the mean square error (MSE) loss multiplied by λ P . λ P is a regularization term that controls the convergence between the Wing loss and the MSE loss. The method for the determination of λ P and its effects on the convergence of the losses are described in Equation (6) through (9). The Wing loss L w i n g is given by Equation (4):
L w i n g = w ln ( 1 + | x | / ϵ ) if | x | < w | x | C otherwise ,
where w denotes the non-negative value to set the range of the nonlinear part, ϵ limits the curvature region, and C = w w ln ( 1 + w / ϵ ) is the smoothness between the linear and nonlinear parts. L denotes the landmark vector and x denotes the L1 norm of the Euclidean distance between the predicted landmark vector L P R E D and the ground truth landmark vector L G T .
Equation (3) also contains the loss of the mean square error, commonly used for regression tasks, as given in Equation (5):
L M S E = 1 N L i = 1 N L ( L P R E D i L G T i ) 2 ,
where N L represents the total number of landmarks, L P R E D i is the i-th landmark coordinate of the predicted landmark vector, and L G T i is the i-th landmark coordinate of the ground truth landmark vector. Our MSE loss, combined with λ P in Equation (3), focuses more on large errors to emphasize the importance of accurate landmark prediction in the presence of pose variations, making it a counterpart to the Wing loss [49] in our facial landmark detection task.
To compute λ P , we need to derive the projected landmark vector L P from the mean 3D landmark L 3 D used to calculate the samples in 300W-LP [22] by using the ground truth pose values of the pitch, yaw, and roll. The mean 3D landmark vector, denoted as L 3 D , comprises 68 landmark points, each represented by three-dimensional Euclidean space coordinate vectors x , y , and z . This vector is derived from the average landmark vector of samples in the frontal view of 300W-LP with pitch, yaw, and roll values close to zero. This mean 3D landmark is then multiplied by the projection matrix to locate the projected landmark vector given by
L 3 D = x y z T ,
Following the derivation of the mean 3D landmark, the rotated point L 3 D was calculated. This point is a product of the rotation matrix R , taken from Rodrigues’ formula [69], and the original 3D landmark, L 3 D , resulting in new coordinates x , y , and z . The interaction between these components is summarized in Equation (7):
L 3 D = R · L 3 D = x y z T ,
where the projected 2D landmark vector L P = ( x P , y P ) can be obtained using perspective projection to the rotated point L 3 D in Equation (8):
L P = x P y P = f x 0 c x 0 f y c y · L 3 D ,
where f x , f y are the focal lengths; c x , c y are the principal points; and x P , y P are 2D projected landmark coordinate vectors. These values can be determined based on the input shapes or assigned arbitrarily. λ P is calculated using Equation (9):
λ P = mean ( L P | | L P | | L P R E D | | L P R E D | | ) ,
where λ P is the mean of the Euclidean distance between the norms of the predicted landmark vector L P R E D , projected landmark vector L P , and visualization representation shown in Figure 8. The primary intent behind employing λ P is not necessarily to enhance the accuracy but to facilitate the initial learning of the landmarks. Given the possibility that the pseudo-labeling of the ground truth is erroneous, λ P also serves as a counterpart to address potential inaccuracies, acting as a stabilizing component in our multitask loss function.
With regard to the loss function for face recognition, we acknowledge the effectiveness of representing the embedding features in the angular domain through the angular margin, which has been proven to enhance the performance significantly. Consequently, we adopted the marginal cosine loss function method, conceptualized in approaches such as CosFace [38], SphereFace [37], and ArcFace [4], due to the novel representation in the angular domain.
Building on the foundational work on the large margin cosine loss [38], we focus on the angular region specific to each ID. To achieve this, we apply the pose ratio as a margin, allowing a greater emphasis on the angular region of each ID. As shown in Figure 9, the allocation of feature regions according to the pose values of each ID tends to drift away from the center of the feature region that the ID holds as the pose values increase. To address this concern and assign feature regions unaffected by the pose, we set margins to allow more challenging pose values closer to the center of their respective ID. L F R represents the large margin cosine loss originally proposed in [38]. However, in this study, we adapted it to incorporate the additional variance associated with pose values in the margin, as given in Equation (10).
L F R = L LPMC ( θ , y ) = 1 N F i l o g e s · c o s ( θ y i i m p ) e s · c o s ( θ y i i m p ) + j y i e s · c o s θ j i ,
subject to
c o s ( θ j i ) = W j T , x i ,
where N F is the number of training samples, and x i is the i-th normalized feature vector with a corresponding ground truth class of y i . W j is the normalized weight vector of the j-th class, and θ j is the angle between W j and x j , as conceptualized in CosFace [38]. A novel margin, m p = m 1 + m 2 γ P , is proposed, where γ P is the average of the three poses—yaw, pitch, and roll—summed and then divided by 180 to ensure a value ranging between 0 and 1.
When using our proposed large pose margin cosine loss, we assign more margin values as a weighted indication that the class is correct as the head pose value of the input face image increases. This enables the network to achieve tighter class clustering despite the diverse variations in head pose. Incorporating pose-invariant feature regions provides a means to effectively address and accommodate more difficult pose values, creating an improved and reliable face recognition model.

4. Experimental Results

In this section, we compare the results against those of various SOTA methods for each task and provide further understanding through an ablation study. We prepared three different network backbones, namely ResNet50 for our pretrained network, IResNet50 for our single-task learning (STL), and IResNet50-Multi for our multitask learning (MTL) model. As described in Section 3.2, the first three layers of each ResNet structure are shared layers, with the remaining layer allocated to each respective task. Our pretrained model, ResNet50 [58], used images of size 128 × 128, whereas our STL and MTL models using IResNet50 [66] were trained with images of size 112 × 112. These networks used λ 1 = 100 and λ 2 = 10 as parameters for the integrated loss L t o t a l in Equation 1, with C = 10 and ϵ = 2 as hyperparameters for the Wing loss [49] in Equation (4). Finally, for L F R in Equation (10), the hyperparameters s = 64 , m 1 = 0.3 , and m 2 = 0.7 were used, which are generally used in face recognition tasks. We started learning with a learning rate of 0.01, multiplying by 0.9 at every epoch, and trained on 8.3 M data per epoch for 20 epochs. The training was conducted on a computer with NVIDIA’s RTX A6000. Each epoch took approximately 3 h when training in FP16 and 6 h in FP32, resulting in a total training time of 60−120 h.

4.1. Face Recognition

4.1.1. Test Dataset

We prepared four datasets for the face recognition task, namely LFW [65], CFP-FP [70], AgeDB [71], and IJB-C [31], which are commonly used for performance comparisons, as shown in Table 2. LFW, CFP-FP, and AgeDB contain 13 K, 7 K, and 16.5 K images, respectively, and evaluations were performed on 6 K, 7 K, and 6 K verification pairs, respectively. These datasets were used to monitor the learning process of face recognition during multitask training. IJB-C has 148.8 K images exhibiting large pose and age variations. Although multiple testing methods, including verification and identification, are available for this dataset, only 1:1 verification was performed in this study. With 15M pairs in this test dataset’s protocol, it is suitable for evaluating the true acceptance rate (TAR) against the false acceptance rate (FAR), and many face recognition methods conduct TAR@FAR evaluations ranging from FAR = 1 × 10 1 to FAR = 1 × 10 6 . This study conducted evaluations across the entire protocol for IJB-C 1:1 verification. Furthermore, we focused on our pose-invariant learning method using our loss function and categorized IJB-C cases based on the pose values into soft, medium, and hard, as measured internally.

4.1.2. Experimental Results

Table 3 presents the accuracies obtained in the LFW [65], CFP-FP [70], and AgeDB [71] tests. Our pretrained models achieved 99.33%, 92.38%, and 94.20%, respectively, using VGGFace2 [28], and 99.85%, 98.70%, and 98.11%, respectively, using the MS1MV3 [29] dataset. Furthermore, our STL model, trained using the pseudo-labeled training dataset (PLTD) that we developed through pseudo-labeling and combining MS1MV3, VGGFace2, and the 300W-LP dataset, achieved 99.72%, 97.73%, and 96.9%, respectively, in the face recognition task. Our MTL model, incorporating the large pose margin cosine loss, achieved 99.70%, 97.26%, and 96.53%, respectively. Compared with the SOTA models, while most comparative models use ResNet100 [58] or IResNet100 [66] as their backbones, our model deploys ResNet50 [58] and IResNet50 [66]. Despite such differences, our best MTL model exhibited competitive performance in IJB-C [31], achieving 94.17% at TAR@FAR = 1 × 10 4 .

4.1.3. Discussion

In Section 3.1, we reassigned face IDs during dataset integration through pseudo-labeling. Duplication may have been misjudged as an error in some instances; these are referred to as outliers in our context. However, the verification performance summarized in Table 3 is comparable to, if not synonymous with, that of other SOTA algorithms. It is inferred that the test performance might have partially diminished because we did not crop the faces tightly or align them during the training stage, as done by other face recognition methods while using the dataset. Additionally, despite training with ResNet100, similar to SOTA configurations, our performance metrics were identical to those achieved with ResNet50.
These results indicate that filtering and learning from them is feasible even if the pseudo-labeled dataset incorporates some outliers. It also exhibits the potential to effortlessly integrate various facial analysis tasks beyond the proposed methods, highlighting the versatility and adaptability of our approach despite the anomalies and discrepancies within the data. In addition, we compared the inference time with that of similar works under identical conditions and conducted a frame-per-second (fps) comparison for the ResNet50 and IResNet50 models used in the comparison works. Unlike the recent trend of using arrays of multiple GPUs to train face recognition methods, our fps comparison was conducted using only a single GPU. Despite performing multitask computations, our proposed method performed well in terms of fps.
Next, in Table 4, we present the experimental results regarding the large pose margin cosine loss designed for our MTL and introduce a new IJB-C [31] test method for the evaluation of pose invariance. The IJB-C 1:1 verification method stores the average of all embedding features of the templates corresponding to each ID within the dataset. Then, it conducts a TAR@FAR performance evaluation by comparing the templates over 15 million pairs. Here, while we store the average of each template’s embedding feature in the gallery similarly, for the probe, we detect the yaw, the z-axis rotation angle of each face in each template, through our MTL model. Based on the yaw values, we name the images within 0 to 30 as soft cases, 30 to 60 as medium cases, and 60 to 90 as hard cases. We save the average of the images corresponding to each case in each template as an embedding feature to use as the probe. Ultimately, among 15,658,489 pairs in the IJB-C 1:1 verification method, 15,070,066 pairs have a soft case, 11,410,991 pairs have a medium case, and 5,926,383 pairs have a hard case, with the possibility of duplicates owing to the presence of all three cases within one template.
In Table 4, for a fair comparison, only our single-task and multitask learning models are used for the evaluation. Both applied the same learning rate and hyperparameters and concluded the training at the same epoch for comparison. If we compare IJB-C [31] TAR@FAR = 1 × 10 5 and 1 × 10 4 , the STL model performs better in all cases; however, our model performs better in medium and hard cases. Accordingly, we observed that the large pose margin cosine loss that we applied could represent more accurate embedding features for such large pose cases. However, compared to the soft cases, many aspects still need improvement. We anticipate that we will attain performance comparable to that of other SOTA methods by repeatedly refining our pseudo-labeled training dataset throughout our entire framework.

4.2. Head Pose Estimation

4.2.1. Test Dataset

For head pose estimation, we used two test datasets. One was AFLW2000-3D [22], which was relabeled from AFLW [55] using the 300W-LP method [22]. While the original and re-annotated versions exist, we only used the original version for testing because the re-annotated version has all faces tightly cropped. This dataset comprises 2000 images, including Euler angles and 68 facial landmarks. The other dataset was BIWI [42], where the evaluation was performed on 15,678 images. In addition, because the Euler angle variation of 300W-LP used for pseudo-labeling was between 99 and 99 , only test samples within the Euler angle were used for AFLW2000-3D and BIWI.

4.2.2. Experimental Results

The evaluation method used for the test employed the MAE of the Euler angle, as in other methods, to conduct the evaluation. Table 5 presents the comparative results for AFLW2000-3D [22]. Our multitask learning model achieved SOTA performance with an AFLW2000-3D MAE score of 3.32. Furthermore, as observed from Table 6, it achieved state-of-the-art results with an MAE score of 3.54.

4.2.3. Discussion

We achieved SOTA performance on both AFLW2000-3D [22] and BIWI [42], the representative test datasets for head pose estimation. This demonstrates that our pseudo-labeled training dataset allowed our multitask learning framework to achieve positive performance improvements in each task. By adding a large amount of face IDs and diverse background noise to address the lack of diversity in existing datasets, as identified in Section 1, we achieved higher accuracy for unseen images in the head pose estimation task.

4.3. Facial Landmark Detection

4.3.1. Test Dataset

We evaluated our model on the AFLW2000-3D [22] dataset, which was also used for head pose estimation. We measured the performance of our model using the normalized mean square error (NMSE):
NMSE = 1 N L i = 1 N L ( L P R E D i L G T i ) 2 d ,
where N L denotes the number of landmarks in landmark vector L , L P R E D i is the i-th landmark coordinate of the predicted landmark vector, L G T i is the i-th landmark coordinate of the ground truth landmark vector, and d is the size of the facial bounding box for the normalized mean square error.

4.3.2. Experimental Results

Table 7 shows a comparison of the performance of various methods and the proposed model in terms of the NMSE. Evaluations were carried out in three different pose ranges based on the yaw rotation angle of the face: from 0 to 30 , from 30 to 60 , and from 60 to 90 . The mean values were calculated by averaging the results within each pose range. Our MTL model demonstrated NMSE scores of 2.62, 3.45, and 4.51 in the respective pose variations. Although our model did not achieve SOTA performance in all variations, it showed superior performance within the lower range of pose variations from 0 to 30 , exceeding the results of SynergyNet [7].

4.3.3. Discussion

Based on the ground truth, our model achieved its best performance in the range of yaw values from 0 to 30 . Unfortunately, it did not perform as well in other angle ranges. This result is probably due to the characteristics of the pseudo-labeled training dataset that we created. As discussed previously, pseudo-labeling allows outliers to be included in the dataset, and these outliers are likely to be included as the pose variation of the face images used for pseudo-labeling increases. The pretrained model that we used for pseudo-labeling in Section 3.1 was trained using 300W-LP [22], which has a higher degree of distortion in images with a large pose variation and a nonuniform definition of invisible landmark points, and we suspect that these features were reflected in the pseudo-labeling. However, despite these challenges, our MTL model achieved near-SOTA performance, as shown in Table 7. This notable result highlights the robustness and adaptability of our approach in handling multiple tasks and varying conditions effectively, especially considering that the model achieves this while sharing the network parameters with other tasks for multitask learning.

4.4. Ablation Study

In this section, we analyze through additional experiments whether multitask learning affects the performance improvement, the influence of the hyperparameters λ 1 and λ 2 of the loss function designed in Section 3.2, and the influence of the regularization term λ P .

4.4.1. Influence of Multitask Learning

To demonstrate the effectiveness and stability of our proposed multitask learning framework through pseudo-labeling, we conducted experiments, as shown in Table 8. Table 8 shows that multitask learning configurations can outperform single-task learning models. All models were trained with a pseudo-labeled training dataset (PLTD). While the STL models did not perform optimally due to mislabeled outliers in the training dataset, the MTL models were trained with an optimal network parameter distribution for each task.

4.4.2. Influence of Regularization Term λ P

The λ P described in Equation (3) is a regularization term that we adopted because the proportion of large pose variation data in the large datasets was relatively small. If we did not adopt λ P , we would have been required to learn by excluding the difficult data in the large datasets, by the principle of convergence to the mean in deep learning, which affects the convergence of head pose estimation, as shown in Table 9. A λ P value of 0 indicates that no regularization strength is used, a λ P value of 1 indicates fixed strength for the facial landmark detection loss function, and a λ P value of α indicates regularization strength that is dynamically computed during training iterations.

4.4.3. Influence of Hyperparameters λ 1 and λ 2

In multitask learning, adjusting the learning weights for the loss function designed for each task is essential for the optimization process. In this paper, we take the loss function for face recognition as the baseline in the unified loss function designed in Equation (1) and set the hyperparameter coefficients of λ 1 and λ 2 to the other tasks, facial landmark detection and head pose estimation, respectively. Table 10 shows the results of the experiments conducted to find the optimal hyperparameters. In detail, the loss convergence ratio is the ratio of L F R , λ 1 L F L D , and λ 2 L F L D in Equation (1) based on the loss value of each task when the learning converges to the maximum. The results show that the convergence ratio should be distributed to a certain extent to converge to more optimal multitask learning. Learning by increasing the hyperparameter coefficients for the remaining two tasks is more effective in our proposed multitask learning framework than learning by expanding the hyperparameter coefficients for face recognition, which constitutes most of the pseudo-labeled training dataset.

4.5. Visualization

In this section, we compare our model with SynergyNet [7], a SOTA method that implements multitask learning for facial landmark detection and head pose estimation to produce visualization results. The results of AFLW2000-3D [22] are visualized in Figure 10. Both models could identify the landmarks and Euler angles with a high degree of accuracy; however, unsuccessful detection occurred in some cases with a large pose variation.
Figure 11 shows the performance of our model and SynergyNet [7] when applied to different ROIs of the face in images from IJB-C [31]. It demonstrates that our trained model can accurately detect landmarks in different face ROIs by learning diverse face ID information and background noises in the training dataset through pseudo-labeling. In contrast, SynergyNet [7] requires the precise detection of the face area through a detector owing to its insufficient training in various situations. The results in Figure 11 suggest that our model can be used with various face detectors. Our training framework provides more reliable performance for face recognition, head pose estimation, and facial landmark detection, making it more suitable for practical applications.

5. Conclusions

In this study, we addressed the problem of tasks with insufficient facial analysis performing well on test datasets but performing poorly when applied in the real world. To solve this problem, we used a dataset of face recognition tasks with a large number of face IDs and performed data enlargement through pseudo-labeling to create a multitask training dataset. We constructed a multitask learning method and achieved SOTA or near-SOTA performance. By using multitask learning, we configured a pose-guided loss function to achieve pose-invariant performance in face recognition and accurate learning in facial landmark detection, even when using a pseudo-labeled dataset containing outliers. We believe that our framework can be applied to various facial analyses.

Author Contributions

Conceptualization, Y.L.; Methodology, Y.L. and S.J.; Software, Y.L.; Validation, Y.L., H.B.B. and T.J.; Formal analysis, S.J. and H.B.B.; Writing—original draft, Y.L.; Writing—review & editing, S.J., H.B.B. and T.J.; Visualization, S.J. and H.B.B.; Supervision, S.L. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Research Foundation of Korea (NRF) grant funded by the Korea government (MSIT)(No. RS-2024-00423362).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

No new data were created or analyzed in this study. Data sharing is not applicable to this article.

Acknowledgments

We would like to express our gratitude to the Korea government, Ministry of Science and ICT (MSIT) and Yonsei University Office of Research Affairs.

Conflicts of Interest

Author Taejae Jeon was employed by the company MX Division, Samsung Electronics Co., Ltd. The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

References

  1. Deng, J.; Guo, J.; Zhou, Y.; Yu, J.; Kotsia, I.; Zafeiriou, S. Retinaface: Single-stage dense face localisation in the wild. arXiv 2019, arXiv:1905.00641. [Google Scholar]
  2. Li, J.; Liu, L.; Li, J.; Feng, J.; Yan, S.; Sim, T. Toward a Comprehensive Face Detector in the Wild. IEEE Trans. Circuits Syst. Video Technol. 2019, 29, 104–114. [Google Scholar] [CrossRef]
  3. Kim, T.K.; Kittler, J. Design and Fusion of Pose-Invariant Face-Identification Experts. IEEE Trans. Circuits Syst. Video Technol. 2006, 16, 1096–1106. [Google Scholar] [CrossRef]
  4. Deng, J.; Guo, J.; Xue, N.; Zafeiriou, S. Arcface: Additive angular margin loss for deep face recognition. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 4690–4699. [Google Scholar]
  5. An, X.; Deng, J.; Guo, J.; Feng, Z.; Zhu, X.; Yang, J.; Liu, T. Killing two birds with one stone: Efficient and robust training of face recognition cnns by partial fc. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 4042–4051. [Google Scholar]
  6. Lee, H.J.; Kim, S.T.; Lee, H.; Ro, Y.M. Lightweight and Effective Facial Landmark Detection using Adversarial Learning with Face Geometric Map Generative Network. IEEE Trans. Circuits Syst. Video Technol. 2020, 30, 771–780. [Google Scholar] [CrossRef]
  7. Wu, C.Y.; Xu, Q.; Neumann, U. Synergy between 3dmm and 3d landmarks for accurate 3d facial geometry. In Proceedings of the 2021 International Conference on 3D Vision (3DV), London, UK, 1–3 December 2021; pp. 453–463. [Google Scholar]
  8. Bae, H.B.; Jeon, T.; Lee, Y.; Jang, S.; Lee, S. Non-visual to visual translation for cross-domain face recognition. IEEE Access 2020, 8, 50452–50464. [Google Scholar] [CrossRef]
  9. Cho, M.; Kim, T.; Kim, I.J.; Lee, K.; Lee, S. Relational deep feature learning for heterogeneous face recognition. IEEE Trans. Inf. Forensics Secur. 2020, 16, 376–388. [Google Scholar] [CrossRef]
  10. Hu, W.; Hu, H. Orthogonal Modality Disentanglement and Representation Alignment Network for NIR-VIS Face Recognition. IEEE Trans. Circuits Syst. Video Technol. 2022, 32, 3630–3643. [Google Scholar] [CrossRef]
  11. Wu, W.; Yin, Y.; Wang, Y.; Wang, X.; Xu, D. Facial expression recognition for different pose faces based on special landmark detection. In Proceedings of the 2018 24th International Conference on Pattern Recognition (ICPR), Beijing, China, 20–24 August 2018; pp. 1524–1529. [Google Scholar]
  12. Jeon, T.; Bae, H.; Lee, Y.; Jang, S.; Lee, S. Stress recognition using face images and facial landmarks. In Proceedings of the 2020 International Conference on Electronics, Information, and Communication (ICEIC), Barcelona, Spain, 19–22 January 2020; pp. 1–3. [Google Scholar]
  13. Kuhnke, F.; Ostermann, J. Deep head pose estimation using synthetic images and partial adversarial domain adaption for continuous label spaces. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27–28 October 2019; pp. 10164–10173. [Google Scholar]
  14. Valle, R.; Buenaposada, J.M.; Valdés, A.; Baumela, L. Face alignment using a 3D deeply-initialized ensemble of regression trees. Comput. Vis. Image Underst. 2019, 189, 102846. [Google Scholar] [CrossRef]
  15. Jin, H.; Liao, S.; Shao, L. Pixel-in-pixel net: Towards efficient facial landmark detection in the wild. Int. J. Comput. Vis. 2021, 129, 3174–3194. [Google Scholar] [CrossRef]
  16. Bafti, S.M.; Chatzidimitriadis, S.; Sirlantzis, K. Cross-domain multitask model for head detection and facial attribute estimation. IEEE Access 2022, 10, 54703–54712. [Google Scholar] [CrossRef]
  17. Wan, M.; Zhu, S.; Luan, L.; Prateek, G.; Huang, X.; Schwartz-Mette, R.; Hayes, M.; Zimmerman, E.; Ostadabbas, S. Infanface: Bridging the infant–adult domain gap in facial landmark estimation in the wild. In Proceedings of the 2022 26th International Conference on Pattern Recognition (ICPR), Montreal, QC, Canada, 21–25 August 2022; pp. 4486–4492. [Google Scholar]
  18. Kuhnke, F.; Ostermann, J. Domain Adaptation for Head Pose Estimation Using Relative Pose Consistency. IEEE Trans. Biom. Behav. Identity Sci. 2023, 5, 348–359. [Google Scholar] [CrossRef]
  19. Zhang, K.; Zhang, Z.; Li, Z.; Qiao, Y. Joint face detection and alignment using multitask cascaded convolutional networks. IEEE Signal Process. Lett. 2016, 23, 1499–1503. [Google Scholar] [CrossRef]
  20. Wang, Z.; He, K.; Fu, Y.; Feng, R.; Jiang, Y.G.; Xue, X. Multi-task deep neural network for joint face recognition and facial attribute prediction. In Proceedings of the 2017 ACM on International Conference on Multimedia Retrieval, Bucharest, Romania, 6–9 June 2017; pp. 365–374. [Google Scholar]
  21. Qin, L.; Wang, M.; Deng, C.; Wang, K.; Chen, X.; Hu, J.; Deng, W. SwinFace: A Multi-task Transformer for Face Recognition, Expression Recognition, Age Estimation and Attribute Estimation. IEEE Trans. Circuits Syst. Video Technol. 2023, 34, 2223–2234. [Google Scholar] [CrossRef]
  22. Zhu, X.; Lei, Z.; Liu, X.; Shi, H.; Li, S.Z. Face alignment across large poses: A 3d solution. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 146–155. [Google Scholar]
  23. Pan, Z.; Wang, Y.; Zhang, S. Joint face detection and Facial Landmark Localization using graph match and pseudo label. Signal Process. Image Commun. 2022, 102, 116587. [Google Scholar] [CrossRef]
  24. Lee, D.H. Pseudo-label: The simple and efficient semi-supervised learning method for deep neural networks. In Proceedings of the Workshop on Challenges in Representation Learning, Atlanta, GA, USA, 16–21 June 2013; Volume 3, p. 896. [Google Scholar]
  25. Yu, X.; Ouyang, B.; Principe, J.C.; Farrington, S.; Reed, J.; Li, Y. Weakly supervised learning of point-level annotation for coral image segmentation. In Proceedings of the Oceans 2019 MTS/IEEE, Seattle, WA, USA, 27–31 October 2019; pp. 1–7. [Google Scholar]
  26. Caron, M.; Bojanowski, P.; Joulin, A.; Douze, M. Deep clustering for unsupervised learning of visual features. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 132–149. [Google Scholar]
  27. Ranjan, R.; Sankaranarayanan, S.; Castillo, C.D.; Chellappa, R. An all-in-one convolutional neural network for face analysis. In Proceedings of the 2017 12th IEEE International Conference on Automatic Face & Gesture Recognition (FG 2017), Washington, DC, USA, 30 May–3 June 2017; pp. 17–24. [Google Scholar]
  28. Cao, Q.; Shen, L.; Xie, W.; Parkhi, O.M.; Zisserman, A. Vggface2: A dataset for recognising faces across pose and age. In Proceedings of the 2018 13th IEEE International Conference on Automatic Face & GESTURE recognition (FG 2018), Xi’an, China, 15–19 May 2018; pp. 67–74. [Google Scholar]
  29. Deng, J.; Guo, J.; Zhang, D.; Deng, Y.; Lu, X.; Shi, S. Lightweight face recognition challenge. In Proceedings of the IEEE/CVF International Conference on Computer Vision Workshops, Seoul, Republic of Korea, 27–28 October 2019. [Google Scholar]
  30. Sagonas, C.; Tzimiropoulos, G.; Zafeiriou, S.; Pantic, M. 300 faces in-the-wild challenge: The first facial landmark localization challenge. In Proceedings of the IEEE International Conference on Computer Vision Workshops, Sydney, Australia, 2–8 December 2013; pp. 397–403. [Google Scholar]
  31. Maze, B.; Adams, J.; Duncan, J.A.; Kalka, N.; Miller, T.; Otto, C.; Jain, A.K.; Niggel, W.T.; Anderson, J.; Cheney, J.; et al. Iarpa janus benchmark-c: Face dataset and protocol. In Proceedings of the 2018 International Conference on Biometrics (ICB), Gold Coast, Australia, 20–23 February 2018; pp. 158–165. [Google Scholar]
  32. Yi, D.; Lei, Z.; Liao, S.; Li, S.Z. Learning face representation from scratch. arXiv 2014, arXiv:1411.7923. [Google Scholar]
  33. Guo, Y.; Zhang, L.; Hu, Y.; He, X.; Gao, J. Ms-celeb-1m: A dataset and benchmark for large-scale face recognition. In Proceedings of the Computer Vision–ECCV 2016: 14th European Conference, Amsterdam, The Netherlands, 11–14 October 2016; Proceedings, Part III 14. Springer: Cham, Switzerland, 2016; pp. 87–102. [Google Scholar]
  34. Nech, A.; Kemelmacher-Shlizerman, I. Level playing field for million scale face recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 7044–7053. [Google Scholar]
  35. Zhu, Z.; Huang, G.; Deng, J.; Ye, Y.; Huang, J.; Chen, X.; Zhu, J.; Yang, T.; Lu, J.; Du, D.; et al. Webface260m: A benchmark unveiling the power of million-scale deep face recognition. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 10492–10502. [Google Scholar]
  36. Liu, W.; Wen, Y.; Yu, Z.; Yang, M. Large-margin softmax loss for convolutional neural networks. arXiv 2016, arXiv:1612.02295. [Google Scholar]
  37. Liu, W.; Wen, Y.; Yu, Z.; Li, M.; Raj, B.; Song, L. Sphereface: Deep hypersphere embedding for face recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 212–220. [Google Scholar]
  38. Wang, H.; Wang, Y.; Zhou, Z.; Ji, X.; Gong, D.; Zhou, J.; Li, Z.; Liu, W. Cosface: Large margin cosine loss for deep face recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 5265–5274. [Google Scholar]
  39. Martins, P.; Batista, J. Accurate single view model-based head pose estimation. In Proceedings of the 2008 8th IEEE International Conference on Automatic Face & Gesture Recognition, Amsterdam, The Netherlands, 17–19 September 2008; pp. 1–6. [Google Scholar]
  40. Rocca, F.; Mancas, M.; Gosselin, B. Head pose estimation by perspective-n-point solution based on 2d markerless face tracking. In Proceedings of the Intelligent Technologies for Interactive Entertainment: 6th International Conference, INTETAIN 2014, Chicago, IL, USA, 9–11 July 2014; Proceedings 6. Springer: Cham, Switzerland, 2014; pp. 67–76. [Google Scholar]
  41. Gross, R.; Matthews, I.; Cohn, J.; Kanade, T.; Baker, S. Multi-pie. Image Vis. Comput. 2010, 28, 807–813. [Google Scholar] [CrossRef] [PubMed]
  42. Fanelli, G.; Weise, T.; Gall, J.; Van Gool, L. Real time head pose estimation from consumer depth cameras. In Proceedings of the Joint Pattern Recognition Symposium, Frankfurt, Germany, 31 August–2 September 2011; Springer: Cham, Switzerland, 2011; pp. 101–110. [Google Scholar]
  43. Joo, H.; Liu, H.; Tan, L.; Gui, L.; Nabbe, B.; Matthews, I.; Kanade, T.; Nobuhara, S.; Sheikh, Y. Panoptic studio: A massively multiview system for social motion capture. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; pp. 3334–3342. [Google Scholar]
  44. Yang, T.Y.; Chen, Y.T.; Lin, Y.Y.; Chuang, Y.Y. Fsa-net: Learning fine-grained structure aggregation for head pose estimation from a single image. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 16–17 June 2019; pp. 1087–1096. [Google Scholar]
  45. Cao, Z.; Chu, Z.; Liu, D.; Chen, Y. A vector-based representation to enhance head pose estimation. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Virtual, 5–9 January 2021; pp. 1188–1197. [Google Scholar]
  46. Dai, D.; Wangkit Wong, Z.C. RankPose: Learning Generalised Feature with Rank Supervision for Head Pose Estimation; Ping An Technology: Shenzhen, China, 2020. [Google Scholar]
  47. Zhang, Z.; Luo, P.; Loy, C.C.; Tang, X. Facial landmark detection by deep multi-task learning. In Proceedings of the Computer Vision–ECCV 2014: 13th European Conference, Zurich, Switzerland, 6–12 September 2014; Proceedings, Part VI 13;. Springer: Cham, Switzerland, 2014; pp. 94–108. [Google Scholar]
  48. Bulat, A.; Tzimiropoulos, G. How far are we from solving the 2d & 3d face alignment problem? (and a dataset of 230,000 3d facial landmarks). In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 1021–1030. [Google Scholar]
  49. Feng, Z.H.; Kittler, J.; Awais, M.; Huber, P.; Wu, X.J. Wing loss for robust facial landmark localisation with convolutional neural networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 2235–2245. [Google Scholar]
  50. Zou, X.; Xiao, P.; Wang, J.; Yan, L.; Zhong, S.; Wu, Y. Towards Unconstrained Facial Landmark Detection Robust to Diverse Cropping Manners. IEEE Trans. Circuits Syst. Video Technol. 2021, 31, 2070–2075. [Google Scholar] [CrossRef]
  51. Liao, X.; Wang, Y.; Wang, T.; Hu, J.; Wu, X. FAMM: Facial Muscle Motions for Detecting Compressed Deepfake Videos Over Social Networks. IEEE Trans. Circuits Syst. Video Technol. 2023, 33, 7236–7251. [Google Scholar] [CrossRef]
  52. Blanz, V.; Vetter, T. Face recognition based on fitting a 3D morphable model. IEEE Trans. Pattern Anal. Mach. Intell. 2003, 25, 1063–1074. [Google Scholar] [CrossRef]
  53. Guo, J.; Zhu, X.; Yang, Y.; Yang, F.; Lei, Z.; Li, S.Z. Towards fast, accurate and stable 3d dense face alignment. In Proceedings of the European Conference on Computer Vision, Glasgow, UK, 23–28 August 2020; Springer: Cham, Switzerland, 2020; pp. 152–168. [Google Scholar]
  54. Zhu, X.; Liu, X.; Lei, Z.; Li, S.Z. Face alignment in full pose range: A 3d total solution. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 41, 78–92. [Google Scholar] [CrossRef] [PubMed]
  55. Koestinger, M.; Wohlhart, P.; Roth, P.M.; Bischof, H. Annotated facial landmarks in the wild: A large-scale, real-world database for facial landmark localization. In Proceedings of the 2011 IEEE International Conference on Computer Vision Workshops (ICCV Workshops), Barcelona, Spain, 6–13 November 2011; pp. 2144–2151. [Google Scholar]
  56. Burgos-Artizzu, X.P.; Perona, P.; Dollár, P. Robust face landmark estimation under occlusion. In Proceedings of the IEEE International Conference on Computer Vision, Sydney, Australia, 1–8 December 2013; pp. 1513–1520. [Google Scholar]
  57. Wu, W.; Qian, C.; Yang, S.; Wang, Q.; Cai, Y.; Zhou, Q. Look at boundary: A boundary-aware face alignment algorithm. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 2129–2138. [Google Scholar]
  58. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  59. Lugaresi, C.; Tang, J.; Nash, H.; McClanahan, C.; Uboweja, E.; Hays, M.; Zhang, F.; Chang, C.L.; Yong, M.; Lee, J.; et al. MediaPipe: A framework for perceiving and processing reality. In Proceedings of the Third Workshop on Computer Vision for AR/VR at IEEE Computer Vision and Pattern Recognition (CVPR) 2019, Long Beach, CA, USA, 17 June 2019. [Google Scholar]
  60. Hu, G.; Xiao, Y.; Cao, Z.; Meng, L.; Fang, Z.; Zhou, J.T.; Yuan, J. Towards real-time eyeblink detection in the wild: Dataset, theory and practices. IEEE Trans. Inf. Forensics Secur. 2019, 15, 2194–2208. [Google Scholar] [CrossRef]
  61. Song, F.; Tan, X.; Liu, X.; Chen, S. Eyes closeness detection from still images with multi-scale histograms of principal oriented gradients. Pattern Recognit. 2014, 47, 2825–2838. [Google Scholar] [CrossRef]
  62. Jesorsky, O.; Kirchberg, K.J.; Frischholz, R.W. Robust face detection using the hausdorff distance. In Proceedings of the Audio-and Video-Based Biometric Person Authentication: Third International Conference, AVBPA 2001, Halmstad, Sweden, 6–8 June 2001; Proceedings 3. Springer: Cham, Switzerland, 2001; pp. 90–95. [Google Scholar]
  63. Martinez, A.; Benavente, R. The ar Face Database: Cvc Technical Report No. 24. 1998. Available online: https://portalrecerca.uab.cat/en/publications/the-ar-face-database-cvc-technical-report-24 (accessed on 8 May 2024).
  64. Gao, W.; Cao, B.; Shan, S.; Chen, X.; Zhou, D.; Zhang, X.; Zhao, D. The CAS-PEAL large-scale Chinese face database and baseline evaluations. IEEE Trans. Syst. Man Cybern. Part Syst. Hum. 2007, 38, 149–161. [Google Scholar]
  65. Huang, G.B.; Mattar, M.; Berg, T.; Learned-Miller, E. Labeled faces in the wild: A database forstudying face recognition in unconstrained environments. In Proceedings of the Workshop on Faces in ‘Real-Life’ Images: Detection, Alignment, and Recognition, Marseille, France, 17–20 October 2008. [Google Scholar]
  66. Duta, I.C.; Liu, L.; Zhu, F.; Shao, L. Improved Residual Networks for Image and Video Recognition. arXiv 2020, arXiv:2004.04989. [Google Scholar]
  67. Andriyanov, N.; Dementev, V.; Vasiliev, K.; Tashlinskii, A. Investigation of methods for increasing the efficiency of convolutional neural networks in identifying tennis players. Pattern Recognit. Image Anal. 2021, 31, 496–505. [Google Scholar] [CrossRef]
  68. Wu, W.; Peng, H.; Yu, S. Yunet: A tiny millisecond-level face detector. Mach. Intell. Res. 2023, 20, 656–665. [Google Scholar] [CrossRef]
  69. Lynch, K.M.; Park, F.C. Modern Robotics; Cambridge University Press: Cambridge, MA, USA, 2017. [Google Scholar]
  70. Sengupta, S.; Chen, J.C.; Castillo, C.; Patel, V.M.; Chellappa, R.; Jacobs, D.W. Frontal to profile face verification in the wild. In Proceedings of the 2016 IEEE Winter Conference on Applications of Computer Vision (WACV), Lake Placid, NY, USA, 7–10 March 2016; pp. 1–9. [Google Scholar]
  71. Moschoglou, S.; Papaioannou, A.; Sagonas, C.; Deng, J.; Kotsia, I.; Zafeiriou, S. Agedb: The first manually collected, in-the-wild age database. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Honolulu, HI, USA, 21–26 July 2017; pp. 51–59. [Google Scholar]
  72. Sun, Y.; Cheng, C.; Zhang, Y.; Zhang, C.; Zheng, L.; Wang, Z.; Wei, Y. Circle loss: A unified perspective of pair similarity optimization. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 6398–6407. [Google Scholar]
  73. Huang, Y.; Wang, Y.; Tai, Y.; Liu, X.; Shen, P.; Li, S.; Li, J.; Huang, F. Curricularface: Adaptive curriculum learning loss for deep face recognition. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 5901–5910. [Google Scholar]
  74. Meng, Q.; Zhao, S.; Huang, Z.; Zhou, F. Magface: A universal representation for face recognition and quality assessment. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 14225–14234. [Google Scholar]
  75. Ruiz, N.; Chong, E.; Rehg, J.M. Fine-grained head pose estimation without keypoints. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Salt Lake City, UT, USA, 18–23 June 2018; pp. 2074–2083. [Google Scholar]
  76. Hsu, H.W.; Wu, T.Y.; Wan, S.; Wong, W.H.; Lee, C.Y. Quatnet: Quaternion-based head pose estimation with multiregression loss. IEEE Trans. Multimed. 2018, 21, 1035–1046. [Google Scholar] [CrossRef]
  77. Bhagavatula, C.; Zhu, C.; Luu, K.; Savvides, M. Faster than real-time facial alignment: A 3d spatial transformer network approach in unconstrained poses. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 3980–3989. [Google Scholar]
  78. Feng, Y.; Wu, F.; Shao, X.; Wang, Y.; Zhou, X. Joint 3D face reconstruction and dense alignment with position map regression network. In Proceedings of the ECCV, Munich, Germany, 8–14 September 2018. [Google Scholar]
Figure 1. Overview of the integrated framework, beginning with pretraining on individual tasks, followed by pseudo-labeling to generate the pseudo-training dataset, which is then used in our multitask learning. The direction of all arrows indicates the order of application of the framework.
Figure 1. Overview of the integrated framework, beginning with pretraining on individual tasks, followed by pseudo-labeling to generate the pseudo-training dataset, which is then used in our multitask learning. The direction of all arrows indicates the order of application of the framework.
Sensors 24 03212 g001
Figure 2. Selecting the face recognition datasets to address the lack of face IDs and diversity in the background in the 300W-LP [22] dataset. Each arrow indicates why we use the VGGFace2 [28] and MS1MV3 [29] datasets.
Figure 2. Selecting the face recognition datasets to address the lack of face IDs and diversity in the background in the 300W-LP [22] dataset. Each arrow indicates why we use the VGGFace2 [28] and MS1MV3 [29] datasets.
Sensors 24 03212 g002
Figure 3. Overview of pseudo-labeled training dataset generation for multitask learning. The upper part illustrates the application of the pseudo-labeling framework of landmarks and poses to the face recognition datasets. The bottom part describes the method of pseudo-labeling face IDs for our training datasets, which involves duplicate ID verification and assigns new face ID labels to the VGGFace2 [28], MS1MV3 [29], and 300W-LP [22] datasets. The direction of all arrows indicates the order of application of the framework.
Figure 3. Overview of pseudo-labeled training dataset generation for multitask learning. The upper part illustrates the application of the pseudo-labeling framework of landmarks and poses to the face recognition datasets. The bottom part describes the method of pseudo-labeling face IDs for our training datasets, which involves duplicate ID verification and assigns new face ID labels to the VGGFace2 [28], MS1MV3 [29], and 300W-LP [22] datasets. The direction of all arrows indicates the order of application of the framework.
Sensors 24 03212 g003
Figure 4. Illustration of handling of closed or partially closed eyes using an eye blink network to address the limitations of the FLD task. We train the ResNet18 model on cropped eye blink datasets to predict the probability of eyes being open or closed. The red dots represent the eye’s landmarks before adjustment and the green dots represent the landmarks after adjustment.
Figure 4. Illustration of handling of closed or partially closed eyes using an eye blink network to address the limitations of the FLD task. We train the ResNet18 model on cropped eye blink datasets to predict the probability of eyes being open or closed. The red dots represent the eye’s landmarks before adjustment and the green dots represent the landmarks after adjustment.
Sensors 24 03212 g004
Figure 5. Illustration of sampling of a pseudo-landmark label from Algorithm 1. We used our pretrained network and Google’s MediaPipe [59] to generate a diverse set of pseudo-landmark samples. We applied index matching to select 68 landmarks from the 468 landmarks output by MediaPipe. The green arrows and dots show the flow and output of the pretrained network on the Algorithm 1, while the blue arrows and dots show the flow and output of the MediaPipe.
Figure 5. Illustration of sampling of a pseudo-landmark label from Algorithm 1. We used our pretrained network and Google’s MediaPipe [59] to generate a diverse set of pseudo-landmark samples. We applied index matching to select 68 landmarks from the 468 landmarks output by MediaPipe. The green arrows and dots show the flow and output of the pretrained network on the Algorithm 1, while the blue arrows and dots show the flow and output of the MediaPipe.
Sensors 24 03212 g005
Figure 6. Visualization of our pseudo-labeled training dataset. It contains 8.3 M images and 93 K face IDs with pose and landmark label annotations.
Figure 6. Visualization of our pseudo-labeled training dataset. It contains 8.3 M images and 93 K face IDs with pose and landmark label annotations.
Sensors 24 03212 g006
Figure 7. Illustration of our multitask learning network based on the modified IResNet [66], consisting of four layers. The first three layers serve as shared layers across all tasks, and the last layer is duplicated as a sublayer for each task. The direction of all arrows indicates the order of application of our network layers.
Figure 7. Illustration of our multitask learning network based on the modified IResNet [66], consisting of four layers. The first three layers serve as shared layers across all tasks, and the last layer is duplicated as a sublayer for each task. The direction of all arrows indicates the order of application of our network layers.
Sensors 24 03212 g007
Figure 8. Illustration of the calculation of λ P , depicting the computation of the mean Euclidean distance between the normalized projected landmark vector L P and the normalized predicted landmark vector L P R E D .
Figure 8. Illustration of the calculation of λ P , depicting the computation of the mean Euclidean distance between the normalized projected landmark vector L P and the normalized predicted landmark vector L P R E D .
Sensors 24 03212 g008
Figure 9. Visualization of loss functions of CosFace [38] and ours with t-sne results. We applied a new margin m p to the loss of CosFace for pose-invariant learning.
Figure 9. Visualization of loss functions of CosFace [38] and ours with t-sne results. We applied a new margin m p to the loss of CosFace for pose-invariant learning.
Sensors 24 03212 g009
Figure 10. Visualization of ground truth, SynergyNet [7], and our model for HPE and FLD on AFLW2000-3D [22].
Figure 10. Visualization of ground truth, SynergyNet [7], and our model for HPE and FLD on AFLW2000-3D [22].
Sensors 24 03212 g010
Figure 11. Visualized results of the FLD task between SynergyNet [7] and our model under various face ROIs. The original SynergyNet uses its face detector before facial landmark detection; we compared our model with SynergyNet with and without the detector.
Figure 11. Visualized results of the FLD task between SynergyNet [7] and our model under various face ROIs. The original SynergyNet uses its face detector before facial landmark detection; we compared our model with SynergyNet with and without the detector.
Sensors 24 03212 g011
Table 1. Datasets for face recognition, head pose estimation, and facial landmark detection.
Table 1. Datasets for face recognition, head pose estimation, and facial landmark detection.
Dataset# Images# IDsImage/ID# LandmarksPose
CASIA-WebFace [32]0.5 M10 K47-
VGGFace2 [28]3.3 M9 K363-
MS1MV3 [29]5.2 M93 K56-
MegaFace2 [34]4.7 M0.6 M7-
WebFace260M [35]260 M4 M21-
300W [30]3.8 K--68
AFLW [55]20 K--21
COFW [56]1.3 K--29
WFLW [57]7.5 K--98
BIWI [42]15 K---
CMU-Panoptic [43]1.3 M---
300W-LP [22]61 K--68
Table 2. Face recognition test datasets used for comparison of verification accuracy.
Table 2. Face recognition test datasets used for comparison of verification accuracy.
Dataset# of ID# of Images# of Verification Pairs
LFW [65]5.7 K13 K6 K
CFP-FP [70]5007 K7 K
AgeDB [71]56816.5 K6 K
IJB-C [31]3.5 K148.8 K15 M
Table 3. Evaluation results (%) for face recognition datasets LFW, CFP-FP, AgeDB, and IJB-C. Using TAR@FAR = 1 × 10 4 for IJB-C. Bold denotes the best performance, and higher values reflect better results.
Table 3. Evaluation results (%) for face recognition datasets LFW, CFP-FP, AgeDB, and IJB-C. Using TAR@FAR = 1 × 10 4 for IJB-C. Bold denotes the best performance, and higher values reflect better results.
MethodDatasetLFWCFP-FPAgeDBIJB-CFPS
SphereFace [37]CASIA99.42----
CosFace [38]CASIA99.73----
SphereFace (Re-Imp)MS1MV299.6798.4698.1791.77-
CosFace (Re-Imp)MS1MV299.7898.2698.1795.56-
CircleLoss [72]MS1M99.7396.02-93.95-
CurricularFace [73]MS1MV299.8098.3798.3296.10329.8
ArcFace [4]MS1MV299.8298.4998.0596.03331.5
ArcFace [4]MS1MV399.8399.0398.1796.5331.5
MagFace [74]MS1MV299.8398.4698.1795.97254.3
Partial FC [5]MS1MV399.8598.798.1196.08255.1
SwinFace [21]MS1MV299.8798.6098.1596.7370.3
Ours (pretrained)—ResNet50VGGFace299.3392.3894.2088.05321.7
Ours (pretrained)—ResNet50MS1MV399.8598.7098.1196.08321.7
Ours (STL)—IResNet50PLTD99.7297.7396.9094.05253.2
Ours (MTL)—IResNet50-MultiPLTD99.7097.2696.5394.17221.2
Ours (MTL)—IResNet100-MultiPLTD99.6897.2296.2994.09112.0
Table 4. Evaluation results (%) for our single-task learning (STL) and multitask learning (MTL) models on the IJB-C test dataset across all, soft, medium, and hard cases, with thresholds ranging from 1 × 10 5 to 1 × 10 1 . Bold denotes better performance between our STL and MTL models.
Table 4. Evaluation results (%) for our single-task learning (STL) and multitask learning (MTL) models on the IJB-C test dataset across all, soft, medium, and hard cases, with thresholds ranging from 1 × 10 5 to 1 × 10 1 . Bold denotes better performance between our STL and MTL models.
Training Dataset
(PLTD)
IJB-C (TAR@FAR)
1 × 10 5 1 × 10 4 1 × 10 3 1 × 10 2 1 × 10 1
Ours
(STL)
All Cases90.7394.0596.2897.8998.95
Soft Case89.0392.594.9496.7798.23
Medium Case76.4184.1989.3193.5196.47
Hard Case55.3468.2777.6485.9692.9
Ours
(MTL)
All Cases89.8694.1796.697.9598.95
Soft Case87.1892.4295.2696.9698.43
Medium Case77.2186.0391.2794.5796.99
Hard Case56.6471.4881.4988.6294.23
Table 5. Mean absolute error (MAE) comparison on AFLW2000-3D [22] for head pose estimation. Bold denotes the best performance, and lower values reflect better results.
Table 5. Mean absolute error (MAE) comparison on AFLW2000-3D [22] for head pose estimation. Bold denotes the best performance, and lower values reflect better results.
MethodYawPitchRollMAEFPS
FAN (12 points) [48]6.3612.38.719.12-
HopeNet [75]6.476.565.446.16323.2
FSANet [44]4.56.084.645.07389.6
3DDFA-TPAMI [54]4.335.984.304.87-
3DDFA-V2 [53]4.065.263.484.27-
QuatNet [76]3.975.623.924.15-
TriNet [45]4.25.774.043.97311.3
RankPose [46]2.994.753.253.66291.7
SynergyNet [7]3.424.092.553.35128.1
Ours (Pretrained)2.894.773.353.67320.8
Ours (MTL)2.804.272.93.32221.2
Table 6. Mean absolute error (MAE) comparison on BIWI [42] for head pose estimation. Bold denotes best performance.
Table 6. Mean absolute error (MAE) comparison on BIWI [42] for head pose estimation. Bold denotes best performance.
MethodYawPitchRollMAEFPS
3D-FAN [48]8.537.487.637.89-
HopeNet [75]4.816.613.274.90323.2
FSANet [44]4.274.962.764.00389.6
QuatNet [76]4.015.492.944.15-
TriNet [45]4.114.763.053.97311.3
RankPose [46]3.594.772.763.71291.7
Ours (pretrained)4.345.182.614.04320.8
Ours (MTL)3.235.032.363.54221.2
Table 7. Normalized mean square error (NMSE) percentage comparison on AFLW2000-3D [22] for 3D facial landmark detection across three yaw angle ranges, 0 to 30 , 30 to 60 , 60 to 90 , and the overall mean error. Bold denotes the best performance, and lower values reflect better results.
Table 7. Normalized mean square error (NMSE) percentage comparison on AFLW2000-3D [22] for 3D facial landmark detection across three yaw angle ranges, 0 to 30 , 30 to 60 , 60 to 90 , and the overall mean error. Bold denotes the best performance, and lower values reflect better results.
Method 0 30 30 60 60 90 MeanFPS
3DSTN [77]3.154.335.984.49-
3D-FAN [48]3.163.534.63.7684.3
PRNet [78]2.753.514.613.62261.3
3DDFA-PAMI [54]2.843.574.963.79-
3DDFA-v2 [53]2.633.424.483.51-
SynergyNet [7]2.653.304.273.41128.1
Ours (pretrained)3.083.744.533.78322.5
Ours (MTL)2.623.454.513.53221.2
Table 8. Comparison of the performance of the baseline model in single-task learning (STL) and multitask learning (MTL) configurations across a variety of tasks, including face recognition (FR), head pose estimation (HPE), and facial landmark detection (FLD).
Table 8. Comparison of the performance of the baseline model in single-task learning (STL) and multitask learning (MTL) configurations across a variety of tasks, including face recognition (FR), head pose estimation (HPE), and facial landmark detection (FLD).
MethodFRHPEFLDIJB-CAFLW2000-3D (MSE)AFLW2000-3D (MAE)# ParamsFPS
STL--94.05--24.6M321.7
STL---3.67-23.5M320.9
STL----3.7823.8M322.5
MTL-94.163.54-56.8M243.8
MTL-94.17-3.3460.1M237.9
MTL--3.523.3347.3M252.7
MTL94.173.533.3273.3M221.2
Table 9. Experimental results when changing the value of λ P in Equation (3) demonstrate the impact of the regularization intensity on the model performance across multiple benchmarks.
Table 9. Experimental results when changing the value of λ P in Equation (3) demonstrate the impact of the regularization intensity on the model performance across multiple benchmarks.
MethodIJB-CAFLW2000-3D (NMSE)AFLW2000-3d (MAE)BIWI (MAE)
L F L D ( λ P = 0 )94.163.553.583.70
L F L D ( λ P = 1 )94.173.563.673.81
L F L D ( λ P = α )94.173.533.323.54
Table 10. Experimental results showing the impact of multitask learning hyperparameters λ 1 and λ 2 on the test accuracy with various benchmarks. These hyperparameters are used in Equation (1) to balance the contributions of each task in the overall multitask learning framework. To check the change caused by λ 1 and λ 2 , we obtained the loss convergence ratio based on the convergence value of the loss function of each task.
Table 10. Experimental results showing the impact of multitask learning hyperparameters λ 1 and λ 2 on the test accuracy with various benchmarks. These hyperparameters are used in Equation (1) to balance the contributions of each task in the overall multitask learning framework. To check the change caused by λ 1 and λ 2 , we obtained the loss convergence ratio based on the convergence value of the loss function of each task.
λ 1 λ 2 IJB-C ( 1 × 10 4 )AFLW2000-3D (NMSE)AFLW2000-3D (MAE)Loss Convergence Ratio
0.10.194.225.124.140.98:0.00:0.02
0.1194.214.773.420.86:0.00:0.14
0.11094.224.563.350.37:0.00:0.63
0.110094.234.983.350.08:0.00:0.92
10.194.204.613.980.96:0.01:0.03
1194.194.213.410.85:0.01:0.14
11094.124.373.390.37:0.01:0.62
110094.184.273.380.08:0.01:0.91
100.194.883.993.940.94:0.05:0.02
10194.193.653.540.82:0.04:0.14
101094.173.613.330.35:0.02:0.63
1010094.173.603.330.06:0.00:0.93
1000.194.183.743.990.67:0.32:0.01
100194.193.553.590.61:0.29:0.10
1001094.173.533.320.31:0.15:0.54
10010094.173.543.320.04:0.03:0.94
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Lee, Y.; Jang, S.; Bae, H.B.; Jeon, T.; Lee, S. Multitask Learning Strategy with Pseudo-Labeling: Face Recognition, Facial Landmark Detection, and Head Pose Estimation. Sensors 2024, 24, 3212. https://doi.org/10.3390/s24103212

AMA Style

Lee Y, Jang S, Bae HB, Jeon T, Lee S. Multitask Learning Strategy with Pseudo-Labeling: Face Recognition, Facial Landmark Detection, and Head Pose Estimation. Sensors. 2024; 24(10):3212. https://doi.org/10.3390/s24103212

Chicago/Turabian Style

Lee, Yongju, Sungjun Jang, Han Byeol Bae, Taejae Jeon, and Sangyoun Lee. 2024. "Multitask Learning Strategy with Pseudo-Labeling: Face Recognition, Facial Landmark Detection, and Head Pose Estimation" Sensors 24, no. 10: 3212. https://doi.org/10.3390/s24103212

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop