Next Article in Journal
A Novel Impedance Matching of Class DE Inverter for High Efficiency, Wide Impedance WPT System
Previous Article in Journal
Spatial-Temporal Attention TCN-Based Link Prediction for Opportunistic Network
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Face-Inception-Net for Recognition

1
College of Information Science and Engineering, Henan University of Technology, Zhengzhou 450001, China
2
Key Laboratory of Grain Information Processing and Control, Henan University of Technology, Ministry of Education, Zhengzhou 450001, China
3
Henan Grain Big Data Analysis and Application Engineering Research Center, Henan University of Technology, Zhengzhou 450001, China
*
Author to whom correspondence should be addressed.
Electronics 2024, 13(5), 958; https://doi.org/10.3390/electronics13050958
Submission received: 29 December 2023 / Revised: 23 February 2024 / Accepted: 27 February 2024 / Published: 1 March 2024

Abstract

:
Face recognition in general scenarios has been saturated in recent years, but there is still room to enhance model performance in extreme scenarios and fairness situations. Inspired by the successful application of Transformer and ConvNet in computer vision, we propose a FIN-Block, which gives a more flexible composition paradigm for building a novel pure convolution model and provides a foundation for constructing a new framework for general face recognition in both extreme scenarios and fairness situations. FIN-Block-A uses a combination of stacked large-size convolution kernels and parallel convolution branches to ensure a large spatial receptive field while improving the module’s deep feature embedding and extraction capabilities. FIN-Block-B takes advantage of stacked orthogonal convolution kernels and parallel branches to balance model size and performance. By applying FIN-Block with an adapted convolution kernel size in different stages, we built a reasonable and novel framework Face-Inception-Net, and the performance of the model is highly competitive with ConvNeXt and InceptionNeXt. The models were trained on CASIA-WebFace and MS-wo-RFW databases and evaluated on 14 mainstream benchmarks, including LFW, extreme scene, and fairness test sets. The proposed Face-Inception-Net achieved the highest average [email protected] of 95.9% in all used benchmarks, fully demonstrating effectiveness and generality in various scenarios.

1. Introduction

In recent years, deep learning has made significant progress in the field of face recognition [1,2,3,4,5,6], with high-precision face recognition models trained on public face datasets now widely used in industrial fields, such as security, surveillance, and mobile applications. The tremendous success of face recognition is mainly attributed to three aspects: (1) large-scale training datasets, including CASIA Webface, VGGFace2, and MS-wo-RFW [5,6,7], have greatly improved the training accuracy of deep models [8]; (2) the network structure, such as GoogleNet [9] and ResNet [10] models, which form the backbone of face recognition models, help improve the performance of deep face recognition; (3) various loss functions are used as control signals during model training, introduced based on Euclidean distance [3] or angular/cosine margin [1,11] to enhance the distinctiveness of face features. These loss functions effectively guide the network to learn more discriminative features by emphasizing inter-class differences and intra-class consistency. However, the performance of DNNs in general scenarios, extreme conditions, and fairness still needs to be improved, making the design of new universal face recognition models a challenge. In IMAN [7], the impact of race-biased face datasets on model performance is demonstrated. In SFace [8], experiments show that the recognition rate of face recognition models in extreme conditions needs to be improved. Therefore, integrating existing achievements in face recognition with new technologies to design innovative universal face recognition models has significant scientific importance and potential application value.
Face recognition is a highly versatile task, facing challenges such as varying poses, lighting, obstructions, facial expressions, and age-related changes [12]. While current research often concentrates on specific aspects of face recognition, there is a pressing need for a universally applicable solution. Recent studies have made notable advancements. Zhang et al. [13] improved the Inception module, significantly enhancing performance on datasets featuring masked faces. LIAAD [14] developed a novel, lightweight attention-based method that, through knowledge distillation, improves accuracy and robustness against age variations in face recognition. Zhang et al. delved into face recognition and its applications, including expression recognition [15], generation [15], and face reconstruction [16,17]. The evolution of hardware and new technologies have spurred the development of more advanced model architectures. Taking inspiration from the successful incorporation of GELU, AdamW [18,19], and Dropout layers in Transformers, ConvNet led to ConvNext [20], a new convolutional model architecture that leverages the ResNet framework. ConvNext has surpassed Transformer models in performance, renewing interest in convolutional models in the computer vision field. InceptionNeXt [21] was introduced; it employs parallel branches, avoiding large convolutional kernels, which leads to superior performance and faster inference, offering new avenues for refining face recognition models. Despite these advancements, the quest for a universally applicable face recognition model has seen limited progress in recent years. The unique demands of face recognition tasks pose significant challenges in making precise adjustments to model structures to enhance versatility across various scenarios.
On the LFW dataset [22], face recognition technologies based on ResNet and Trans-former [23] have reached a performance saturation point. To thoroughly evaluate the effectiveness of our proposed improvements, we introduced verification datasets from multiple scenarios [7,22,24,25,26,27,28,29,30,31] and measured the models’ universality by calculating their average precision across these datasets. Much of the current research in face recognition focuses on Inception and its ResNet variants, demonstrating the possibility of achieving exceptional performance at lower computational costs by optimizing local network topology. Additionally, multilayer perceptrons (MLPs) [32] have been proven to achieve results comparable to complex attention mechanisms, providing theoretical support for enhancing the performance of pure convolutional network structures. Inspired by the parallel branches of InceptionNeXt [21] and the stacked convolutional kernels of Inception, we introduce two types of modules, FIN-Block-A/B. FIN-Block-A combines the design of stacked large convolutional kernels with parallel branches, aimed at expanding the model’s spatial receptive field and enhancing the efficiency of deep feature embedding and extraction. FIN-Block-B, on the other hand, optimizes the balance between model size and performance through stacking orthogonal convolutional kernels and parallel branches.
Based on these innovations, we constructed a novel Face-Inception-Net model, following the InceptionNeXt [21] stage structure design and adjusting the size of convolutional kernels at different stages to form a framework that is both rational and innovative. Compared to existing ConvNet [20] and InceptionNeXt [21] models, our model demonstrates strong competitive performance. The novelty of this study lies in the meticulous stage-wise improvements to the model structure: using large pure convolutional kernels for feature extraction in Stage 1; employing stacked large convolutional kernels for deeper feature mining in Stage 2; and using orthogonal square kernels of varying sizes in Stages 3/4 to optimize the balance between model performance and computational efficiency. We used the average [email protected] across different scenarios as a metric for model universality. Our experimental results not only proved that increasing model width is more effective than depth in large-scale face recognition tasks but also demonstrated the wide applicability of our method through evaluations on the extreme scene and fairness test sets [7].
The main contributions of our work are summarized as follows.
  • We present an innovative FIN block that utilizes a hybrid of state-of-the-art convolution and Inception technologies. This architecture is designed to offer an expansive spatial receptive field alongside advanced capabilities for the extraction of deep feature embeddings.
  • We use convolutional branches that combine different stages and dimensions to build the novel Face-Inception-Net. Evaluated on the extreme scene and fairness test set, experiments show that our model outperforms the compared models.
  • By analyzing the experimental data, we found that our model can better balance the gap between the optimal solution and the average solution of the model. This means that our model is more powerful and stable.
Our model helps to accelerate the research of face recognition model architecture in complex cases.

2. Related Work

In this section, we commence with an overview of deep learning-based algorithms specifically tailored for deep face recognition. Subsequently, we delve into methodologies for integrating transformations with convolutions, exploring their synergistic potential. In conclusion, we highlight recent innovations in face recognition algorithms, tailored for a variety of scenarios, thereby presenting a comprehensive update on the state-of-the-art in the field.

2.1. Deep Face Recognition

Deep face recognition systems, underpinned by convolutional neural networks (CNNs) [33], have exhibited notable performance. These algorithms excel in extracting discriminative features from facial images, subsequently encoding them into an n-dimensional vector space. This space is distinguished by minimized intra-class distances and maximized inter-class distances, enabling precise identification and differentiation of individual faces.
During the training process of deep face recognition models, two key elements play a crucial role: (1) network architecture and (2) loss functions. For the network architecture, researchers have employed both traditional and novel network structures to design these models [6]. Well-known classical network structures, such as AlexNet [34], VGGNet [2], GoogleNet [9], ResNet [10], and SENet [35], are frequently utilized as backbone networks for face recognition models. For instance, Taigman et al. [36] pioneered the use of a nine-layer CNN with multiple locally connected layers for face recognition. In addition, novel architectures specifically designed for face recognition have been proposed to improve model performance [6]. For the latter role, researchers have proposed novel loss functions or modified existing ones to improve the discriminative capabilities of convolutional networks for distinguishing individuals. In early deep face recognition algorithms [37,38], cross-entropy loss is used for feature learning. Schroff et al. [39] introduced triplet loss as a means to directly optimize the embedded features, further improving the performance of the model. Recent research [1,3,11,40] has predominantly adopted angle/cosine margin loss, integrating an angle margin penalty to amplify the discriminative capabilities of features. By leveraging true labels as supervisory signals, these loss functions facilitate feature separability, fostering a pronounced differentiation among various categories.
Face recognition has realized remarkable achievements through years of development and technological advancement. Building on these accomplishments, we have undertaken our research. The optimization of the model and the application of loss functions are complex tasks, each requiring meticulous attention to detail and a deep understanding of underlying principles.

2.2. The Fusion of Transformer and Convolution

Transformer [23] is a widely used architecture in natural language processing (NLP) based on the self-attention architecture. The Transformer model has achieved top results in computer classification and segmentation tasks, but the model execution efficiency is far behind convolution. How to effectively combine Transformer and convolution has become a hot issue.
The application of Transformer [23] to various computer vision tasks has become a prominent trend. So far, research on models has branched out in several directions. (1) The fusion of two models has led to the development of hybrid architectures, such as Conformer (by the National University of Science and Technology) [41], Next-ViT [42], and CoTNet [43]. (2) The combination of the CNN module and the Transformer layer can be used to construct novel models, such as the module-parallel design scheme proposed by Mobile-Former [39] and CMT [38]. (3) Researchers have explored specialized approaches, including using the convolutional model as a guide to training tiny Transformers [44] and using the Transformer to improve the ConvNet convolutional training technique [20]. The Transformer model has gained prominence in computer vision, but CNNs have also made significant progress. More specifically, pure convolutional networks, such as ConvNeXt [20] and InceptionNeXt [21], exploit the inherent strengths of convolution in computer vision. By integrating “network modernization” techniques [41] with training techniques from Transformer [45,46] and other technologies, these models have achieved superior performance on Imagenet, surpassing that of the Transformer model.
The performance of convolutional network models in unconstrained face recognition has reached a plateau of accuracy. Nonetheless, with the relentless advancement of hardware and the continuous emergence of novel technologies, there is an ongoing trend toward designing new model structures that boast enhanced performance.

2.3. Face Recognition in Special Scenes

Conventional face recognition technology has matured to the point of fulfilling commercial deployment criteria, indicating a saturation in its development for standard applications. However, when it comes to extreme scenarios and fairness assessments, the model’s performance reveals room for improvement.
Face Transformer [47] proves that Transformer can be used for face recognition. Considering that the original Transformer may ignore the inter-patch information, Face Transformer modifies the patch generation process and makes the tokens with sliding patches that overlap each other. Inspired by the successful application of self-attention in computer vision, authors of [48] proposed a convolutional visual self-attention network (CVSAN), which uses self-attention to improve convolutional operators. A study conducted by [7] revealed the existence of racial bias in face recognition algorithms, resulting in significant performance degradation in real-world face recognition systems. To address racial bias, a novel IMAN method was proposed to bridge the domain gap, and its potential and effectiveness were validated through experiments. Authors of [14] proposed a novel lightweight attention angle distillation (LIAAD) method, which introduces a learning paradigm that can effectively distill the age-invariant attention and angle knowledge of these teachers into a lightweight student network, enabling it to be more robust with higher accuracy and robustness to the age factor.
The assessment of a model’s performance is fundamentally linked to the test set employed. In light of the saturation encountered in routine face recognition tasks, we incorporated test sets designed for fairness [7] and extreme scenarios [8]. This approach allowed for a thorough evaluation of our model’s capabilities, with the objective of showcasing its adaptability and versatility across a broad range of conditions.

3. Method

3.1. FIN-Block

Formulation of ConvNeXt Block. ConvNeXt [20] is a contemporary CNN model known for its streamlined architecture. In each ConvNeXt block, the input X undergoes an initial processing step with a depth convolution to propagate information across spatial dimensions. As shown in Figure 1, the token mixer is effectively established by using a depthwise convolution in ConvNeXt,
X = TokenMixer ( X ) = DWConv k × k C C ( X )
where DWConv k × k C C denotes depthwise convolution with a kernel size of k × k. In ConvNeXt, k is set as 7 by default.
Formulation of Inception depthwise convolution. As shown in Figure 1, conventional depthwise convolution with a large kernel size significantly hampers model speed. Therefore, InceptionNeXt [21] leaves partial channels unchanged and denotes them as a branch of identity mapping. For the processing channels, InceptionNeXt decomposes the depthwise operations with Inception style [9,49,51], using 3 × 3, 1 × k w , and k h × 1 kernels as branches. Specifically, for input X, we split it into four groups along the channel dimension,
X hw , X w , X h , X id = Split ( X ) = X : , : g , X : g : 2 g , X : 2 g : 3 g , X : 3 g : ,
where g is the channel numbers of convolution branches. Set ratio r g to determine the branch channel numbers by g = r g C . The splitting inputs are fed into different parallel branches,
X hw = DWConv k s × k s g g g X hw , X w = DWConv 1 × k b g g g X w , X h = DWConv k b × 1 g g g X h , X id = X id ,
where k s denotes the small square kernel size set as 3 by default and k b represents the band kernel size set as 11 by default. Finally, the outputs from each branch are concatenated:
X = Concat ( X h w , X w , X h , X i d ) .
Formulation of FIN-Block. As shown in Figure 1, Inception [9] shows excellent performance with relatively low computational cost. MLPs [32,49,50] have also been shown to achieve comparable performance to attention mechanisms. Inception has shown that 1 × 1 convolutions are used to compute reductions before the expensive 3 × 3 and 5 × 5 convolutions. Besides being used as reductions, they include the use of rectified linear activation, which makes them dual-purpose. The ubiquitous use of dimension reduction allows for shielding a large number of input filters from the last stage to the next layer, reducing their dimension before convolving over them with a large patch size. Another practical and useful aspect of this design is that it aligns with the intuition that visual information should be processed at various scales and then aggregated so that the next stage can abstract features from different scales simultaneously. The design of the FIN-Block is inspired by Inception, which uses deep convolutional stacking and orthogonal bands kernel stacking to improve the performance of the model. The modified formula is shown below.
X = Conv 1 × 1 C C / 3 ( X ) , X v = DWConv k s × k s g g g ( X ) , X w 1 = DWConv k b × k b g g g X , X w 2 = DWConv k b × k b g g g X w 1 , X o = Concat X , X v , X w 2 .
X = Conv 1 × 1 C C / 2 ( X ) , X w 1 = DWConv 1 × k b g g g X , X w 2 = DWConv k b × 1 g g g X w 1 , X o = Concat ( X , X w 2 ) ,
where k s and k b denote the small square kernel size set as 11 by default. Finally, the outputs from each are concatenated. The equation represents FIN-Block-A on the up and FIN-Block-B on the down. FIN-Block-A comprises a 1 × 1 small square kernel, large square kernel, and stack of two large square kernel. FIN-Block-B comprises a 1 × 1 small square kernel and stack of two orthogonal bands kernel.

3.2. Face-Inception-Net

Based on FIN-Block-A and FIN-Block-B, we developed a set of models called Face-Inception-Net. Considering ConvNeXt [20] and InceptionNeXt as our primary baselines for comparison, we mainly followed their approach to construct models of different sizes. Similar to ResNet [10] and ConvNeXt, we adopted a 4-stage framework. For tiny models, the stage numbers are (3, 3, 9, 3), and for the base model, the stage numbers are (3, 3, 27, 3). Additionally, we experimented with stage numbers (5, 5, 10, 5) and dimensions (96, 192, 384, 786) for TV2 and stage numbers (5, 5, 10, 5) and dimensions (192, 384, 576, 786) for SV2. Our model uses BatchNorm instead of LayerNorm in the stem to prioritize speed. Detailed model configurations are shown in Table 1 and Figure 2. In terms of model structure, we mainly experimented with two popular stage-block combinations and explored the impact of block kernel size, model depth, and width on performance to find the optimal model that meets the requirements of face recognition.

3.3. Loss Function

In our experiments, after evaluating various loss functions [1,11], we selected ArcFace [1] as our definitive choice. ArcFace introduces a unified formula that blends the angular margin principles of both ArcFace and CosFace [11] with traditional softmax loss, providing a versatile framework for facial recognition. This adaptable formula, incorporating angular margin penalties from SphereFace [4], ArcFace [1], and CosFace [11], enhances the discriminative power of embeddings in face recognition models. The integration of these three penalty methods is succinctly captured in the following formula:
( cos ( m 1 θ + m 2 ) m 3 ) ,
where m 1 represents the multiplicative angular margin used in SphereFace, m 2 denotes the additive angular margin found in ArcFace, and m 3 signifies the additive cosine margin from CosFace. The generalized loss expression is as follows:
L = 1 N i = 1 N log e s cos m 1 θ y i + m 2 m 3 e s cos m 1 θ y i + m 2 m 3 + j = 1 , j y i n e s   cos θ j ,
where N is the number of samples; s denotes the scaling factor to control the logits’ magnitude; m 1 , m 2 and m 3 are parameters that adjust the margin effect, allowing the formula to adapt between Softmax ( m 1 = 1 , m 2 = 0 , m 3 = 0 ) , SphereFace ( m 1 1 , m 2 = 0 , m 3 = 0 ) , CosFace ( m 1 = 1 , m 2 = 0 , m 3 = m ) , and ArcFace ( m 1 = 1 , m 2 = m , m 3 = 0 ) ; m denotes the margin added to the angle θ y i between the feature vector of the i-th sample and the weight vector of its corresponding true class y i , which is central to the angular margin concept; and n represents the total number of classes.

4. Experiment

Following the Open Face Recognition pipeline [4], we trained Face-Inception-Net on a face database comprising images (X) and corresponding labels (y) in a supervised manner. We utilized a carefully designed loss function [1,3,11], which enhanced the discriminative capacity of the face image embedding. The LFW dataset served as the benchmark for evaluating face recognition performance. We introduced the dataset, including 2 training datasets and 14 test datasets, and assessed Face-Inception-Net’s effectiveness through various experiments. To ensure a fair comparison with established baselines, we maintained consistent hyperparameters during training. The experimental outcomes, achieved using learning rate decay and without model fine-tuning, are detailed further in Appendix A Table A1.

4.1. Dataset and Evaluate

For Training. We utilized CASIA-WebFace and MS-wo-RFW [5,6,7] as our training datasets. CASIA-WebFace comprises approximately 500,000 images of 1000 celebrities, featuring a wide variety of poses, lighting conditions, expressions, ages, and races. This diversity makes it an ideal dataset for developing robust face recognition systems. MS-wo-RFW, containing 3.5 million images from 7500 celebrities, includes specific test sets for African, Asian, Caucasian, and Indian demographics, facilitating region-specific evaluations. The CASIA-WebFace dataset is made available by InsightFace, while MS-wo-RFW is provided by IMAN [7].
For Testing. Given privacy concerns in face recognition, benchmarks typically encompass 5000 to 8000 face-matching tasks. We utilized 14 benchmarks for accuracy assessment. These include LFW [22], which tests face recognition algorithms under real-world conditions, AgeDB-30 [24] for age invariance, CFP-FF and CFP-FP [25] for pose variations, and VGG2-FP [26] for frontal to profile comparisons. For extreme scenarios, we used CALFW [27] to assess real-world variations, CPLFW [28] for cross-pose recognition, MLFW [29] for ethnic diversity, SLLFW [30] for low-light conditions, and TALFW [31] for temporal aging variations. Additionally, RFW [7] provided fairness metrics by evaluating racial fairness across diverse racial groups (Africa-test, Asian-test, Caucasian-test, and Indian-test). AgeDB-30, VGG2-FP, and CALFW were also pivotal in evaluating cross-age recognition performance.
For Evaluation. To thoroughly assess model performance, we utilized 14 test sets, categorizing them for detailed analysis. Our evaluation metrics include mean accuracy, ROC curve plots, and performance analysis across common, extreme, and fairness scenarios. For training, we adopted the ConvNeXt method. Unlike traditional approaches that focus solely on the optimal value, disregarding its correlation with fluctuating epoch results (where the optimum across different benchmarks rarely coincides in a single epoch), we averaged the accuracy across benchmarks for a more comprehensive performance and generalization assessment. Furthermore, we compared the optimal value and average [email protected] differences between our model and established models [34], like ConvNeXt and InceptionNeXt. Our model demonstrates exceptional performance in extreme scenarios and on racially balanced face datasets.

4.2. Implementation and Training Techniques

The effectiveness of ConvNeXt and InceptionNeXt shows that the final performance of the network is influenced by the training process as well as the architecture design. Transformers introduce different model designs for vision along with different training methods (such as the AdamW [18,19] optimizer) to the computer vision (CV) domain. Optimization strategies and related hyperparameter settings can help the model find better saddle points, but more hyperparameters also increase the difficulty of model training. Because of the variety of training tasks and loss functions, we experimented with different optimizers and learning rate decays. In our study, we first trained a baseline model using ResNet50 and Inception-ResNet-V1. After many rounds of pre-training, the number of rounds was set to approximately 120 epochs in the loop dataset. The optimizer adopts the Yogi optimizer proposed by Google [18,19,46,52]. By controlling the increase of the effective learning rate, better performance could be obtained under the similar convergence theory guarantee. Then, the EMA wrapper optimizer [37] was used to replace the training parameters with the moving average of the deep network training parameters. The learning rate decay uses learning rate warmup and cosine learning rate decay. The full set of hyperparameters we used can be found in Appendix A. In terms of the learning rate itself, this training method led to a 0.43% improvement in accuracy.
In addition, the improvement in training speed of technical teams, such as Mixed Precision [53], XLA [54], and TF32, is huge, reducing the time cost by more than half. This shows that training skills can affect performance in face recognition. We used a fixed training method with the same hyperparameters throughout the comparative training process, and each reported [email protected] is the average of all benchmarks obtained from training. The experiments were implemented in Tensorflow and performed on a Linux operating system using two Nvidia A100-40G GPUs.

4.3. Result

Training on CASIA-WebFace and MS-wo-RFW. Our initial learning rate was lr = 0.002, and the epoch was 120. There was a 1/10 × epoch linear warmup and then a cosine decay schedule; the learning rate decay multiple was 1/10. Batchsize = 256 for the CASIA-WebFace dataset, and Batchsize = 512 for the MS-wo-RFW dataset. We did not use data augmentation. For regularization, we used label smoothing [9], stochastic depth [55], and weight decay. Like ConvNeXt, we also used LayerScale [56] in the stem, a technique that helps train deep models. We used the exponential moving average (EMA) [57] because we find that it facilitates the overfitting of larger models. Our code is based on the TensorFlow [58,59,60] and TensorFlow Addons [60] libraries.
Figure 3 and Figure 4 below show our first experiments where the values in the MLFW, SLLFW, and TALFW test sets were not used. We compare the improved Face-Inception-Net model with the state-of-the-art ConvNeXt and InceptionNeXt models. As shown in Figure 3 and Figure 4, our proposed model shows strong competitiveness on both the CASIA-WebFace and MS-wo-RFW datasets. With similar model sizes and flops, Face-Inception-Net consistently outperforms ConvNeXt and InceptionNeXt in terms of accuracy. And, comparable experimental results were obtained on two different scale datasets, CASIA-WebFace and MS-wo-RFW.
The training results of the CASIA-WebFace dataset are shown in Figure 3 and Figure 5. The accuracy Acc-mean([email protected]) obtained by the ResNet and Inception-ResNet models are 0.8263, 0.8137, 0.8157, and 0.8167, respectively. This is the result of using the old model training. Then, we reproduced ConvNeXt-Tiny and InceptionNeXt-Tiny, and the obtained accuracy was 0.8807 and 0.8833, respectively. So far, we conclude that using Inception to transform ConvNeXt is effective in the field of face recognition. We designed five models of Face-Inception-Net (FIN-T, FIN-S, FIN-B, FIN-TV2, and FIN-SV2), among which FIN-TV2 obtained the highest Acc-mean([email protected]) value of 0.9106. When we experimented with deeper networks, the performance of the model decreased, such as the small and basic models of ConvNeXt and InceptionNeXt. The FIN-S, FIN-B, and FIN-SV2 models we developed also showed a decrease in model performance. In this regard, we suggest that the size of the CASIA-WebFace dataset is too small and the complexity of the model is too high, resulting in overfitting. Furthermore, the performance of our model on the TALFW (Transferable Adversarial Labeled Faces in the Wild) dataset is worse, which is a bad sign that our model is not well-defended and vulnerable to attacks.
The results of training on the MS-wo-RFW dataset are shown in Figure 4 and Figure 6. The accuracy Acc-mean([email protected]) obtained by the ResNet and Inception-ResNet models are 0.9040, 0.8878, 0.9059, and 0.8933, respectively. Training on the CASIA-WebFace dataset, we found that the use of a deep model resulted in performance degradation. The result of training with ConvNeXt is that the performance degradation of the model is suppressed but not completely eliminated. The results of the InceptionNeXt model improved compared to ConvNeXt, but the results were unstable. We propose that the model achieved an improvement in accuracy when the stage was (3,3,9,3) and (5,5,10,5). And, under the deeper small and base models, the accuracy of our model improved, which solved the problem of model performance degradation. In comparison, under the same parameters of our model, the accuracy of the FIN-T model is 0.9533, which is higher than the 0.9477 and 0.9494 of ConvNeXt-Tiny and InceptionNeXt-Tiny. The highest performance of FIN-SV2 is 0.9599, which is the best result in our experiments. Furthermore, the performance of our model on the TALFW (Transferable Adversarial Labeled Faces in the Wild) dataset leads to the same conclusion as on CASIA-WebFace.
On the fairness and extreme scene test sets. Our findings exceed those achieved by IMAN, based on comparative data sourced from their paper. The analysis shows our model, FIN-SV2, significantly surpasses previous models in accuracy while reducing racial bias. This improvement is particularly notable in test sets for African, Asian, and Indian demographics, where our model achieved accuracies over 0.94. These results, detailed in Table 2, suggest that using large datasets can mitigate racial bias in model performance effectively. Additionally, we assessed the model’s accuracy in various special scenarios, with results shown as ROC curves in Figure 5 and Figure 6. The experiments confirm that our improvements enhance facial recognition performance in complex scenarios. An analysis of the difference between optimal and average values further demonstrates the versatility and robustness of our approach.
Optimal and mean analysis. In our experiments, we used several benchmarks and averaged all [email protected] results for consistent accuracy evaluation. The analysis showed a difference between the optimal and average values for models trained on various datasets, as shown in Table 3. Using a larger training set reduces this gap. Our FIN-Tiny and FIN-Tiny-V2 models outperformed others on the CASIA-WebFace dataset, whereas the InceptionNeXt series was more effective with the MS-wo-RFW dataset, highlighting our model’s improved performance. Further examination of the model structure revealed that a deeper stacked convolution structure boosts feature extraction, thus enhancing model performance. The gap between the optimal and average values reflects the model’s generalizability, with a smaller gap indicating stronger stability and generalizability across different datasets. However, this balance is not as optimal as with InceptionNeXt, indicating room for improvement.
Ablation studies. Table 4 highlights the three main improvements in our model training: the design of FIN-Block-A and FIN-Block-B, model structure enhancements, and training techniques. We conducted Face-Inception-Net ablation studies using FIN-Tiny as a baseline, focusing on the utilization of different FIN-Blocks at various stages. Specifically, ConvNeXt-proposed blocks are employed in the stem, while in the subsequent three stages, we systematically remove each block to assess our module’s effectiveness. Notably, removing the FIN-Block-B-11 model in stage 3 results in the most significant performance drop, from 90.48 to 89.96. These experiments underscore the importance of these blocks, as large convolution kernels increase the model’s receptive field, enabling deeper feature extraction and achieving an optimal balance between model parameters and performance.
As shown in Table 4, the enhancement of our model training primarily encompasses three aspects. The first is changing the Inception stage configuration from (3,3,9,3) to (5,5,10,5), leading to the proposal of FIN-Tiny-V2. Furthermore, we adjusted the dimensions from (96,192,384,786) to (192,384,576,786), introducing FIN-Small-V2. This experiment validated our improvement direction, showing increased parameters and FLOPs for FIN-Tiny-V2 alongside enhanced model performance. Specifically, in the CASIA-WebFace database, FIN-Tiny-V2 outperformed both FIN-Small and FIN-Base, demonstrating the superiority of the (5,5,10,5) configuration for face recognition. FIN-Small-V2 also delivered remarkable results on the MS-wo-RFW dataset, with fewer parameters yet significant performance gains compared to FIN-Base. This confirms that, in face recognition, widening the model is more efficacious than deepening it, aligning with conclusions from the inception paper. Additionally, the adoption of a learning rate warmup and decay strategy, as opposed to a fixed learning rate, further improved our model accuracy by 0.43%.

5. Conclusions and Future Work

This study introduces Face-Inception-Net, utilizing advanced convolution techniques for improved performance. Unlike other architectures, FIN-Block-A and FIN-Block-B offer versatile composition, enabling swift adjustments in model width and depth for various tasks. FIN-Block-B’s extensive convolutional kernel stack provides a wide spatial receptive field and strong deep feature embedding, achieving an ideal balance between performance and model size. Our experiments showed that expanding the model and adjusting stage sizes significantly enhances face recognition performance. Moreover, results from the extreme scene and fairness test sets significantly demonstrate our method’s versatility and effectiveness.
Our findings indicate that incorporating fine-grained blocks at various stages enhances performance. These results were achieved in a single training session without subsequent fine-tuning, indicating the potential for further optimization and mean analysis improvement relative to InceptionNeXt. Future efforts will focus on simplifying the model structure and integrating pre-training with fine-tuning to maximize model performance.

Author Contributions

Conceptualization, Q.Z. and M.Z.; Methodology, Q.Z.; Software, X.W.; Validation, X.W.; Investigation, L.L.; Data curation, L.L.; Writing—original draft, Q.Z. and X.W.; Writing—review & editing, M.Z.; Visualization, P.L.; Supervision, M.Z.; Funding acquisition, Q.Z. and M.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This work is partially supported by the Key R&D Project of Henan Province (No. 231111222100), the Science and Technology Research Project of Henan Province (No. 222102320039), the Collaborative Innovation Project of Zhengzhou (No. 22ZZRDZX41) and the High-level Talent Fund Project of Henan University of Technology (No. BS2019027).

Data Availability Statement

We used two training sets CASIA-WebFace and MS-wo-RFW as well as LFW, AgeDB-30, CFP (CFP-FF, CFP-FP), MLFW, SLLFW, TALFW, RFW (Africa-test, Asian-test, Caucasian-test, and Indian-tests) 14 validation datasets. CASIA-WebFace train dataset and LFW, AgeDB-30, CFP (CFP-FF, CFP-FP), MLFW, SLLFW, TALFW, RFW test dataset analysed during the current study are available in the Github repository, https://github.com/deepinsight/insightface/wiki/Dataset-Zoo (accessed on 26 February 2024). MS-wo-RFW and RFW data that support the findings of this study are available from Pattern Recognition and Intelligent System Laboratory (PRIS Lab) but restrictions apply to the availability of these data, which were used under license for the current study, and so are not publicly available. Data are however available from the authors upon reasonable request and with permission of Pattern Recognition and Intelligent System Laboratory (PRIS Lab). This data set can be obtained by contacting [email protected]. Details can be found at http://whdeng.cn/RFW/Trainingdataste.html (accessed on 26 February 2024).

Acknowledgments

We offer thanks to the public computing service platform of Henan University of Technology for providing computing power support. We also thank Pattern Recognition and Intelligent System Laboratory (PRIS Lab), Beijing University of Posts and Telecommunications, for providing the dataset.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A

In this appendix, we provide the experimental results in Table A2 and Table A3 and hyperparameters in Table A1.
Table A1. Hyperparameters of Face-Inception-Net on CASIA-WebFace and MS-wo-RFW face recognition.
Table A1. Hyperparameters of Face-Inception-Net on CASIA-WebFace and MS-wo-RFW face recognition.
Face-Inception-Net
CASIA-WebFaceMS-wo-RFW
Input resolution112 × 112112 × 112
Epochs120120
Batch size256512
OptimizerYogiYogi
Learning rate0.0020.0025
Learning rate decayCosine 0.1Cosine 0.1
Warmup epochs0.1 × epochs0.1 × epochs
Weight decayNoneNone
Dropout0.60.6
LayerScale initialization1.00 × 10−61.00 × 10−6
EMA decay rate0.990.99
Table A2. Using the performance of the model trained by CASIA-WEBFACE, we recorded the accuracy of the model on each test set, including the optimal value and the optimal average value obtained in the round. * in the model brackets represents the Optim Value of the training process. Since the extreme scene face recognition dataset was not fully used during training, there are no optimal data for MLFW, SLLFW and TALFW in the table.
Table A2. Using the performance of the model trained by CASIA-WEBFACE, we recorded the accuracy of the model on each test set, including the optimal value and the optimal average value obtained in the round. * in the model brackets represents the Optim Value of the training process. Since the extreme scene face recognition dataset was not fully used during training, there are no optimal data for MLFW, SLLFW and TALFW in the table.
Average Value/Optim Value (*) ResNet50ResNet101Incep-RV1Incep-RV2Convn Ext-TConvn Ext-SConvn Ext-BInnext-TInnext-SInnext-BFIN-TFIN-SFIN-BFIN-TV2FIN-SV2
lfw96.6296.9096.0596.6398.8298.4898.4298.9098.6098.3899.2299.0598.9299.2099.05
agedb_3084.0583.6085.6283.2790.4389.6088.9890.3889.8889.6092.0791.4891.8592.6592.43
cfp_ff96.3196.2795.7095.8798.7397.8098.1398.4098.0398.1199.0698.5698.5699.1698.89
cfp_fp80.3777.8984.3779.2790.6789.9789.7991.1988.5088.7693.3991.6791.4994.1494.06
calfw83.3584.8886.5786.6291.5590.2289.1391.1089.8790.1592.1391.8591.8392.4892.23
cplfw75.8276.8777.9575.4784.4883.4582.8383.9582.6782.4886.0085.4585.7287.5387.20
vgg2_fp80.9881.5483.5280.7889.4289.3088.1088.9287.5288.5491.5891.1090.9692.1692.38
African_test69.3569.5366.5068.2878.2875.4875.9378.5876.8375.1881.9279.5877.7882.4381.25
Asian_test73.0072.8871.8871.3579.7077.5278.1079.7578.6877.0383.3780.2379.9283.6282.17
Caucasian_test81.6382.3780.5781.7888.8886.7886.4088.2087.3286.1091.8289.6289.3091.8590.18
Indian_test75.8876.3074.0075.9282.3280.1780.3382.2382.1881.1086.4583.8883.1586.3385.50
MLFW61.8858.9559.8358.0070.7566.1367.2069.4366.7567.6271.4270.1870.4873.4871.57
SLLFW87.7887.9888.0388.1595.7793.9593.9894.8594.8293.5796.7296.0295.8397.2596.85
TALFW81.1783.9575.6775.6769.7872.2571.3570.5873.5873.0267.9872.8573.4067.7367.35
lfw (*)96.9297.1096.9597.0798.8798.4898.4598.9598.6598.3899.2799.0799.1099.3799.27
agedb_30 (*)86.5084.2584.5383.5090.7590.0789.8291.4590.5290.5093.4892.2391.9393.4293.02
cfp_ff (*)96.8796.0196.3496.2798.3998.0098.1398.5698.1998.0798.9698.5998.6099.0098.97
cfp_fp (*)84.3981.9183.4780.2092.6790.6190.1091.6090.2989.5993.3792.3791.7795.0094.36
calfw (*)87.7786.6286.6285.5391.0090.5090.7391.3091.0090.5392.9791.9091.9592.8592.67
cplfw (*)77.5375.4776.5277.7084.8083.7283.2284.5083.3783.0386.7785.3885.7287.7887.57
vgg2_fp (*)83.7882.2482.6681.4290.3689.3088.7889.3488.8088.7691.9091.1090.9692.7692.68
African_test (*)69.3268.8869.7270.7577.2576.4776.4879.0076.8375.2282.0879.8778.9582.8881.43
Asian_test (*)73.1872.9774.5072.7779.7078.1778.0779.7878.9078.8783.1880.9581.0583.2582.17
Caucasian_test (*)82.2381.7581.9782.2288.2787.3087.3788.6587.5586.9891.0789.6389.5892.0791.08
Indian_test (*)76.7875.8276.8276.8782.3081.9581.7283.1282.4581.8785.7884.7083.8086.3886.12
MLFW (*)NoneNoneNoneNoneNoneNoneNoneNoneNoneNoneNoneNoneNoneNoneNone
SLLFW (*)NoneNoneNoneNoneNoneNoneNoneNoneNoneNoneNoneNoneNoneNoneNone
TALFW (*)NoneNoneNoneNoneNoneNoneNoneNoneNoneNoneNoneNoneNoneNoneNone
Table A3. Using the performance of the model trained by MS-wo-RFW, we recorded the accuracy of the model on each test set, including the optimal value and the optimal average value obtained in the round. * in the model brackets represents the Optim Value of the training process. Since the extreme scene face recognition dataset was not fully used during training, there are no optimal data for MLFW, SLLFW and TALFW in the table.
Table A3. Using the performance of the model trained by MS-wo-RFW, we recorded the accuracy of the model on each test set, including the optimal value and the optimal average value obtained in the round. * in the model brackets represents the Optim Value of the training process. Since the extreme scene face recognition dataset was not fully used during training, there are no optimal data for MLFW, SLLFW and TALFW in the table.
Average Value Optim Value (*)Resnet 50Resnet 101Inception-RV1Inception-RV2Convn Ext-TConvn Ext-SConvn Ext-BInnext-TInnext-SInnext-BFIN-TFIN-SFIN-BFIN-TV2FIN-SV2
lfw99.4799.1299.1598.8399.6599.6099.7099.5599.7299.5899.6399.7099.7399.7399.75
agedb_3094.0391.6393.7592.3397.2097.6597.1097.3297.4797.4797.3097.3797.5097.7297.52
cfp_ff98.8098.2498.9398.3199.6099.6199.5999.7199.7399.7399.7399.6999.6199.7799.81
cfp_fp86.6086.3987.6087.4693.2993.5093.3793.1492.8692.7194.5495.0194.5095.5495.26
calfw94.1593.1394.0892.5295.5395.7595.5595.5595.6795.6295.6795.8295.7895.5295.82
cplfw85.2084.3884.6583.0288.9389.4788.9788.5089.0288.4089.1789.7789.3090.0390.38
vgg2_fp89.9889.7489.7489.2893.7093.5092.8293.9293.4893.5693.9094.1294.0294.4094.74
African_test83.7881.6883.4280.2391.8891.7591.7292.3592.3593.2293.1094.0793.7293.5294.33
Asian_test84.3881.7884.3880.2791.6091.3390.7092.5292.7292.4793.3093.5393.8093.6594.27
Caucasian_test92.7091.3291.9290.2297.0297.4297.0397.7798.0097.7398.0298.2798.2298.5398.73
Indian_test87.4084.2586.8883.4892.8892.9892.6293.9293.9394.2394.6394.8895.0294.8595.33
MLFW70.2070.1373.2770.2382.2383.5383.5081.0782.8381.8885.1285.9084.7385.9384.82
SLLFW97.3396.0896.5395.5399.0399.0799.0899.1899.3299.2099.2099.4799.3399.3899.38
TALFW73.5877.3871.2778.0266.9869.3367.6868.1869.4267.1864.7269.6569.1266.2865.02
lfw (*)99.4299.1299.4799.2799.68None99.7299.6899.7299.7399.78None99.7799.7899.78
agedb_30 (*)94.1092.4794.2892.8597.32None97.4597.3297.4897.5397.47None97.6297.7297.60
cfp_ff (*)99.0098.3699.0998.6099.77None99.6799.7799.7799.7699.74None99.7499.7999.83
cfp_fp (*)88.7388.0787.6787.3493.24None93.3793.2493.3192.7194.63None94.5095.5495.69
calfw (*)94.1292.9794.1793.5595.77None95.6895.7795.7295.7895.72None95.8795.7595.88
cplfw (*)84.9783.8785.3784.3888.82None89.1588.8289.3588.6089.63None89.8290.2290.47
vgg2_fp (*)90.3689.5290.6089.6693.92None93.1293.9293.4893.8494.22None94.4294.6295.02
African_test (*)83.3580.2584.1381.7092.78None91.7292.7892.6793.2293.28None93.9594.3295.05
Asian_test (*)84.4281.3784.8382.2592.60None91.3092.6092.8092.7393.50None93.8393.6794.27
Caucasian_test (*)92.9390.1393.1791.4397.95None97.0397.9598.1797.9298.25None98.4398.5398.78
Indian_test (*)87.3384.4787.5384.9894.22None93.0594.2294.1294.3894.72None95.4095.4395.63
MLFW (*)NoneNoneNoneNoneNoneNoneNoneNoneNoneNoneNoneNoneNoneNoneNone
SLLFW (*)NoneNoneNoneNoneNoneNoneNoneNoneNoneNoneNoneNoneNoneNoneNone
TALFW (*)NoneNoneNoneNoneNoneNoneNoneNoneNoneNoneNoneNoneNoneNoneNone

References

  1. Deng, J.; Guo, J.; Xue, N.; Zafeiriou, S. Arcface: Additive angular margin loss for deep face recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 4690–4699. [Google Scholar]
  2. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  3. Schroff, F.; Kalenichenko, D.; Philbin, J. Facenet: A unified embedding for face recognition and clustering. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 815–823. [Google Scholar]
  4. Liu, W.; Wen, Y.; Yu, Z.; Li, M.; Raj, B.; Song, L. Sphereface: Deep hypersphere embedding for face recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 212–220. [Google Scholar]
  5. Wang, M.; Deng, W. Mitigating bias in face recognition using skewness-aware reinforcement learning. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 9322–9331. [Google Scholar]
  6. Wang, M.; Deng, W. Deep face recognition: A survey. Neurocomputing 2021, 429, 215–244. [Google Scholar] [CrossRef]
  7. Wang, M.; Deng, W.; Hu, J.; Tao, X.; Huang, Y. Racial faces in the wild: Reducing racial bias by information maximization adaptation network. In Proceedings of the IEEE International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 692–702. [Google Scholar]
  8. Zhong, Y.; Deng, W.; Hu, J.; Zhao, D.; Li, X.; Wen, D. SFace: Sigmoid-constrained hypersphere loss for robust face recognition. Proc. IEEE Trans. Image Process. 2021, 30, 2587–2598. [Google Scholar] [CrossRef] [PubMed]
  9. Szegedy, C.; Ioffe, S.; Vanhoucke, V.; Alemi, A. Inception-v4, inception-resnet and the impact of residual connections on learning. In Proceedings of the AAAI Conference on Artificial Intelligence, San Francisco, CA, USA, 4–9 February 2017; Volume 31. [Google Scholar]
  10. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  11. Wang, H.; Wang, Y.; Zhou, Z.; Ji, X.; Gong, D.; Zhou, J.; Li, Z.; Liu, W. Cosface: Large margin cosine loss for deep face recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 5265–5274. [Google Scholar]
  12. Wang, S.; Liu, P. A New Feature Extraction Method Based on the Information Fusion of Entropy Matrix and Covariance Matrix and Its Application in Face Recognition. Entropy 2015, 17, 4664–4683. [Google Scholar] [CrossRef]
  13. Zhang, M.; Zhang, Y.; Zhang, Q. Attention-Mechanism-Based Models for Unconstrained Face Recognition with Mask Occlusion. Electronics 2023, 12, 3916. [Google Scholar] [CrossRef]
  14. Truong, T.D.; Duong, C.N.; Quach, K.G.; Le, N.; Bui, T.D.; Luu, K. LIAAD: Lightweight attentive angular distillation for large-scale age-invariant face recognition. Neurocomputing 2023, 543, 126198. [Google Scholar] [CrossRef]
  15. Zhang, Y.; Xu, X.; Zhao, Y.; Wen, Y.; Tang, Z.; Liu, M. Facial prior guided micro-expression generation. IEEE Trans. Image Process. 2023, 33, 525–540. [Google Scholar] [CrossRef] [PubMed]
  16. Shahreza, H.O.; Marcel, S. Template inversion attack against face recognition systems using 3d face reconstruction. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Paris, France, 2–6 October 2023; pp. 19662–19672. [Google Scholar]
  17. Yang, H.; Zhu, H.; Wang, Y.; Huang, M.; Shen, Q.; Yang, R.; Cao, X. Facescape: A large-scale high quality 3d face dataset and detailed riggable 3d face prediction. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 601–610. [Google Scholar]
  18. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
  19. Loshchilov, I.; Hutter, F. Fixing Weight Decay Regularization in Adam. 2018. Available online: https://api.semanticscholar.org/CorpusID:3312944 (accessed on 26 February 2024).
  20. Liu, Z.; Mao, H.; Wu, C.Y.; Feichtenhofer, C.; Darrell, T.; Xie, S. A convnet for the 2020s. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 11976–11986. [Google Scholar]
  21. Yu, W.; Zhou, P.; Yan, S.; Wang, X. Inceptionnext: When inception meets convnext. arXiv 2023, arXiv:2303.16900. [Google Scholar]
  22. Huang, G.B.; Mattar, M.; Berg, T.; Learned-Miller, E. Labeled faces in the wild: A database forstudying face recognition in unconstrained environments. In Proceedings of the Workshop on Faces in’Real-Life’Images: Detection, Alignment, and Recognition, Marseille, France, 12–18 October 2008. [Google Scholar]
  23. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is all you need. In Proceedings of the Advances in Neural Information Processing, Long Beach, CA, USA, 4–9 December 2017; Volume 30. [Google Scholar]
  24. Moschoglou, S.; Papaioannou, A.; Sagonas, C.; Deng, J.; Kotsia, I.; Zafeiriou, S. Agedb: The first manually collected, in-the-wild age database. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Honolulu, HI, USA, 21–26 July 2017; pp. 51–59. [Google Scholar]
  25. Sengupta, S.; Chen, J.C.; Castillo, C.; Patel, V.M.; Chellappa, R.; Jacobs, D.W. Frontal to profile face verification in the wild. In Proceedings of the IEEE Winter Conference on Applications of Computer Vision, Lake Placid, NY, USA, 7–10 March 2016; pp. 1–9. [Google Scholar]
  26. Cao, Q.; Shen, L.; Xie, W.; Parkhi, O.M.; Zisserman, A. Vggface2: A dataset for recognising faces across pose and age. In Proceedings of the IEEE International Conference on Automatic Face & Gesture Recognition, Xi’an, China, 15–19 May 2018; pp. 67–74. [Google Scholar]
  27. Zheng, T.; Deng, W.; Hu, J. Cross-age lfw: A database for studying cross-age face recognition in unconstrained environments. arXiv 2017, arXiv:1708.08197. [Google Scholar]
  28. Zheng, T.; Deng, W. Cross-pose lfw: A database for studying cross-pose face recognition in unconstrained environments. Beijing Univ. Posts Telecommun. Tech. Rep. 2018, 5, 1–6. [Google Scholar]
  29. Wang, C.; Fang, H.; Zhong, Y.; Deng, W. Mlfw: A database for face recognition on masked faces. In Proceedings of the Chinese Conference on Biometric Recognition, Beijing, China, 11–13 November 2022; Springer: Berlin/Heidelberg, Germany, 2022; pp. 180–188. [Google Scholar]
  30. Deng, W.; Hu, J.; Zhang, N.; Chen, B.; Guo, J. Fine-grained face verification: FGLFW database, baselines, and human-DCMN partnership. Pattern Recognit. 2017, 66, 63–73. [Google Scholar] [CrossRef]
  31. Zhong, Y.; Deng, W. Towards transferable adversarial attack against deep face recognition. Proc. IEEE Trans. Inf. Forensics Secur. 2020, 16, 1452–1466. [Google Scholar] [CrossRef]
  32. Tolstikhin, I.O.; Houlsby, N.; Kolesnikov, A.; Beyer, L.; Zhai, X.; Unterthiner, T.; Yung, J.; Steiner, A.; Keysers, D.; Uszkoreit, J.; et al. Mlp-mixer: An all-mlp architecture for vision. Adv. Neural Inf. Process. Syst. 2021, 34, 24261–24272. [Google Scholar]
  33. Kim, J.; Lee, J.K.; Lee, K.M. Deeply-recursive convolutional network for image super-resolution. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 26 June–1 July 2016; pp. 1637–1645. [Google Scholar]
  34. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. ImageNet classification with deep convolutional neural networks. Commun. ACM 2017, 60, 84–90. [Google Scholar] [CrossRef]
  35. Hu, J.; Shen, L.; Sun, G. Squeeze-and-excitation networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 7132–7141. [Google Scholar]
  36. Taigman, Y.; Yang, M.; Ranzato, M.; Wolf, L. Deepface: Closing the gap to human-level performance in face verification. In Proceedings of the IEEE conference on computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 1701–1708. [Google Scholar]
  37. Zhang, M.; Lucas, J.; Ba, J.; Hinton, G.E. Lookahead optimizer: K steps forward, 1 step back. In Proceedings of the 33rd Conference on Neural Information Processing Systems (NeurIPS 2019), Vancouver, BC, Canada, 8–14 December 2019; Volume 32. [Google Scholar]
  38. Guo, J.; Han, K.; Wu, H.; Tang, Y.; Chen, X.; Wang, Y.; Xu, C. Cmt: Convolutional neural networks meet vision transformers. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 12175–12185. [Google Scholar]
  39. Chen, Y.; Dai, X.; Chen, D.; Liu, M.; Dong, X.; Yuan, L.; Liu, Z. Mobile-former: Bridging mobilenet and transformer. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 5270–5279. [Google Scholar]
  40. Zhang, M.M.; Shang, K.; Wu, H. Learning deep discriminative face features by customized weighted constraint. Neurocomputing 2019, 332, 71–79. [Google Scholar] [CrossRef]
  41. Peng, Z.; Huang, W.; Gu, S.; Xie, L.; Wang, Y.; Jiao, J.; Ye, Q. Conformer: Local features coupling global representations for visual recognition. In Proceedings of the IEEE international Conference on Computer Vision, Montreal, QC, Canada, 11–17 October 2021; pp. 367–376. [Google Scholar]
  42. Li, J.; Xia, X.; Li, W.; Li, H.; Wang, X.; Xiao, X.; Wang, R.; Zheng, M.; Pan, X. Next-vit: Next generation vision transformer for efficient deployment in realistic industrial scenarios. arXiv 2022, arXiv:2207.05501. [Google Scholar]
  43. Li, Y.; Yao, T.; Pan, Y.; Mei, T. Contextual transformer networks for visual recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 45, 1489–1500. [Google Scholar] [CrossRef] [PubMed]
  44. Elliott, D.; Otero, C.E.; Wyatt, S.; Martino, E. Tiny transformers for environmental sound classification at the edge. arXiv 2021, arXiv:2103.12157. [Google Scholar]
  45. Liu, Z.; Lin, Y.; Cao, Y.; Hu, H.; Wei, Y.; Zhang, Z.; Lin, S.; Guo, B. Swin transformer: Hierarchical vision transformer using shifted windows. In Proceedings of the IEEE International Conference on Computer Vision, Montreal, QC, Canada, 11–17 October 2021; pp. 10012–10022. [Google Scholar]
  46. Zaheer, M.; Reddi, S.; Sachan, D.; Kale, S.; Kumar, S. Adaptive methods for nonconvex optimization. In Proceedings of the 32nd International Conference on Neural Information Processing Systems, Montreal, QC, Canada, 3–8 December 2018; Volume 31. [Google Scholar]
  47. Zhong, Y.; Deng, W. Face transformer for recognition. arXiv 2021, arXiv:2103.14803. [Google Scholar]
  48. Ge, Y.; Liu, H.; Du, J.; Li, Z.; Wei, Y. Masked face recognition with convolutional visual self-attention network. Neurocomputing 2023, 518, 496–506. [Google Scholar] [CrossRef]
  49. Szegedy, C.; Vanhoucke, V.; Ioffe, S.; Shlens, J.; Wojna, Z. Rethinking the inception architecture for computer vision. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 2818–2826. [Google Scholar]
  50. Benz, P.; Ham, S.; Zhang, C.; Karjauv, A.; Kweon, I. Adversarial robustness comparison of vision transformer and mlp-mixer to cnns. arXiv 2021, arXiv:2110.02797. [Google Scholar]
  51. Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; Rabinovich, A. Going deeper with convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 1–9. [Google Scholar]
  52. Bottou, L. Stochastic gradient descent tricks. In Neural Networks: Tricks of the Trade: Second Edition; Springer: Berlin/Heidelberg, Germany, 2012; pp. 421–436. [Google Scholar]
  53. Micikevicius, P.; Narang, S.; Alben, J.; Diamos, G.; Elsen, E.; Garcia, D.; Ginsburg, B.; Houston, M.; Kuchaiev, O.; Venkatesh, G.; et al. Mixed precision training. arXiv 2017, arXiv:1710.03740. [Google Scholar]
  54. Abdelfattah, A.; Anzt, H.; Boman, E.G.; Carson, E.; Cojean, T.; Dongarra, J.; Fox, A.; Gates, M.; Higham, N.J.; Li, X.S.; et al. A survey of numerical linear algebra methods utilizing mixed-precision arithmetic. Int. J. High Perform. Comput. Appl. 2021, 35, 344–369. [Google Scholar] [CrossRef]
  55. Huang, G.; Sun, Y.; Liu, Z.; Sedra, D.; Weinberger, K.Q. Deep networks with stochastic depth. In Proceedings of the Computer Vision–ECCV 2016: 14th European Conference, Amsterdam, The Netherlands, 11–14 October 2016; Part IV 14. Springer: Berlin/Heidelberg, Germany, 2016; pp. 646–661. [Google Scholar]
  56. Touvron, H.; Cord, M.; Sablayrolles, A.; Synnaeve, G.; Jégou, H. Going deeper with image transformers. In Proceedings of the IEEE International Conference on Computer Vision, Montreal, QC, Canada, 10–17 October 2021; pp. 32–42. [Google Scholar]
  57. Polyak, B.T.; Juditsky, A.B. Acceleration of stochastic approximation by averaging. SIAM J. Control. Optim. 1992, 30, 838–855. [Google Scholar] [CrossRef]
  58. Abadi, M. TensorFlow: Learning functions at scale. In Proceedings of the 21st ACM SIGPLAN International Conference on Functional Programming, Nara, Japan, 18–22 September 2016; p. 1. [Google Scholar]
  59. Agarwal, A.; Barham, P.; Brevdo, E.; Chen, Z.; Citro, C.; Corrado, G. Tensorflow: A system for large-scale machine learning. In Proceedings of the 12th USENIX Conference on Operating Systems Design and Implementation, USENIX Association, Savannah, GA, USA, 2–4 November 2016. [Google Scholar]
  60. Silaparasetty, N. Machine Learning Concepts with Python and the Jupyter Notebook Environment: Using Tensorflow 2.0; Springer: Berlin/Heidelberg, Germany, 2020. [Google Scholar]
Figure 1. Block illustration of ConvNeXt [20], InceptionNeXt [21], FIN-Block-A, and FIN-Block-B. ConvNeXt uses a large convolution kernel 7 × 7 instead of the original 3 × 3, uses the depth convolution layer to move up to replace the original convolution layer, and combines Norm and MLP [32,49,50] to form the ConvNeXt block. Compared to ConvNeXt, InceptionNeXt is efficient in that it decomposes the expensive large kernel depth convolution into four efficient parallel branches. FIN-Block-A and FIN-Block-B refer to the Inception structure, using 1 × 1 convolution and stacked depth convolution instead of large kernel depth convolution. The variable convolution kernel size allows for a more reasonable model structure.
Figure 1. Block illustration of ConvNeXt [20], InceptionNeXt [21], FIN-Block-A, and FIN-Block-B. ConvNeXt uses a large convolution kernel 7 × 7 instead of the original 3 × 3, uses the depth convolution layer to move up to replace the original convolution layer, and combines Norm and MLP [32,49,50] to form the ConvNeXt block. Compared to ConvNeXt, InceptionNeXt is efficient in that it decomposes the expensive large kernel depth convolution into four efficient parallel branches. FIN-Block-A and FIN-Block-B refer to the Inception structure, using 1 × 1 convolution and stacked depth convolution instead of large kernel depth convolution. The variable convolution kernel size allows for a more reasonable model structure.
Electronics 13 00958 g001
Figure 2. The overall architecture of Face-Inception-Net. The FIN-Block-A and FIN-Block-B structures are shown in Figure 1. The dim-s, k-s, and num arrays of each stage are variable, and we experimented with the combinations in Table 1.
Figure 2. The overall architecture of Face-Inception-Net. The FIN-Block-A and FIN-Block-B structures are shown in Figure 1. The dim-s, k-s, and num arrays of each stage are variable, and we experimented with the combinations in Table 1.
Electronics 13 00958 g002
Figure 3. The performance of the model using CASIA-WebFace training. The specific data can be seen in the Appendix A Table A2. We evaluated the model performance according to three aspects: LFW and age test set, extreme scene test set, and fairness test set. Except for the TALFW test set, our model achieves the best performance on all other benchmarks.
Figure 3. The performance of the model using CASIA-WebFace training. The specific data can be seen in the Appendix A Table A2. We evaluated the model performance according to three aspects: LFW and age test set, extreme scene test set, and fairness test set. Except for the TALFW test set, our model achieves the best performance on all other benchmarks.
Electronics 13 00958 g003
Figure 4. The performance of the model using MS-wo-RFW training. The specific data can be seen in Appendix A Table A3. The model trained with MS-wo-RFW outperforms CASIA-WebFace in all aspects and is also largely consistent in distribution, demonstrating the validity of our experiments. We evaluated the model performance according to three aspects: LFW and age test set, extreme scene test set, and fairness test set. Except for the TALFW test set, our model achieved the best performance on all other benchmarks.
Figure 4. The performance of the model using MS-wo-RFW training. The specific data can be seen in Appendix A Table A3. The model trained with MS-wo-RFW outperforms CASIA-WebFace in all aspects and is also largely consistent in distribution, demonstrating the validity of our experiments. We evaluated the model performance according to three aspects: LFW and age test set, extreme scene test set, and fairness test set. Except for the TALFW test set, our model achieved the best performance on all other benchmarks.
Electronics 13 00958 g004
Figure 5. ROC performance on CASIA-WebFace for related models. We divided the benchmark into five aspects to plot the ROC curve and analyze the model’s performance on various specific tasks, i.e., ALL Test Set ROC Curve, LFW and common Test Set ROC Curve, Extreme scene Test Set ROC Curve, Fairness Test Set ROC Curve, and Age Test Set ROC Curve. Our model achieves optimal performance on all ROC curves, as seen in the locally zoomed images. In one of the ROC curves for the extreme scene, the negative effect of TALFW is masked by the inclusion of five datasets.
Figure 5. ROC performance on CASIA-WebFace for related models. We divided the benchmark into five aspects to plot the ROC curve and analyze the model’s performance on various specific tasks, i.e., ALL Test Set ROC Curve, LFW and common Test Set ROC Curve, Extreme scene Test Set ROC Curve, Fairness Test Set ROC Curve, and Age Test Set ROC Curve. Our model achieves optimal performance on all ROC curves, as seen in the locally zoomed images. In one of the ROC curves for the extreme scene, the negative effect of TALFW is masked by the inclusion of five datasets.
Electronics 13 00958 g005
Figure 6. ROC performance on MS-wo-RFW for related models. The use of MS-wo-RFW training results in a better manipulative treatment than CASIA-WebFace, which can be seen to have a better effect on the Large-scale data sets. The distribution of other ROC curves is similar for the model trained on the two datasets.
Figure 6. ROC performance on MS-wo-RFW for related models. The use of MS-wo-RFW training results in a better manipulative treatment than CASIA-WebFace, which can be seen to have a better effect on the Large-scale data sets. The distribution of other ROC curves is similar for the model trained on the two datasets.
Electronics 13 00958 g006
Table 1. Face-Inception-Net model configuration. It has a similar model configuration to Swin [45] and ConvNeXt [20].
Table 1. Face-Inception-Net model configuration. It has a similar model configuration to Swin [45] and ConvNeXt [20].
Stage# of TokensLayer SpecificationFace-Inception-Net
TSBTV2SV2
1
Stem
H 4 × H 4 Down-
sampling
Kernel size4 × 4, stride 4
Embed. dim.9612896192
ConvNeXt
block
DWConv kernel size7 × 7
MLP ratio4
# of blocks35
2
Stage
H 8 × H 8 Down-
sampling
Kernel size2 × 2, stride 2
Embed. dim.192256192384
FIN-Block-ADWConv kernel size ( 11 × 11 ) , 11 × 11 11 × 11
Conv. group ratio1/8
MLP ratio4
# of blocks35
3
Stage
H 16 × H 16 Down-
sampling
Kernel size2 × 2, stride 2
Embed. dim.384512384576
FIN-Block-BDWConv kernel size 11 × 1 1 × 11
Conv. group ratio1/8
MLP ratio4
# of blocks92710
4
Stage
H 32 × H 32 Down-
sampling
Kernel Size2 × 2, stride 2
Embed. dim.7861024786
FIN-Block-BDWConv kernel size 7 × 1 1 × 7
Conv. group ratio1/8
MLP ratio3
# of blocks35
BatchNorm + Dropout + Flatten + Dense (512) + BatchNorm
Table 2. Face fairness test set performance. The first three sets of data in the table come from the paper. The rest of the data are obtained from our experiments.
Table 2. Face fairness test set performance. The first three sets of data in the table come from the paper. The rest of the data are obtained from our experiments.
ModelsCaucasian-TestIndian-TestAsian-TestAfrican-Test
Casia [7]92.1588.0083.9884.93
MS-wo-RFW [7]93.9292.9890.6090.98
Iman [7]94.7894.1591.1591.42
Inception-RV1 [9]91.9286.8783.3584.32
ResNet50 [10]92.7087.4083.7884.38
ConvNeXt-T-v1 [20]97.6893.9592.1092.53
InceptionNeXt-T [21]97.7793.9592.2892.57
FIN-Tiny 98.2394.2293.0293.22
FIN-Small98.0894.7593.7893.10
FIN-Base98.2295.0293.7293.80
FIN-Tiny-V298.5394.8593.5293.65
FIN-Small-V298.7395.3394.3394.27
Table 3. The gap between the optimal and average performance reflects the model’s generality. The difference between the model’s optimal and average performance is presented in the table for the Casia and RFW datasets, indicating the disparity between the best and average values. This metric reflects the model’s generalizability; a smaller value signifies stronger generalizability.
Table 3. The gap between the optimal and average performance reflects the model’s generality. The difference between the model’s optimal and average performance is presented in the table for the Casia and RFW datasets, indicating the disparity between the best and average values. This metric reflects the model’s generalizability; a smaller value signifies stronger generalizability.
Model#Param. (M)FLOPs (G)Value Casia(%)Value RFW (%)
Inception-RV1 [9]331.230.580.39
Inception-RV2 [9]62.92.40.720.36
ResNet-50 [10]491.271.170.35
ResNet-101 [10]70.62.420.540.31
ConvNEXT-T-v1 [20]39.52.410.510.32
ConvNeXt-S-v1 [20]61.24.530.540.25
ConvNeXt-B-v1 [20]101.47.970.610.19
InceptionNeXt-T [21]32.91.020.420.16
InceptionNeXt-S [21]54.32.060.580.15
InceptionNeXt-B [21]90.63.660.580.14
FIN-Tiny 34.61.090.320.21
FIN-Small57.32.210.310.20
FIN-Base98.44.340.360.19
FIN-Tiny-V244.41.480.280.19
FIN-Small-V267.44.150.370.19
Table 4. Ablation for Face-Inception-Net on CASIA-WebFace face recognition benchmark. FIN-Tiny was utilized as the baseline for the ablation study. Acc-mean([email protected]) accuracy on the validation set is reported.
Table 4. Ablation for Face-Inception-Net on CASIA-WebFace face recognition benchmark. FIN-Tiny was utilized as the baseline for the ablation study. Acc-mean([email protected]) accuracy on the validation set is reported.
AblationVariant#Param. (M)FLOPs. (G)Acc-Mean ([email protected])
BaselineFIN-Tiny34.641.0990.48
BranchWithout FIN-Block-A34.561.0890.23
Without FIN-Block-B-1134.101.0789.96
Without FIN-Block-B-733.851.0990.28
Stagestage (3,3,9,3)–(5,5,10,5)44.411.4891.06
dim (96,192,384,786)–(192,384,576,786)67.404.1590.49
Training SkillWithout linear warmup and a cosine decaying schedule afterward.34.641.0990.05
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhang, Q.; Wang, X.; Zhang, M.; Lu, L.; Lv, P. Face-Inception-Net for Recognition. Electronics 2024, 13, 958. https://doi.org/10.3390/electronics13050958

AMA Style

Zhang Q, Wang X, Zhang M, Lu L, Lv P. Face-Inception-Net for Recognition. Electronics. 2024; 13(5):958. https://doi.org/10.3390/electronics13050958

Chicago/Turabian Style

Zhang, Qinghui, Xiaofeng Wang, Mengya Zhang, Lei Lu, and Pengtao Lv. 2024. "Face-Inception-Net for Recognition" Electronics 13, no. 5: 958. https://doi.org/10.3390/electronics13050958

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop