You are currently viewing a new version of our website. To view the old version click .
Sensors
  • Article
  • Open Access

27 July 2023

Innovative Hybrid Approach for Masked Face Recognition Using Pretrained Mask Detection and Segmentation, Robust PCA, and KNN Classifier

,
,
and
1
Computer Science Department, Faculty of Computing and Artificial Intelligence, Beni Suef University, Beni-Suef 62511, Egypt
2
Computer Science Department, Faculty of Science, Minia University, Minia 61519, Egypt
3
Computer Science Department, Faculty of Computers and Artificial Intelligence, University of Sadat City, Sadat City 32897, Egypt
4
Electrical Engineering Department, Faculty of Engineering, Minia University, Minia 61519, Egypt
This article belongs to the Special Issue Biometric Recognition System Based on Iris, Fingerprint and Face

Abstract

Face masks are widely used in various industries and jobs, such as healthcare, food service, construction, manufacturing, retail, hospitality, transportation, education, and public safety. Masked face recognition is essential to accurately identify and authenticate individuals wearing masks. Masked face recognition has emerged as a vital technology to address this problem and enable accurate identification and authentication in masked scenarios. In this paper, we propose a novel method that utilizes a combination of deep-learning-based mask detection, landmark and oval face detection, and robust principal component analysis (RPCA) for masked face recognition. Specifically, we use pretrained ssd-MobileNetV2 for detecting the presence and location of masks on a face and employ landmark and oval face detection to identify key facial features. The proposed method also utilizes RPCA to separate occluded and non-occluded components of an image, making it more reliable in identifying faces with masks. To optimize the performance of our proposed method, we use particle swarm optimization (PSO) to optimize both the KNN features and the number of k for KNN. Experimental results demonstrate that our proposed method outperforms existing methods in terms of accuracy and robustness to occlusion. Our proposed method achieves a recognition rate of 97%, which is significantly higher than the state-of-the-art methods. Our proposed method represents a significant improvement over existing methods for masked face recognition, providing high accuracy and robustness to occlusion.

1. Introduction

The COVID-19 pandemic has affected how people interact with each other. To mitigate the spread of the pandemic, governments have instituted rules such as wearing masks, staying away from others, and staying at home. Wearing masks helps protect people from the epidemic; however, it makes it difficult for facial recognition systems to recognize people’s faces.
It is hard to keep important information safe using traditional passwords and security measures; therefore, researchers have been focusing on a special kind of technology called biometric technology that is proficient at maintaining security and is very difficult for people to imitate [1]. Consequently, face biometric technology has received more attention to recognize a person correctly or recognize facial emotions [2,3]. Biometric facial recognition technology is widely used in various applications such as security systems, access control, and law enforcement.
With the widespread use of masks due to the COVID-19 pandemic, facial recognition systems face significant challenges in identifying masked faces. This has created a need for new methods that can accurately identify individuals wearing masks. Because of the COVID-19 pandemic, the masked face detection and recognition problems have become the most important area to innovate new methods and algorithms to detect and recognize people who are wearing or not wearing mask to reduce and prevent spread of COVID-19. Matching a masked face with unmasked or masked faces is the goal of the masked face recognition problem (MFR).

1.1. Motivation

Face masks are used in a variety of settings and industries, and there are many jobs that require the use of face masks to protect workers and the public. Some examples of jobs that commonly require the use of face masks include healthcare workers, food service workers, construction workers, manufacturing workers, retail workers, hospitality workers, transportation workers, educational workers, and public safety workers.
Healthcare workers, including doctors, nurses, and other medical professionals, use face masks to protect themselves and their patients from the spread of infectious diseases.
Workers in the food service industry, including cooks, servers, and other staff, use face masks to prevent the spread of germs and bacteria in food preparation and service areas. This helps to protect both workers and customers from the spread of infectious diseases.
Construction workers often work near one another, making it difficult to maintain social distancing. Face masks are used to protect workers from the spread of respiratory droplets on job sites and in work trailers.
Workers in manufacturing settings often work in close proximity to one another on assembly lines or in production areas. Face masks are used to protect workers from the spread of respiratory droplets and other airborne particles.
Retail workers, including cashiers and sales associates, use face masks to protect themselves and customers from the spread of infectious diseases. This is particularly important in situations where social distancing is not possible, such as in crowded stores or during busy shopping periods.
For hospitality workers, workers in the hospitality industry, including hotel staff, restaurant servers, and bartenders, use face masks to protect themselves and patrons from the spread of infectious diseases. This is particularly important in situations where social distancing is not possible, such as in dining areas or at bars.
Transportation workers, including bus drivers, taxi drivers, and airline staff, use face masks to protect themselves and passengers from the spread of infectious diseases. This is especially important in situations where social distancing is not possible, such as in crowded buses or airplanes.
Educational workers, including teachers, school staff, and other workers in educational settings, use face masks to protect themselves and students from the spread of infectious diseases. This is particularly important in situations where social distancing is not possible, such as in classrooms or during school events.
Public safety workers, including police officers, firefighters, and other public safety workers, use face masks to protect themselves and the public from the spread of infectious diseases. This is particularly important in situations where close contact with others is likely, such as during emergency responses or public events.

1.2. Contribution

Face detection systems face a number of challenges when it comes to detecting faces that are partially or fully covered by face masks. Here are some of the key challenges that face detection systems encounter when faced with masked faces:
  • Reduced accuracy: Face detection systems use machine learning algorithms to identify and track faces. When a face is partially or fully covered by a mask, the system may not be able to accurately identify the face, leading to reduced accuracy in the detection process.
  • False positives and negatives: In some cases, face detection systems may mistake a mask for a face or may fail to detect a face that is partially covered by a mask. This can lead to false positives or negatives, which can compromise the accuracy of the system.
  • Need for retraining: Face detection systems that were trained on unmasked faces may not perform well when faced with masked faces. To accurately detect faces that are partially or fully covered by masks, the system may need to be retrained using data that includes masked faces. This can be time-consuming and resource intensive.
  • Dependence on other features: In some cases, face detection systems may rely on other features, such as the shape of the head or the position of the eyes, to accurately identify faces. When a face is partially or fully covered by a mask, these other features may not be visible, making it more difficult for the system to accurately identify the face.
  • Privacy concerns: The use of face detection systems, particularly in public spaces, can raise privacy concerns. When individuals are wearing masks, the system may not be able to accurately identify them, which can lead to concerns about government surveillance and tracking.
  • Limitations in low-light conditions: Face detection systems may also be limited in low-light conditions, particularly when individuals are wearing masks. This can make it difficult for the system to accurately detect and track faces, leading to reduced performance and increased false positives.
  • Adaptation to different mask types: There are many different types of masks available, including surgical masks, cloth masks, and N95 masks. Each type of mask may present different challenges for face detection systems, which may need to be adapted to detect faces that are partially or fully covered by different types of masks.
  • Impact on system performance: When a face detection system is faced with many masked faces, the performance of the system may be impacted. This can lead to slower processing times, increased false positives, and reduced accuracy.
To address this issue, an integrated approach is proposed, combining the following: (1) Two pretrained deep-learning-based algorithms; one is based on Single Shot Multibox Detector (SSD)-MobileNetV2 for mask detection and the other is based on MediaPipe model for landmark and oval face detection; (2) Robust Principal Component Analysis (RPCA) for robust occlusion and accuracy in identifying faces with masks face recognition; (3) the K-Nearest Neighbors (KNN) classifier for face recognition; (4) the Particle Swarm Optimization (PSO) algorithm to select the features used for training the KNN classifier and optimize the k number.

1.3. Paper Organization

The paper is organized as follows: In Section 2, the literature of existing studies is presented. Section 3 introduces our proposed hybrid method in detail. Section 4 presents the experimental results and discussion. Section 5 concludes this paper. Finally, Section 6 points out the future research ideas.

3. Methodology

This section presents the methodology employed in this paper, which proposes a novel approach to masked face recognition.
The proposed method integrates deep-learning-based mask detection, landmark and oval face detection, and robust principal component analysis (RPCA) to achieve accurate recognition of masked faces. The following steps outline the methodology:
  • Mask Detection: A pretrained ssd-MobileNetV2 model is utilized to detect the presence and location of masks on a face. This deep-learning-based approach effectively identifies whether a person is wearing a mask or not.
  • Landmark and Oval Face Detection: Landmark and oval face detection techniques are employed to identify key facial features. This step helps in precisely localizing facial landmarks and obtaining a better representation of the face.
  • Robust Principal Component Analysis (RPCA): RPCA is utilized to separate occluded and non-occluded components of an image. By extracting the non-occluded components, the proposed method becomes more reliable in identifying faces with masks, as it focuses on the visible facial features.
  • Optimization using Particle Swarm Optimization (PSO): To enhance the performance of the proposed method, particle swarm optimization (PSO) is employed. PSO optimizes both the KNN features and the number of k (nearest neighbors) for KNN, resulting in improved accuracy and robustness.
Pseudocode of our proposed method are mentioned below:
Step 1: Input Masked Face Image.
Step 2: Run Pretrained Mask Detection and Landmark Detection.
   2.1: Mask Segmentation and Conversion to Black Pixels.
Step 3: Prepare, input data set and set the parameter.
Step 4: Normalize the Data.
Step 5: Initialize arrays for feature vectors and labels.
Step 6: Use Robust PCA for dimension reduction.
Step 7: Extract features using LBP.
   7.1: Convert binary feature selection vector to logical indexing vector.
Step 8: Split the data into training and testing sets.
Step 9: Use PSO in feature selection.
   9.1: Define the fitness function for PSO-KNN.
Step 10: Use PSO in optimizing the number of k.
Step 11: Use PSO-KNN for face recognition with optimized k value.
Step 12: Evaluate the performance of the PSO-KNN classifier.
   12.1: Train KNN classifier using training data and optimized k.
   12.2: Test the classifier.
We will explain every step of our proposed methods in the following subsections in detail. Figure 1 shows our proposed pipeline and Figure 2. Shows an example of mask separation, mask face image, reference image, separated mask, and separated face.
Figure 1. The Flowchart of the Proposal Method.
Figure 2. Example of mask separation. (a) Mask face image. (b) Reference image. (c) Separated mask. (d) Separated face.
The distance between a camera and an object can be calculated using various methods, depending on the type of camera, the environment, and the object being measured. Here are some of the most common methods:
  • Triangulation: This method involves using two cameras to measure the distance to an object. The cameras are placed at a known distance from each other and are pointed at the object. By comparing the position of the object in the two camera images, the distance can be calculated using triangulation.
  • Time-of-flight: This method involves emitting a signal, such as a laser or infrared light, from the camera towards the object and measuring the time it takes for the signal to bounce back. The distance can be calculated by multiplying the time of flight by the speed of light.
  • Stereo vision: This method uses two cameras placed a known distance apart to create a 3D image of the scene. By comparing the difference in position of an object in the two camera images, the distance to the object can be calculated.
  • Structured light: This method involves projecting a pattern of light onto the object and measuring the distortion of the pattern caused by the object’s surface. The distance can be calculated by analyzing the distortion of the pattern.
  • Focus: This method involves adjusting the focus of the camera lens until the object is in sharp focus. The distance can be calculated by using the known focal length of the lens and the distance from the lens to the image sensor.
We used the focus technique. The accuracy of these methods can vary depending on the environment, lighting conditions, and the size and shape of the object being measured.

3.1. Deep-Learning-Based Mask Detection and Face Oval Detection

The technique of recognizing distinct features on a person’s face, such as the corners of the eyes, the tip of the nose, and the borders of the lips, is known as facial landmark detection. Machine learning techniques are frequently used by facial landmark identification algorithms to recognize these landmarks. In our work, MediaPipe framework [40] is used for detecting the landmarks of the face images. By using MediaPipe framework, we can detect the face oval which is the outline of the face constructed by connecting the outer face landmarks as shown in Figure 3. We use a pretrained deep-learning-based mask detection method for detecting masks. This pretrained model is based on SSD-MobileNetV2. SSD-MobileNetV2 is a widely used object detection model that is specifically designed to operate quickly and accurately on devices with limited computational power, such as smartphones. It employs a combination of a base network called MobileNetV2 and a detection layer known as SSD (Single Shot Detector) to predict the bounding boxes and class labels of objects within an image.
Figure 3. (a) Face Landmarks Detection and (b) Face Oval Detection.
The reason why SSD-MobileNetV2 is chosen for object detection tasks is due to its numerous advantages over other models, such as the following:
  • It is lightweight and efficient, meaning it has fewer parameters and operations than other models, which reduces memory and power consumption. It can achieve real-time inference (30 frames per second) even on mobile devices.
  • It employs depthwise separable convolutions, which are a type of convolution that divides the standard convolution into two steps: a depthwise convolution that applies a single filter to each input channel, and a pointwise convolution that combines the outputs of the depthwise convolution. This reduces the number of computations and parameters by a factor of 8 to 9.
  • It utilizes inverted residual blocks, which are a type of residual block that have thin bottleneck layers at the input and output, and a thick expansion layer in the middle. This allows the network to learn more complex features with fewer parameters and computations.
  • It is one-stage, meaning it directly outputs the bounding boxes and class labels without using any intermediate steps, such as region proposals or feature pyramids. This makes it faster and simpler than two-stage models, such as Faster R-CNN or Mask R-CNN.

3.2. Robust Principal Component Analysis

One of the well-known dimension reduction techniques is the principal component analysis (PCA) [41]. In computer vision, it is used to represent an image with a relatively small dimensional feature vector. However, PCA is fragile with respect to outliers. For tackling this drawback, Candès et al. developed a statistical method called Robust PCA (RPCA) [42], where RPCA is used for decomposing data into its principal components (i.e., the underlying structure of the data), while also identifying and removing outliers and noise. This makes RPCA a powerful tool for a variety of applications, including image and video processing, signal processing, and machine learning. It is often used in situations where the data contain noise or outliers that can distort the results of traditional PCA. In our case, the outliers are the mask pixels that occlude the lower part of the face images.
Before using RPCA, the masks that occlude the face images are segmented by extracting the intersection region between the bounding box surrounding the masks and the face ovals detected using mediapipe library. Then, the segmented masks are converted to black pixels to not disturb the RPCA, where we investigate that the colored masks are not classified well as outliers. After that, RPCA is used for obtaining the low-rank matrix of face features. Figure 4 shows an example of the RPCA technique.
Figure 4. RPCA Technique.
In our algorithm, RPCA assumes that the matrix consisting of face images features vectors (denoted as X) is a combination of a low-rank component which represents the eigen faces and a sparse component that contain the occlusion pixels. The RPCA method aims to factorize an input matrix X into the sum of a low-rank matrix L and a sparse matrix S such that X = L + S. This can be formulated as the following optimization problem,
m i n L , S   rank   L + S 0   subject   to   L + S = X ,
where   . 0   denotes the L0 norm. We can find the best L and S with a high probability by using a simpler way called convex relaxation, where the rank is relaxed to nuclear norm and the L0 norm is relaxed to the L1 norm. After convex relaxation, the equation becomes as follows,
m i n L , S   L + λ S 1   subject   to   L + S = X
where   .   denotes the nuclear norm and   . 1   denote the L1 norm. There are several algorithms that can be used to perform RPCA, including the Principal Component Pursuit (PCP) algorithm and the Alternating Direction Method of Multipliers (ADMM) algorithm [42]. These algorithms solve the optimization problem required for RPCA and can efficiently and effectively separate the low-rank and sparse components. After obtaining the low-rank matrix L which represents the eigen faces without holes or black pixels, KNN is then used for face recognition.

3.3. K-Nearest Neighborhood Algorithm and Particle Swarm Optimization (PSO) for Face Recognition

People can be recognized by looking at certain things in their faces, like the color and texture. Texture is especially important for biometric face recognition because it helps the computer recognizes patterns in the face image. Researchers use a well-known method called the local binary pattern (LBP) to extract the most important features of the face image. LBP is effective because it works even if the face image looks different under conditions such as lighting, the person’s facial expressions, or pose. Finding these important parts of the face is very important for computers to recognize people’s faces accurately.
Feature selection is an important step for a wide range of machine learning approaches and computer vision tasks, including facial recognition. Feature selection aims to identify the most relevant and informative features in the data, while eliminating the redundant and irrelevant ones that may negatively affect the classifier performance. This can help to reduce the dimensionality of the feature space, which can in turn improve the accuracy and efficiency of recognition algorithms [43]. Moreover, the KNN classifier is a facial recognition method that is both easy to use and efficient computationally. However, the initialization of the parameter k causes it to suffer. Therefore, k is also required to optimize.
To select the features used for training the KNN classifier and optimize the k number, evolutionary algorithms such as Particle Swarm Optimization (PSO) and Genetic Algorithm (GA) are used because they are global metaheuristic optimization techniques, where they are not affected by the issue of local minima. Based on the results of [7], PSO is the best choice used for optimum features selection and k selection.

4. Experimental Settings and Results

4.1. Experimental Settings

This subsection mentions the setting for the experiments. Two data sets—simulated and real masked face data sets—are used for testing the performance of different masked face recognition methods. The simulated one is the Labeled Face in the Wild Stimulated Masked Face Data Set (LFW-SMFD) [44]. The data set consists of 13,117 faces of 5713 people. Sample images from simulated masked and non-masked faces images are shown in Figure 5.
Figure 5. Pairs of simulated masked and non-masked face images.
For the real data set, we chose to use actual photographs of people wearing masks rather than computer-generated masked face images to provide a more realistic testing environment. We searched Google for masked pictures of well-known people, such as politicians and celebrities, to build our database. The photographs we chose were carefully chosen to be of the highest caliber and to be free of duplication. Sample images from real masked and non-masked faces images are shown in Figure 6.
Figure 6. Pairs of real masked and non-masked face images.

4.2. Experimental Results and Discussion

In this section, the performance of our proposed algorithm will be evaluated compared by other methods such as the method of Ejaz et al. [45] which uses the PCA technique, Latent Part Detection method [11], and Rodriguez et al.’s [46] method that uses the mixture of Gaussian method.
The experimental results were obtained using a computer system running Windows 11 Home as its operating system. The processor used in this system is the AMD Ryzen™ 7 6800H, which has a maximum boosted frequency of up to 4.7 GHz, 3 MB L16 cache, 8 cores, and 16 series processing. The system utilizes AMD’s on-chip system technology for its chipset. The graphics cards are separate and consist of an NVIDIA® GeForce RTX™ 3070 Ti Laptop GPU with 6 GB of dedicated GDDR8 memory. The system is equipped with 16 GB of DDR4800-16 MHz RAM (2 × 8 GB) with transfer rates of up to 4800 MB/s. Additionally, the system has a Gen4 SSD with 1 TB of storage. For video conferencing and recording, the system includes an HP Wide Vision 720p HD camera with temporary noise reduction and integrated dual-array digital microphones.
Table 2 and Figure 7 shows the performance of our proposed methods compared to the other methods in terms of accuracy in recognition of masked images. From Table 1, it is clearly shown that our proposed method achieves accuracy 97% for the masked data set, which outperforms the compared methods. Moreover, it is evident that the performance of PCA alone [45] is the worst one compared with the other results.
Table 2. Performance of our proposed methods compared to the other methods in terms of accuracy in masked face recognition.
Figure 7. Analysis of the performance of the accuracy of different algorithms in masked face recognition.
Table 3 and Figure 8 show the performance of our proposed methods compared to the other methods in terms of accuracy in recognition of unmasked images. From Table 2, it is clearly shown that our proposed method achieves accuracy 98.4% for the unmasked data set, which outperforms the compared methods. Moreover, it is evident that the performance of PCA alone [25] is also the worst one compared with the other results.
Table 3. Performance of our proposed methods compared to the other methods in terms of accuracy in unmasked face image.
Figure 8. Analysis of the performance of the accuracy of different algorithms in unmasked face recognition.
For the KNN part, the features are obtained from LBP after applying the RPCA then the KNN classifier optimized by PSO is used and it is denoted as PSO-KNN. The parameters PSO algorithm is reported in Table 4. The effectiveness of the PSO-KNN algorithm has been evaluated in comparison to conventional benchmark classifiers.
Table 4. PSO parameters.
Table 5 describes the accuracy of the proposed KNN optimized by PSO compared with unoptimized k number of KNN in the cases of using different optimization techniques such as PSO and GA for feature extraction.
Table 5. Comparative analysis of accuracy of proposed method with existing benchmark algorithms in case of unmasked data set.
The results show that PSO-KNN outperforms KNN in all scenarios, achieving a higher classification accuracy. For example, when using actual features, KNN achieved an accuracy of 89%, while PSO-KNN achieved an accuracy of 93%. When using GA for feature selection, KNN achieved an accuracy of 94%, while PSO-KNN achieved an accuracy of 96%. When using PSO for feature selection, KNN achieved an accuracy of 95%, while PSO-KNN achieved an accuracy of 98%, which is the highest accuracy among all methods.
The results demonstrate that PSO-KNN is a more effective algorithm for masked face recognition, achieving higher accuracy compared to KNN, especially when PSO is used for feature selection and optimizing the value of k.

4.3. Execution Time Results

The SSD-Mobilenet is a real-time object detection model that is designed to run at a high frame rate of 46 frames per second. This means that it takes approximately 0.02 s to process each frame and detect objects within it. The model achieves this high performance by combining the Single Shot Detector (SSD) algorithm with the efficient MobileNet architecture.
In contrast, the RPCA and KNA algorithms have a lower frame rate of 4 frames per second and are not considered to be real-time. These algorithms take approximately 0.3 s to process each frame, resulting in a total processing time of 0.32 s. While these algorithms may be effective for certain tasks, they are not suitable for applications that require real-time object detection.
The SSD-Mobilenet model is a highly efficient and effective real-time object detection solution that can process frames at a much faster rate than other algorithms such as RPCA and KNA. Its ability to accurately detect objects in real-time makes it a valuable tool for a wide range of applications, from self-driving cars to security systems and beyond.

5. Discussion

The proposed method of combining mask detection, facial landmark detection, oval face detection, and RPCA represents a promising approach for achieving higher accuracy in masked face recognition. By detecting and localizing the mask and key facial features, the method is able to extract relevant non-occluded facial components for recognition. The use of RPCA to separate occluded and non-occluded components helps improve occlusion robustness. Finally, optimizing the KNN features and parameters using PSO further enhances performance. This is evidenced by the achieved recognition rate of 97%, outperforming existing methods. The proposed method thus demonstrates significant advancements for addressing the challenges of masked face recognition. We can summarize our contribution discussion as follows:
  • The combination of deep learning and traditional computer vision techniques achieves state-of-the-art performance.
  • The mask detection and facial feature localization improve robustness to occlusion.
  • The use of RPCA to separate occluded and non-occluded features is effective for masked face recognition.
  • Optimizing the KNN parameters using PSO further boosts the accuracy.
  • The 97% recognition rate significantly outperforms existing methods.
  • The proposed method represents an important advancement for masked face recognition.
  • The approach is potentially applicable to real-world scenarios where face masks are common.
  • The methodology can be extended to other types of facial occlusion.
  • The techniques are generalizable and could be applied to other computer vision tasks.
  • The approach provides a foundation for future research on masked face recognition.

6. Limitations

While the proposed method shows high accuracy and robustness, there are several limitations that could be addressed in future work.
  • The method relies on a data set of individuals wearing different types of masks. The performance may degrade for novel mask types not present in the data set. Collecting a more comprehensive masked face data set could help address this.
  • The accuracy may be sensitive to mask placement, facial poses, and lighting conditions. Collecting a more diverse data set that covers more mask variations, poses, and conditions could help improve robustness.
  • The method has only been evaluated on still images and may not generalize well to video. Extending the method to video-based masked face recognition could enable real-world applications.
  • The computational cost of combining multiple deep learning and computer vision techniques may be high. Methods for optimizing efficiency could be explored to enable on-device applications.
  • The ability to handle occlusion from other objects has not been evaluated.

7. Future work

In future work, we can focus on extending our proposed method to handle other types of occlusions, such as sunglasses, scarfs, and hats, which can also pose challenges to face recognition systems. Additionally, integrating other biometric modalities, such as voice and fingerprint, can further enhance the accuracy and reliability of the system. Another direction for future work is to investigate the privacy implications of masked face recognition. As masks have become a ubiquitous part of daily life, concerns about facial recognition technology’s impact on privacy have increased.
Our proposed method relies on face recognition, which raises concerns about the potential for misuse and abuse of the technology. Therefore, future research can focus on developing ethical guidelines and regulations for the use of masked face recognition technology.

8. Conclusions

The use of face masks has become essential in various industries and jobs, necessitating the development of effective masked face recognition technologies. In this paper, we proposed a novel method that combines deep-learning-based mask detection, landmark and oval face detection, and robust principal component analysis (RPCA) for accurate masked face recognition. Our proposed method utilizes pretrained ssd-MobileNetV2 for mask detection and RPCA to separate occluded and non-occluded components of an image. To optimize the performance of our proposed method, we used particle swarm optimization (PSO) to optimize both the KNN features and the number of k for KNN. Experimental results showed that our proposed method outperformed existing methods in terms of accuracy and robustness to occlusion, achieving a recognition rate of 97%, significantly higher than the state-of-the-art methods. Overall, our proposed method represents a significant improvement over existing methods for masked face recognition, providing high accuracy and robustness to occlusion.

Author Contributions

This work was carried out in collaboration among all authors. All Authors designed the study, performed the statistical analysis, and wrote the protocol. Authors M.E., T.M.M., M.M.I. and T.A.E.-H. managed the analyses of the study, managed the literature searches, and wrote the first draft of the manuscript. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Data Availability Statement

The data that support the findings of this study are available in https://www.kaggle.com/datasets/muhammeddalkran/lfw-simulated-masked-face-dataset, accessed on 26 June 2023.

Conflicts of Interest

The authors declare that there is no conflict of interest. The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.

References

  1. Jain, A.; Bolle, R.; Pankanti, S. Introduction to Biometrics; Springer: Boston, MA, USA, 1996. [Google Scholar]
  2. Hariri, W. Efficient masked face recognition method during the COVID-19 pandemic. Signal Image Video Process. 2022, 16, 605–612. [Google Scholar] [CrossRef]
  3. Farkhod, A.; Abdusalomov, A.B.; Mukhiddinov, M.; Cho, Y.-I. Development of Real-Time Landmark-Based Emotion Recognition CNN for Masked Faces. Sensors 2022, 22, 8704. [Google Scholar] [CrossRef]
  4. Hemathilaka, S.; Aponso, A. A comprehensive study on occlusion invariant face recognition under face mask occlusion. arXiv 2022, arXiv:2201.09089. [Google Scholar] [CrossRef]
  5. Jeevan, G.; Zacharias, G.C.; Nair, M.S.; Rajan, J. An empirical study of the impact of masks on face recognition. Pattern Recognit. 2022, 122, 108308. [Google Scholar] [CrossRef]
  6. Neto, P.C.; Pinto, J.R.; Boutros, F.; Damer, N.; Sequeira, A.F.; Cardoso, J.S. Beyond masks: On the generalization of masked face recognition models to occluded face recognition. IEEE Access 2022, 10, 86222–86233. [Google Scholar] [CrossRef]
  7. Sasirekha, K.; Thangavel, K. Optimization of K-nearest neighbor using particle swarm optimization for face recognition. Neural Comput. Appl. 2019, 31, 7935–7944. [Google Scholar]
  8. Zhao, W.; Zhu, X.; Shi, H.; Zhang, X.-Y.; Lei, Z. Consistent Sub-Decision Network for Low-Quality Masked Face Recognition. IEEE Signal Process. Lett. 2022, 29, 1147–1151. [Google Scholar] [CrossRef]
  9. Koklu, M.; Cinar, I.; Taspinar, Y.S. CNN-based bi-directional and directional long-short term memory network for determination of face mask. Biomed. Signal Process. Control 2022, 71, 103216. [Google Scholar] [CrossRef]
  10. Adhikarla, E.; Davison, B.D. Face mask detection on real-world Webcam images. In Proceedings of the Conference on Information Technology for Social Good, Rome, Italy, 9–11 September 2021; pp. 139–144. [Google Scholar]
  11. Ding, F.; Peng, P.; Huang, Y.; Geng, M.; Tian, Y. Masked face recognition with latent part detection. In Proceedings of the 28th ACM International Conference on Multimedia, Seattle, WA, USA, 12–16 October 2020; pp. 2281–2289. [Google Scholar]
  12. Vu, H.N.; Nguyen, M.H.; Pham, C. Masked face recognition with convolutional neural networks and local binary patterns. Appl. Intell. 2022, 52, 5497–5512. [Google Scholar] [CrossRef]
  13. Aswal, V.; Tupe, O.; Shaikh, S.; Charniya, N.N. Single camera masked face identification. In Proceedings of the 2020 19th IEEE International Conference on Machine Learning and Applications (ICMLA), Miami, FL, USA, 14–17 December 2020; pp. 57–60. [Google Scholar]
  14. Kocacinar, B.; Tas, B.; Akbulut, F.P.; Catal, C.; Mishra, D. A real-time cnn-based lightweight mobile masked face recognition system. IEEE Access 2022, 10, 63496–63507. [Google Scholar] [CrossRef]
  15. Qi, C.; Yang, L. Face recognition in the scene of wearing a mask. In Proceedings of the 2020 International Conference on Advance in Ambient Computing and Intelligence (ICAACI), Ottawa, ON, Canada, 12–13 September 2020; pp. 77–80. [Google Scholar]
  16. Singh, S.; Ahuja, U.; Kumar, M.; Kumar, K.; Sachdeva, M. Face mask detection using YOLOv3 and faster R-CNN models: COVID-19 environment. Multimed. Tools Appl. 2021, 80, 19753–19768. [Google Scholar] [CrossRef]
  17. Zhu, R.; Yin, K.; Xiong, H.; Tang, H.; Yin, G. Masked face detection algorithm in the dense crowd based on federated learning. Wirel. Commun. Mob. Comput. 2021, 2021, 8586016. [Google Scholar]
  18. Jignesh Chowdary, G.; Punn, N.S.; Sonbhadra, S.K.; Agarwal, S. Face mask detection using transfer learning of inceptionv3. In Proceedings of the Big Data Analytics: 8th International Conference, BDA 2020, Sonepat, India, 15–18 December 2020; Proceedings 8. pp. 81–90. [Google Scholar]
  19. Loey, M.; Manogaran, G.; Taha, M.H.N.; Khalifa, N.E.M. A hybrid deep transfer learning model with machine learning methods for face mask detection in the era of the COVID-19 pandemic. Measurement 2021, 167, 108288. [Google Scholar]
  20. Suresh, K.; Palangappa, M.; Bhuvan, S. Face mask detection by using optimistic convolutional neural network. In Proceedings of the 2021 6th International Conference on Inventive Computation Technologies (ICICT), Coimbatore, India, 20–22 January 2021; pp. 1084–1089. [Google Scholar]
  21. Lodh, A.; Saxena, U.; Motwani, A.; Shakkeera, L.; Sharmasth, V.Y. Prototype for integration of face mask detection and person identification model—COVID-19. In Proceedings of the 2020 4th International Conference on Electronics, Communication and Aerospace Technology (ICECA), Coimbatore, India, 5–7 November 2020; pp. 1361–1367. [Google Scholar]
  22. Rusli, M.H.; Sjarif, N.N.A.; Yuhaniz, S.S.; Kok, S.; Kadir, M.S. Evaluating the masked and unmasked face with LeNet algorithm. In Proceedings of the 2021 IEEE 17th International Colloquium on Signal Processing & Its Applications (CSPA), Langkawi, Malaysia, 5–6 March 2021; pp. 171–176. [Google Scholar]
  23. Saleh, N.A.; Ertunç, H.M.; Saleh, R.A.; Rassam, M.A. A simple mask detection model based on a multi-layer perception neural network. In Proceedings of the 2021 International Conference of Technology, Science and Administration (ICTSA), Taiz, Yemen, 22–24 March 2021; pp. 1–5. [Google Scholar]
  24. Wu, G. Masked face recognition algorithm for a contactless distribution cabinet. Math. Probl. Eng. 2021, 2021, 5591020. [Google Scholar]
  25. Ullah, N.; Javed, A.; Ghazanfar, M.A.; Alsufyani, A.; Bourouis, S. A novel DeepMaskNet model for face mask detection and masked facial recognition. J. King Saud Univ.-Comput. Inf. Sci. 2022, 34, 9905–9914. [Google Scholar] [CrossRef]
  26. Habib, S.; Alsanea, M.; Aloraini, M.; Al-Rawashdeh, H.S.; Islam, M.; Khan, S. An Efficient and Effective Deep Learning-Based Model for Real-Time Face Mask Detection. Sensors 2022, 22, 2602. [Google Scholar]
  27. Farman, H.; Khan, T.; Khan, Z.; Habib, S.; Islam, M.; Ammar, A. Real-time face mask detection to ensure COVID-19 precautionary measures in the developing countries. Appl. Sci. 2022, 12, 3879. [Google Scholar]
  28. Freud, E.; Di Giammarino, D.; Stajduhar, A.; Rosenbaum, R.S.; Avidan, G.; Ganel, T. Recognition of masked faces in the era of the pandemic: No improvement despite extensive natural exposure. Psychol. Sci. 2022, 33, 1635–1650. [Google Scholar] [CrossRef]
  29. Song, Z.; Nguyen, K.; Nguyen, T.; Cho, C.; Gao, J. Spartan Face Mask Detection and Facial Recognition System. Healthcare 2022, 10, 87. [Google Scholar] [CrossRef]
  30. Kwak, N.; Kim, D. Study On Masked Face Detection And Recognition using transfer learning. Int. J. Adv. Cult. Technol. (IJACT) 2022, 10, 294–301. [Google Scholar]
  31. Peng, H.; Xing, Z.; Liu, X.; Gao, Z.; He, H. Toward masked face recognition: An effective facial feature extraction and refinement model in multiple scenes. Expert Syst. 2023, 40, e13166. [Google Scholar]
  32. Ge, Y.; Liu, H.; Du, J.; Li, Z.; Wei, Y. Masked face recognition with convolutional visual self-attention network. Neurocomputing 2023, 518, 496–506. [Google Scholar] [CrossRef] [PubMed]
  33. Golwalkar, R.; Mehendale, N. Masked-face recognition using deep metric learning and FaceMaskNet-21. Appl. Intell. 2022, 52, 13268–13279. [Google Scholar] [CrossRef]
  34. Yuan, G.; Zheng, H.; Dong, J. MSML: Enhancing Occlusion-Robustness by Multi-Scale Segmentation-Based Mask Learning for Face Recognition. In Proceedings of the AAAI Conference on Artificial Intelligence, Virtual, 22 February–1 March 2022; pp. 3197–3205. [Google Scholar]
  35. Huang, B.; Wang, Z.; Wang, G.; Jiang, K.; Han, Z.; Lu, T.; Liang, C. PLFace: Progressive learning for face recognition with mask bias. Pattern Recognit. 2023, 135, 109142. [Google Scholar] [CrossRef]
  36. Fang, M.; Damer, N.; Kirchbuchner, F.; Kuijper, A. Real masks and spoof faces: On the masked face presentation attack detection. Pattern Recognit. 2022, 123, 108398. [Google Scholar] [CrossRef]
  37. Shatnawi, M.; Almenhali, N.; Alhammadi, M.; Alhanaee, K. Deep learning approach for masked face identification. Int. J. Adv. Comput. Sci. Appl. 2022, 13, 296–305. [Google Scholar] [CrossRef]
  38. Pann, V.; Lee, H.J. Effective attention-based mechanism for masked face recognition. Appl. Sci. 2022, 12, 5590. [Google Scholar]
  39. Oumina, A.; El Makhfi, N.; Hamdi, M. Control the COVID-19 pandemic: Face mask detection using transfer learning. In Proceedings of the 2020 IEEE 2nd International Conference on Electronics, Control, Optimization and Computer Science (ICECOCS), Kenitra, Morocco, 2–3 December 2020; pp. 1–5. [Google Scholar]
  40. Lugaresi, C.; Tang, J.; Nash, H.; McClanahan, C.; Uboweja, E.; Hays, M.; Zhang, F.; Chang, C.-L.; Yong, M.G.; Lee, J. Mediapipe: A framework for building perception pipelines. arXiv 2019, arXiv:1906.08172. [Google Scholar]
  41. Abdi, H.; Williams, L.J. Principal component analysis. Wiley Interdiscip. Rev. Comput. Stat. 2010, 2, 433–459. [Google Scholar]
  42. Candès, E.J.; Li, X.; Ma, Y.; Wright, J. Robust principal component analysis? J. ACM (JACM) 2011, 58, 11. [Google Scholar] [CrossRef]
  43. Tubishat, M.; Alswaitti, M.; Mirjalili, S.; Al-Garadi, M.A.; Rana, T.A. Dynamic butterfly optimization algorithm for feature selection. IEEE Access 2020, 8, 194303–194314. [Google Scholar] [CrossRef]
  44. Dalkiran, M. LFW Simulated Masked Face Dataset; Kaggle: San Francisco, CA, USA, 2020; pp. 1–8. [Google Scholar]
  45. Ejaz, M.S.; Islam, M.R.; Sifatullah, M.; Sarker, A. Implementation of principal component analysis on masked and non-masked face recognition. In Proceedings of the 2019 1st International Conference on Advances in Science, Engineering and Robotics Technology (ICASERT), Dhaka, Bangladesh, 3–5 May 2019; pp. 1–5. [Google Scholar]
  46. Nieto-Rodriguez, A.; Mucientes, M.; Brea, V.M. System for medical mask detection in the operating room through facial attributes. In Proceedings of the Pattern Recognition and Image Analysis: 7th Iberian Conference, IbPRIA 2015, Santiago de Compostela, Spain, 17–19 June 2015; Proceedings 7. pp. 138–145. [Google Scholar]
  47. Samaria, F.S.; Harter, A.C. Parameterisation of a stochastic model for human face identification. In Proceedings of the 1994 IEEE Workshop on Applications of Computer Vision, Sarasota, FL, USA, 5–7 December 1994; pp. 138–142. [Google Scholar]
  48. Sikha, O.; Bharath, B. VGG16-random Fourier hybrid model for masked face recognition. Soft Comput. 2022, 26, 12795–12810. [Google Scholar] [CrossRef]
  49. Li, Y.; Guo, K.; Lu, Y.; Liu, L. Cropping and attention based approach for masked face recognition. Appl. Intell. 2021, 51, 3012–3025. [Google Scholar] [CrossRef]
  50. Yi, D.; Lei, Z.; Liao, S.; Li, S.Z. Learning face representation from scratch. arXiv 2014, arXiv:1411.7923. [Google Scholar]
  51. Lane, L. Nist finds flaws in facial checks on people with covid masks. Biom. Technol. Today 2020, 2020, 2. [Google Scholar]
  52. Taigman, Y.; Yang, M.; Ranzato, M.A.; Wolf, L. Deepface: Closing the gap to human-level performance in face verification. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 1701–1708. [Google Scholar]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.