Next Article in Journal
Development of A Micro-CT Scanner with Dual-Energy Option and Endovascular Contrast Agent Administration Protocol for Fetal and Neonatal Virtual Autopsy
Previous Article in Journal
An Improved Path-Finding Method for the Tracking of Centerlines of Tortuous Internal Carotid Arteries in MR Angiography
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Privacy-Preserving Face Recognition Method Based on Randomization and Local Feature Learning

School of Cyberspace, Hangzhou Dianzi University, Hangzhou 310018, China
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
J. Imaging 2024, 10(3), 59; https://doi.org/10.3390/jimaging10030059
Submission received: 31 January 2024 / Revised: 26 February 2024 / Accepted: 27 February 2024 / Published: 28 February 2024
(This article belongs to the Section Biometrics, Forensics, and Security)

Abstract

:
Personal privacy protection has been extensively investigated. The privacy protection of face recognition applications combines face privacy protection with face recognition. Traditional face privacy-protection methods encrypt or perturb facial images for protection. However, the original facial images or parameters need to be restored during recognition. In this paper, it is found that faces can still be recognized correctly when only some of the high-order and local feature information from faces is retained, while the rest of the information is fuzzed. Based on this, a privacy-preserving face recognition method combining random convolution and self-learning batch normalization is proposed. This method generates a privacy-preserved scrambled facial image and an image fuzzy degree that is close to an encryption of the image. The server directly recognizes the scrambled facial image, and the recognition accuracy is equivalent to that of the normal facial image. The method ensures the revocability and irreversibility of the privacy preserving of faces at the same time. In this experiment, the proposed method is tested on the LFW, Celeba, and self-collected face datasets. On the three datasets, the proposed method outperforms the existing face privacy-preserving recognition methods in terms of face visual information elimination and recognition accuracy. The recognition accuracy is >99%, and the visual information elimination is close to an encryption effect.

1. Introduction

Due to the rapid development of technology and the use of large datasets, face recognition technology has been applied in various fields. However, the widespread use of face recognition systems brings some challenges, from which the issue of face privacy protection has been widely mentioned. Face recognition is closely related to property and personal safety. However, in order to rapidly promote face recognition technology, face databases without privacy protection are widely collected, which brings serious personal privacy security risks.
In a face recognition system, facial images are transferred to a server using an acquisition device and are compared with the stored facial feature database on the server. Servers mostly store clear and complete faces. Some PPTs (privacy-protection technologies) employ encryption and decryption methods, wherein the server-based facial feature database and the facial images are encrypted [1]. However, decrypting the incoming and the server-based facial images is necessary during recognition. Therefore, server-based databases encounter potential privacy breaches. Some PPTs use feature extraction, reversible transformation [2], or other technologies to recognize face images after deformation, but the recognition accuracy is often not ideal due to the influence of the original image deformation. In addition, there are some PPTs that can protect face privacy in public databases, such as deep neural networks that are used to remove soft biometric information [3] such as gender, age, identity, etc. Although face images processed by such methods can achieve a certain degree of privacy protection, they cannot be used directly for face recognition.
A clear and complete face image is not necessary in face recognition. Only parts of the high-order and local features of the face are retained, which can also be used for the purpose of achieving face recognition. For example, eigenfaces can be recognized by using PCA (principal component analysis) to propose partial features of the face. However, the eigenface recognition accuracy proposed by PCA is not high, the accuracy is less than 99%, and the privacy-protection effect is not ideal. On this basis, this paper proposes a privacy-preserving face recognition method based on randomization and local feature learning (RLLFPR), which combines privacy protection and recognition to create a fuzzy face to protect privacy. The algorithm uses a deep neural network with random convolution and self-learning batch normalization layers and combines a self-learning update parameter network with a loss-function backpropagation update parameter network. In this process, the fuzziness, revocability, and irreversibility of the privacy preserving of faces are guaranteed. The network randomizes the pixel values in the facial image and maintains the light-and-dark relationship between the pixel values in the region so that the biometric characteristics of the generated facial image cannot be recognized by human vision; this process is irreversible and can only be recognized by the deep neural network model with specific parameters. At the same time, the deep neural network model cannot infer the original face. The contributions of this paper are as follows:
  • We deeply study the learning of hidden human visual information in the end-to-end face image feature learning of deep neural networks and propose that preserving the light-and-dark relationship between facial image pixels and randomizing other information can eliminate human visual information while maintaining the recognizability of facial images. According to the results of our review, the proposed method is the most thorough method to eliminate human visual information in the current privacy-preserving face recognition methods. Using this technology can make a facial image have better privacy-protection ability.
  • A deep neural network framework and RLLFPR method for face privacy protection are proposed. Different from traditional encryption and decryption methods, the proposed framework combines face privacy protection with face recognition optimization, which can jointly compute face privacy protection and recognition.
  • The RLLFPR method produces privacy-preserved faces with fuzziness, revocability, and irreversibility for better privacy protection. All the information stored in the face recognition server, private face recognition model, private facial image, etc., cannot restore or deduce the original face, which improves the privacy protection of facial images in face recognition or authentication systems.

2. Related Work

2.1. Face Recognition

Deep neural networks perform well in recognition and classification tasks [4]. The deep neural network feature vector can effectively close and open the intraclass distances during face recognition. Schroff [5] proposed the FaceNet algorithm and employed a triple-loss method to design the loss functions. FaceNet expanded the class distances more than the traditional Softmax method [6]. However, triplet-loss training was challenging on large datasets. Wen [7] proposed the center-loss method to enhance Softmax’s recognition ability. It is hoped that the sum of the distance between all the image features of a class and the center features of that class is minimized. Liu proposed the Sphereface method [8] and introduced angle boundaries into Softmax to improve the face recognition accuracy. The complex-loss-function training led to network training instability. Wang proposed the Cosface method [9] to increase the target’s cosine angle penalty. Cosface exhibited better performance and was easier to implement than Sphereface. Deng proposed the ArcFace method [10], with an angular margin penalty added to the cosine angle. The ArcFace method exhibited compact class arrangement and higher discrimination, which resulted in further improvements.

2.2. Face Privacy Protection

Face privacy preservation [11,12] is also an important issue arising with the development of deep neural networks and face recognition. Face data are directly related to personal property and are widely valued by researchers. Traditional face privacy protection is achieved by blurring or pixelizing the main area of the face [13], but blurring may lead to the loss of other information except the face information. Recently, deep neural networks have also been widely used in face privacy-protection technology. In order to protect face data in public databases, the use face-swapping technology or face-derecognition technology are good methods, and most of these technologies generate new faces with the help of GAN [14] technology. Face-swapping technology aims to replace the face in the original image with other faces to achieve image privacy protection. Korshunova [15] used a patch-based style loss function and a variety of style images to quickly change faces. Yang [16] realized the face privacy-protection method of face swapping based on a reversible mask network, which generated more realistic face images and had reversibility. Face-swapping technology has a good effect on face privacy protection, but it cannot be applied to face recognition. Face derecognition has been more widely studied than face-swapping technology. People try to remove soft biometric information such as gender and age from face images without changing the original face. Mirjalili [17] proposed PrivacyNet to generate face images that can interfere with race, gender, and age through a GAN model based on consistency loss. Liu [18] proposed an attribute-preserving face derecognition framework, which can discard some privacy attributes while retaining the required facial attributes. Face-derecognition technology applied in the field of face recognition can only play an auxiliary role in face recognition because the deletion of specific soft biometric information will lead to a reduction in accuracy.

2.3. Privacy-Protected Face Recognition

Several privacy-protection technologies exist in face recognition [19]. Recognition by facial encryption and decryption [20] is a fundamental method in this regard. Bai [21] proposed a privacy-preserving face recognition system using homomorphic encryption and multiparty secure computation. Ma [22] proposed a lightweight privacy-preserving adaptive-enhancement (AdaBoost) face recognition framework based on encrypted sharing and edge computing. Boragule [23] stored facial features in smart card memory to realize portable and privacy-preserving recognition. Im [24] proposed the encryption of the feature vectors after deep neural network processing but not of the images. The face recognition privacy was protected by storing the encrypted feature vectors on the server. Zhang [25] proposed a secure and efficient outsourcing protocol for face recognition based on PCA. Through the transformation of the original image information, the information privacy was protected and the resources were saved. Lei [26] proposed a new secure inner product protocol that used a lightweight random mask technique instead of time-consuming public key cryptographic operations to efficiently measure the similarity of facial data. Mai [27] generated reversible random secure sketches through facial feature maps to protect irreversible and deletable templates while maintaining verification performance. The privacy-protection methods of cryptography often need to decrypt and restore the original face for recognition, which has certain security risks. At the same time, the existence of the key is also a burden to the holder. Therefore, these methods are mostly used for face authentication rather than large-scale face recognition. You [13] proposed a reversible privacy-preserving facial expression recognition method in which face information was hidden in mosaics through an adversarial method. However, privacy-preserved images can only complete expression recognition. Walia [28] used the adaptive weighted graph approach to generate multimodal cancelable biometric templates. The multimodal approach increases the complexity of the overall system. Morales [29] used deep neural networks to suppress sensitive information from faces while ensuring correct recognition. Zhang [30] generated perturbed images with hidden attributes while retaining the effect of face verification. Refs. [29,30] decreased the accuracy during face recognition due to the suppression of soft biometric information in the face image. Wu et al. [31,32] proposed the technical idea of using biometric key technology to directly generate a strong biometric key from the biometric characteristics of the client, which means the server does not need to save the biometric template, so as to protect privacy. At present, the stability of the key generated by this technical route needs to be improved. It is difficult to directly blur the original face for face recognition. The existing methods have poor recognition accuracy or privacy-protection effects. By using local differential privacy, Chamikara [33] used differential privacy to apply disturbances to eigenfaces. The privacy-protection effect was better, and the screen was human vision. However, the accuracy was low, only 80%–90%. Zhang [2] achieved the purpose of privacy protection by using an Arnold transformation to process faces and recognize the transformed faces. However, the recognition accuracy of the method needs to be improved. Mi [34] adopted frequency domain collaborative reasoning and proposed DuetFace. The accuracy of the privacy-preserving face recognition was greater than 99%, but the privacy-preserved face retained part of the original face contour. Subsequently, Mi [35] proposed the random selection of frequency components for training and learning, and the recognition accuracy was further improved, but the privacy-protected face still retained the original human part of the information and did not completely shield the image from human vision.

3. Privacy-Preserving Face Recognition Based on Randomization and Local Feature Learning (RLLFPR)

In face recognition, it is not necessary to learn the whole face, only the difference of the pixel values of each organ of the original face so that the face can be accurately recognized. Based on the above observations, this paper proposes the RLLFPR framework and method to recognize faces by randomizing the pixel values in the facial image and maintaining the light-and-dark relationship between the pixel values in the facial image. At the same time, the facial image cannot be recognized by the human eye, and the process is irreversible.
RLLFPR consists of two main networks, the SN network (represented by the blue–green–purple parts in Figure 1) and the ResNet [36] network (represented by the yellow part in Figure 1). The SN network is responsible for generating privacy-preserved faces, while the ResNet network is used for recognizing these privacy-preserved faces. The SN uses SegNet [37] as the main module, and the SegNet network is an encoder–decoder network. The max-pooling layer is used for downsampling in the encoder network. In the decoder network, SegNet is different from previous upsampling methods (such as deconvolution or linear interpolation). It uses the max-pooling layer index generated by the downsampling max-pooling process for upsampling. The concrete SN adopts a five-layer encoder–decoder structure and removes the last Softmax layer, and the number of output channels of the last convolutional layer is one. The green part is composed of three parts: convolutional layer + batch normalization layer + ReLU activation function. The blue part is the max-pooling layer, which not only is used in the downsampling but also produces the max-pooling index (the purple part in Figure 1) for upsampling.

3.1. Privacy-Preserving Face Recognition Based on Randomization and Local Feature Learning (RLLFPR) Training Method

During the training process of the RLLFPR, although the two networks are jointly trained, the backpropagation of the loss function only involves the ResNet network and does not encompass the entire network. The SN network learns through batch normalization layers, enabling self-learning. Figure 2 illustrates the changes in facial images during training and how the parameters of the two networks are updated. The SN network learns through the batch normalization layers and generates SN_PrivacyFace (SNPF) images that are difficult for human vision to recognize, as depicted in Figure 3. The ResNet parameters are learned through the backpropagation of the Cosface loss function (Equation (1)). After training, the ResNet network can accurately recognize SNPF images.
L o s s = 1 N i i log e s ( cos ( θ y i , i ) ) m ) e s ( cos ( θ y i , i ) ) m ) + j y i e s ( cos ( θ y i , i ) )

3.2. Randomized Convolution and Batch Normalization Learning for RLLFPR

In RLLFPR, the SN utilizes random convolution through the convolutional layers to randomize the pixel values of the face. The batch normalization layer, with learned mean and variance, preserves the brightness relationship between the pixel values within the face region. During the initial training phase, the SN randomly initializes the convolutional kernels in all the convolutional layers used in the upsampling and downsampling processes of the network structure. The values of the convolutional kernels follow a normal distribution. Without backpropagation of a loss function, the parameters of the convolutional layers remain unchanged, and the convolutional kernels retain their initial random values. The convolution operation process is shown in Figure 4, where x1–x9 are the random values of the initial convolution kernel.
In the SN, images after multiple rounds of random convolution operations can have high fuzziness. In the convolution operation, it is assumed that the original image is a grayscale image of n × n (n is even). Equation (2) is the output image size formula of randomized convolution; input, p, k, s, and output represent the input image size, zero filling, convolutional kernel size, stride, and output image size, respectively. After the convolution operation with the convolution kernel size k = 3, stride s = 1, and zero filling p = 1, the image size is still n × n (Equations (2) and (3)). In this process, the pixel values in the image are randomized to generate privacy-protected images that are difficult for human vision to recognize.
o u p u t = i n p u t + 2 p k s + 1
a 11 a 12 a 1 ( n 1 ) a 1 n a 21 a 22 a 2 ( n 1 ) a 2 n a ( n 1 ) 1 a ( n 1 ) 2 a ( n 1 ) ( n 1 ) a ( n 1 ) n a n 1 a n 2 a n ( n 1 ) a n n n × n k s = 1 b 11 b 12 b 1 ( n 1 ) b 1 n b 21 b 22 b 2 ( n 1 ) b 2 n b ( n 1 ) 1 a ( n 1 ) 2 b ( n 1 ) ( n 1 ) b ( n 1 ) n b n 1 b n 2 b n ( n 1 ) b n n n × n
The SN network first reduces and then enlarges the facial image, with the downsampling process primarily utilizing the max-pooling layer. The max-pooling layer is similar to a convolutional layer but does not require convolutional kernels. It is used to reduce the size of the feature map and extract the main features. Specifically, the max-pooling operation takes the maximum value within each region and assigns it the value at the corresponding position in the output feature map, as shown in Equation (4). In other words, it summarizes the pixel values within each region by taking the maximum value. The output feature map has a reduced size but retains the most significant features within each region. The output image size is determined by Equation (2), where k represents the region size, s is the stride, and p is the zero padding. Additionally, during the downsampling process, the position indices of the max-pooling layer are saved and used for upsampling in subsequent steps.
a 11 a 12 a 1 ( n 1 ) a 1 n a 21 a 22 a 2 ( n 1 ) a 2 n a ( n 1 ) 1 a ( n 1 ) 2 a ( n 1 ) ( n 1 ) a ( n 1 ) n a n 1 a n 2 a n ( n 1 ) a n n n × n k = 2 , s = 2 , p = 0 a 11 a 1 n a n 1 a n n n 2 × n 2
During the upsampling process, the downsampling max-pooling layer’s index matrix is utilized to enlarge the reduced image. The maximum values obtained from the downsampling process are placed back into their original positions in the matrix using the index positions, while the other values within the region are set to 0. This process is described by Equation (5), where k represents the region size, s is the stride, and index refers to the index matrix. By performing upsampling the same number of times as downsampling, the image can be restored to its original size. Utilizing the max-pooling layer’s index to upscale the image reduces the number of network parameters and improves computational efficiency.
a 11 a 1 n a n 1 a n n n 2 × n 2 k = 2 , s = 2 , i n d e x a 11 0 0 a 1 n 0 0 0 0 0 0 0 0 a n 1 0 0 a n n n × n
During the downsampling process in the SN network, randomization is introduced through convolutional layers, adding randomness to the image. The image size is reduced and some information is removed through the max-pooling layer. The removed information cannot be fully recovered during the unlearned upsampling process, resulting in differences between the generated image and the original image. The combination of randomness and removed information creates privacy-preserved images that are difficult for human vision to recognize.
The convolution operation adds randomness while preserving some information from the original image. This information is enhanced in the batch normalization layer. The self-learning of the batch normalization layer can learn the light-and-dark difference of pixel values in the image, which can be recognized by the subsequent network. The batch normalization layer specifically learns the mean and variance of the data. Different from the convolutional layer, the mean and variance of the batch normalization layer in the training mode can also update the parameters without loss-function backpropagation. The SN structure has a convolutional layer followed by a batch normalization layer (Figure 5). x is the value after the randomized convolutional layer, and y is the value of x after the batch normalization layer; the mean(x), Var(x), represented mean, variance, and eps prevent the divisor from being zero. λ and β represent the updated backpropagation parameters, defaulting to 1 and 0 (Equation (6)). On the basis of the preset hyperparameter m, the variance and mean will constantly self-learn and update during the training process (Equations (7) and (8)). x and x’ denote different batches of data. In the prediction stage, the batch normalization layer processes the data with the mean and variance learned by training, and no more updates are performed.
y = x m e a n ( x ) V a r ( x ) + e p s γ + β
m e a n = ( 1 m ) m e a n ( x ) + m m e a n ( x ) V a r = ( 1 m ) V a r ( x ) + m V a r ( x )
V a r = ( 1 m ) V a r ( x ) + m V a r ( x )
Since the convolution kernel is random, the image is passed through a convolution layer to generate a lower-information image (Figure 6a). After multiple convolutional layers, affected by the randomized convolution kernel, the image has no information. Such images cannot be used for face recognition. However, we observe that images combined by convolutional layers and batch normalization layers can generate preliminarily recognizable facial images (Figure 6b). The facial image in Figure 6b contains more recognition features, such as face contour, eye, nose, mouth, and other contours, and other original image information than the facial image in Figure 6a. The whole process is as follows: the convolution layer randomizes the pixel values of the facial image in the region, and the batch normalization layer maintains the light-and-dark relationship between the pixel values in the region. After multiple rounds of this cycle, the privacy-protected face image that is difficult to recognize by human vision but can be accurately recognized by machine vision is generated.
In the convolution process, the randomness of the convolution kernel also makes the final image revocable and irreversible. The randomness of the convolution kernel makes the same image obtain different results through different convolution kernels, see Equation (3), and this randomness will also be enhanced in multiple convolution operations. Different random convolution kernels are used for each training, so the networks between training cannot communicate with each other. This makes the facial images in the database revocable, and even if the server-side face data are stolen (the server-side stores the privacy-preserved facial images), the whole network can be retrained so that the stolen facial images cannot be recognized normally. At the same time, the convolution operation is irreversible, so the resulting image is also irreversible.

4. Experiment

4.1. Setting

We used RLLFPR with the LFW (Labeled Faces in the Wild) [38], Celeba, and HDU (camera-captured facial images) datasets. The LFW dataset consists of 13,233 face images. The celebrity dataset Celeba randomly selects 10% of its face images (20,177 faces) for training and testing. The HDU dataset collected by our camera consists of 3891 face images of 51 young people (around 20 years old) under different background and lighting conditions. These faces were uniformly aligned and enlarged to 160 × 160 size by the MTCNN [39] method. The ResNet network was updated to ResNet50. The loss function used Cosface (S = 30, M = 0.5) (Equation (1)). The Python language, Pyotch1.8 library, and GTX 2080TI graphics card were used. A Microsoft LifeCam HD-3000 was used to capture the images.
The ResNet50 was first pretrained using the LFW dataset with epoch = 256, learning rate = 0.1, and batchsize = 32. After pretraining the ResNet50, the RLLFPR network was trained, and the SN and ResNet were jointly trained and tested on the training samples. Furthermore, we used Equation (1) as the loss function, epoch = 64, learning rate = 0.001, batchsize = 32.

4.2. The Recognition Performance of Privacy-Preserving Face Recognition Based on Randomization and Local Feature Learning (RLLFPR)

We mainly compared the accuracy, misidentification rate, and F1 score of an RLLFPR image with the original face image without privacy protection and images processed with Arnold transform [2], eigenface [40], differential privacy [33], and PartialFace [35] methods on the three datasets, and the specific results are shown in Table 1. The Arnold transform method processes the face through the method of data processing. The eigenface method uses PCA to extract part of the features of the face for recognition. The differential privacy method adds differential privacy on the basis of the eigenface. PartialFace blurs the face by removing the low-frequency information of the image and partially randomizes the high-frequency information.
The original face images without privacy protection, shown in Table 1, performs the best on the three datasets. The privacy-protected faces generated by the AES encryption method cannot be recognized in the case of ciphertext, so the accuracy is about 0 in the tests on the three datasets. On the LFW dataset, RLLFPR had the highest accuracy of 99.93%. It was better than the 99.82% of the Arnold transform method, 83.73% of differential privacy, and 99.80% of PartialFace. In terms of F1 score, RLLFPR was the same as PartialFace, with 99.67%, which was better than the other privacy-preserving recognition algorithms. The misidentification-rate score of RLLFPR was 0.13%, which was slightly worse than Arnold’s 0.10% but better than the other methods. On the Celeba dataset, RLLFPR had the highest recognition accuracy of 98.77%, the highest F1 score of 98.67%, and the lowest misidentification rate of 0.11%, which was better than the existing methods. In the dataset collected by the camera, due to the small size of the dataset and the environment, the lighting and other factors changed greatly. The method of differential privacy performed poorly on the HDU dataset, with an accuracy and F1 score below 50% and misidentification rate greater than 50%. The accuracy of the eigenface method was only 80.39%, and the accuracy of the Arnold transform was only 90.57%. PartialFace and RLLFPR also had excellent results on the HDU dataset, with an accuracy of more than 99%. RLLFPR was better than PartialFace, with an accuracy of 99.58%, misidentification rate of only 0.59%, and F1 score of 99.23%.
Discussion 1: Compared with the previous privacy-protection methods, RLLFPR had a higher accuracy and F1 score and lower misidentification rate. At the same time, compared with the original face image without protection, the accuracy of RLLFPR only had a slight decrease, which was only 0.02% lower on the LFW dataset. On the Celeba dataset, the maximum decline was 0.92%, and the final accuracy was 98.77%, which was also better than the other privacy protection–recognition methods. On the HDU dataset, the accuracy was reduced by 0.39%, and the accuracy was still more than 99.5%. In general, RLLFPR is better than the previous methods in balancing privacy protection and face recognition and is superior to the existing methods.

4.3. RLLFPR Fuzzing Performance Test

Some privacy-preserved faces produced by RLLFPR in different datasets are shown in Figure 7. Compared with the original faces, the privacy-protected faces generated by RLLFPR in the three datasets changed greatly. From the perspective of human vision, information about the original face cannot be recognized from the privacy-preserved face.

4.3.1. Fuzziness Test Method

In terms of the fuzziness of privacy-protected images generated, because PartialFace retains part of the contours of the face, the recognition accuracy of the differential privacy method was less than 90%. Therefore, the main comparison is between RLLFPR and the Arnold transform method and privacy-protected images produced by AES encryption. RLLFPR, Arnold transform, and AES encryption methods all blur the original face to a large extent, which makes it difficult to identify with human vision.
The privacy-preserved faces generated by RLLFPR cannot be recognized by human vision, and some statistical characteristics of their performance are shown in Figure 8. Figure 8 shows the grayscale histogram of the original image, processed by the RLLFPR method, Arnold transform method, and AES encryption method. Compared with the original image (Figure 8a,e), the gray distribution of the Arnold transform image (Figure 8c,g) is basically unchanged, which is related to the restoration of the original image after multiple rounds of operation. The gray histogram obtained by the RLLFPR method (Figure 8b,f) and the AES encrypted image (Figure 8d,h) show similar distributions for different faces; the RLLFPR method shows a similar bell-shaped distribution; and AES shows a similar white-noise distribution. The similar gray histogram distribution of different faces can improve the blurring performance of facial images.
Figure 9 and Table 2 show the original image, Arnold, AES, as well as RLLFPR image-adjacent pixel correlation analysis. Specifically, 3000 pairs of adjacent pixels from the three R, G, B channels were selected for analysis. As can be seen in Table 2, the original image (Figure 9a) presented a high linear correlation, and the three directional correlation values were all greater than 0.98. Although Arnold transform (Figure 9c) blurred what can be seen with human vision, it only moved pixels between locations, so it still had high correlation, and the three directional phase property values were around 0.8. RLLFPR (Figure 9d) and AES (Figure 9b) showed low correlation in the correlation analysis; RLLFPR showed three correlations around 0.05, and AES had even lower values, around 0.005. Low correlation can indicate that the image pixels are less regular and difficult to predict. In general, the correlation effect of RLLFPR is much higher than that of the Arnold transform and slightly inferior to the AES method.
Table 3 shows a comparison of the PSNR and UACI values of the privacy-protected face images generated by RLLFPR with the Arnold transformed, noisy, and AES encrypted images. The noise-processed face is still recognizable and can be used as an intermediate value for comparison. The PSNR value is the highest and the UACI value is the lowest for the face after noise processing, indicating its high recognizability. The processed image with noise has less distortion and blurring. The RLLFPR, Arnold, and AES encrypted versions of all three images are not visible to the human eye and show better results in terms of PSNR and UACI. The RLLFPR method is similar to the Arnold transform in terms of both data PSNR and UACI, with PSNR values of 11.80 and 11.53 and UACI values of 50.61 and 53.93. All methods resulted in an AES-encrypted ciphertext with values of 8.77 and 75.27. However, AES-encrypted images cannot be recognized.
In addition, we designed three fuzziness tests.
First, we conducted human evaluation experiments by means of questionnaires. The questionnaire consisted of two types of single-choice questions. The experimental group was based on the privacy-preserved faces generated by RLLFPR, and the control group comprised the privacy-preserved images processed by AES encryption. The first single-item choice was entitled “Similarity problem” (Figure 10a), which judged the degree of similarity between two faces by giving the original facial image and the privacy-preserved image of the original face. Five options were given, respectively, is completely not similar (4 points), a small degree of similarity (3 points), not sure (2 points), most similar (1 points), and completely similar (0 points), using scoring judgment of fuzzy degrees. The higher the score, the higher the fuzziness of the image.
The second type of single-choice question (Figure 10b) was a matching question. On the basis of the original face, four faces after privacy protection were given, and the privacy-protected face that the participant believed belongs to the same person as the original face was selected. The lower the accuracy, the higher the image fuzziness.
A total of 100 questionnaires (groups 1 and 2 each had 10 questions) were distributed, and 98 were collected, resulting in an effective response rate of 98%. The survey results are considered valid. The statistical results for each question in the questionnaire are as follows (Table 4).
Discussion 2: The statistical results for each question in the questionnaire are as follows (Table 4). It can be observed that the privacy-preserved face generated by RLLFPR (SNPF) and the face encrypted with AES show similar results. Both methods received high scores in the first group. The similarity values are 3.51 and 3.55, respectively, which reflects that the privacy-preserved images generated by RLLFPR are visually similar to the effect of the AES-encrypted images and that the processed and original images have a low similarity. In the second group, the matching accuracy is close to 25%, indicating that the respondents were close to randomly selecting their answers. Accuracy close to 25% proves that both images could not be matched to the original image, and even if privacy-preserved images were acquired, they could not be matched to a specific person.The results of the two questionnaires show that the privacy-preserved faces generated by RLLFPR have high ambiguity, similar to encryption methods. PartialFace [35] achieved high similarity compared to the original face in the first questionnaire, and in the second questionnaire, humans recognized privacy-protected faces with accuracies reaching or even exceeding 70–80%. Therefore, it was judged that the PartialFace method for human eye visual blurring was unsatisfactory, and no subsequent blurring test was performed.
Second, the idea of convolution is used to calculate the mean value of the image block, and the variance (Equations (9)–(11)) is calculated for all the mean values. s is the block size, which means the mean value of the s × s region size. Sharp images are colorful, and V ( i m g ) tends to be larger, while fuzzy images are smaller. V ( i m g i ) V ( i m g o ) is generally between 0 and 1. i m g i refers to privacy-protected images and i m g o refers to original images:
V ( i m g ) = var ( i m g i k s × s )
k s × s = 1 s s 1 s s 1 s s 1 s s
Fuzziness 2 is shown in Equation (11); the larger the fuzziness of the image, the higher the value.
F u z z i n e s s 2 = 1 V ( i m g i ) V ( i m g o )
The third kind of fuzziness is the method of edge detection (Equations (12) and (13)) + SSIM similarity detection (Equation (14)). Firstly, the edge of the image is detected, and then the SSIM similarity between the privacy-protected image and the original image is calculated. The lower the similarity, the higher the ambiguity. The Laplacian operator is used for edge detection.
L a p l a c i a n ( i m g ) = 2 i m g x 2 + 2 i m g y 2
k l a p l a c i a n = 0 1 0 1 4 1 0 1 0
The SSIM algorithm [41] similarity index measures the similarity of images, ranging from 0 to 1. The larger the value, the higher the similarity of the image; μ and σ represent the mean and standard deviation, respectively.
S S I M ( x , y ) = ( 2 μ x μ y + c 1 ) ( 2 μ x y + c 2 ) ( μ x 2 + μ y 2 + c 1 ) ( σ x 2 + σ y 2 + c 12 )
The image fuzzy degree Fuzziness3 is obtained by subtracting the SSIM similarity from 1 (Equation (15)), and the higher the value, the higher the fuzzy degree.
F u z z i n e s s 3 = 1 S S I M ( l a p ( i m g o ) , l a p ( i m g i ) )

4.3.2. Fuzzy Performance of Privacy-Preserving Face Recognition Based on Randomization and Local Feature Learning (RLLFPR)

Table 5 shows the values obtained by different privacy-protection methods after the fuzziness test methods, and the comparison of privacy-protection performance between different methods is added in Table 5. The Fuzziness1 results are shown in Table 4. As can be seen from the table, the ambiguity of RLLFPR measured by Fuzziness2 is in the range of 0.80–0.95, which is obviously better than that of noise addition, Arnold transformation, and other methods. AES is the encryption algorithm with ambiguity in the range 0.6–0.9, and the encrypted image cannot be directly recognized. Measured by Fuzziness3, RLLFPR also has a good effect. RLLFPR (0.997) is better than Arnold (0.996), and noise (0.992) is second to AES (0.999). In summary, the privacy-preserving facial fuzzy image generated by RLLFPR is better than the additive noise method and Arnold method and is similar to the AES encryption method. The generated faces in RLLFPR can effectively obscure human vision, providing protection against unauthorized access to facial data stored on servers. RLLFPR eliminates the need for encryption keys, reducing the burden on individuals concerned about their identification. In addition, RLLFPR can directly recognize privacy-protected images.
Discussion 3: In terms of security, RLLFPR is irreversible as well as revocable. Irreversibility: The core of the SN network is composed of a convolutional layer, a batch normalization layer, and an activation-function and maximum-pooling layer. The convolution operation in the convolutional layer and the pooling operation in the maximum-pooling layer are irreversible operations. The activation function uses a nonlinear ReLU activation function, which is also irreversible. Therefore, with the stacked use of irreversible operations, an irreversible privacy-protected image will eventually be produced, and even if the privacy-protected image is obtained, the probability of recovering from it the original face image is close to 0. The ciphertext encrypted by AES cannot be recognized directly, and the Arnold transformed image can be used to recover the original image after many transformations. This all increases the risk of privacy leakage.
Revocability: Since RLLFPR generates blurred images that cannot be recognized by the human eye, it is impossible to match the face images in the database with the real face images even if the database is stolen. The Arnold transform method is not revocable since its images can be restored to the original images. Combining the two aspects of blurriness and security, RLLFPR is more advantageous than the previous methods.

4.4. Network Structure Ablation Experiment of RLLFPR Method

During the training process of RLLFPR, the entire SN network does not update its parameters through the loss function. To demonstrate the effectiveness of the structural improvement, experiments were conducted on the SN network with parameter updates through loss backpropagation. In this case, the SN network still used the Cosface loss function (Equation (1)). However, upon careful observation, it was found that the SN network trained with loss backpropagation retained the contour information of the original face, resulting in suboptimal privacy protection (Figure 11, the generated privacy-protected image is enlarged). The experiments tested these two types of images under three different levels of blurriness (Table 6). From Table 6, it can be observed that the facial images generated by the RLLFPR method have higher privacy-protection effectiveness compared to the SN network trained with the loss function. In the first fuzziness test method, the similarity value of SN after self-learning is 3.51, which is higher than that of SN after loss backpropagation, which is 2.11. It is proved that the face image produced by the self-learning SN is shielded from human-eye vision. In the second and third fuzziness test methods, the self-learned SN is 0.80–0.95 and 0.997, respectively, which is better than the SN after loss function backpropagation of 0.75–0.95 and 0.983. Thus, the self-learned SN produces more ambiguous privacy-preserved faces than the SN after loss function backpropagation.
For the recognition of privacy-protected faces on the backbone network, a ResNet was chosen due to its superior performance in face recognition and image classification. Specifically, ResNet50 was selected. Additionally, DenseNet [42] was compared to the ResNet, as it has a similar excellent effect in image classification. For DenseNet, we selected DenseNet121. The results of the training on the LFW dataset are presented in Table 7. Both ResNet50 and DenseNet121 showed similar performance in terms of accuracy, with ResNet50 achieving a slightly higher accuracy of 99.93% compared to DenseNet121’s 99.91%. However, ResNet50 had a higher misidentification rate of 0.13% compared to DenseNet121’s 0.07%. Both models had the same F1 score of 99.67%. Therefore, either backbone network can be selected for DenseNet.

5. Conclusions

In this paper, we study the recognizable space of face images and find that under the condition of appropriately maintaining the structure of face components and local interpixel light-and-shadow relations, randomizing the rest of the information can still allow the model to accurately recognize face images with a recognition accuracy of >99%. On this basis, a face privacy recognition method, RLLFPR, is proposed. Instead of using traditional cryptography or image-processing methods to generate privacy-protected faces, RLLFPR takes advantage of the inherent randomness of the convolutional layer and the self-learning function of the batch normalization layer in deep neural networks. By randomizing the pixel values between face image regions and maintaining the light–dark relationship between face pixels, it generates privacy-protected faces that are difficult to recognize by human vision, with a recognition rate of nearly 25% in a four-choice–one-human recognition test, which is equivalent to the recognition rate of random guessing. At present, the RLLFPR method requires a high number of face training sets and mainly needs to adapt to as many lighting environments as possible. Further research considerations include continuing to explore the identifiable space of facial images and building a more portable, controllable, and effective network structure to generate facial images with better recognition effects and privacy-protection effects.

Author Contributions

Conceptualization, Z.W.; methodology, Z.W. and Y.H.; software, Z.W. and Y.H.; validation, Y.H.; formal analysis, Z.W. and Y.H.; investigation, Z.W.; resources, Z.W.; data curation, Z.W.; writing—original draft preparation, Y.H.; writing—review and editing, Z.W. and Y.H.; visualization, J.C. and H.X.; supervision, J.C. and H.X.; project administration, J.C. and H.X.; funding acquisition, Z.W. All authors have read and agreed to the published version of the manuscript.

Funding

This study is supported by the Key Projects of NSFC Joint Fund of China (No. U1866209), National Key R&D Program of China (No. 2018YFB0804102), and National Natural Science Foundation of China (No. 61772162).

Institutional Review Board Statement

Ethical review and approval were waived for this study due to use of public datasets and private datasets with the consent of all people.

Informed Consent Statement

Both public and private datasets are used in this paper. Informed consent for the private dataset has been obtained from all subjects participating in the study.

Data Availability Statement

The LFW and Celeba face databases used in this paper are publicly available face databases. The images in the HDU face database were captured with the informed consent of the recorded subjects. The HDU database face images disclosed herein have the informed consent of the users. The HDU database is not public.The LFW dataset URL is https://vis-www.cs.umass.edu/lfw/ accessed on 30 January 2024. The Celeba dataset URL is https://mmlab.ie.cuhk.edu.hk/projects/CelebA.html accessed on 30 January 2024.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Boddeti, V.N. Secure face matching using fully homomorphic encryption. In Proceedings of the 2018 IEEE 9th International Conference on Biometrics Theory, Applications and Systems (BTAS), Redondo Beach, CA, USA, 22–25 October 2018; pp. 1–10. [Google Scholar]
  2. Zhang, J.; Shen, W.; Wu, Z. Recognition of face privacy protection using convolutional neural networks. J. Image Graph. 2019, 24, 0744–0752. [Google Scholar]
  3. Wu, Y.; Yang, F.; Xu, Y.; Ling, H. Privacy-protective-GAN for privacy preserving face de-identification. J. Comput. Sci. Technol. 2019, 34, 47–60. [Google Scholar] [CrossRef]
  4. Liu, Z.; Wu, Z.; Li, T.; Li, J.; Shen, C. GMM and CNN hybrid method for short utterance speaker recognition. IEEE Trans. Ind. Inform. 2018, 14, 3244–3252. [Google Scholar] [CrossRef]
  5. Schroff, F.; Kalenichenko, D.; Philbin, J. Facenet: A unified embedding for face recognition and clustering. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 815–823. [Google Scholar]
  6. Cao, Q.; Shen, L.; Xie, W.; Parkhi, O.M.; Zisserman, A. Vggface2: A dataset for recognising faces across pose and age. In Proceedings of the 2018 13th IEEE International Conference on Automatic Face & Gesture Recognition (FG 2018), Xi’an, China, 15–19 May 2018; pp. 67–74. [Google Scholar]
  7. Wen, Y.; Zhang, K.; Li, Z.; Qiao, Y. A discriminative feature learning approach for deep face recognition. In Proceedings of the Computer Vision–ECCV 2016: 14th European Conference, Amsterdam, The Netherlands, 11–14 October 2016; Proceedings, Part VII 14. Springer: Berlin/Heidelberg, Germany, 2016; pp. 499–515. [Google Scholar]
  8. Liu, W.; Wen, Y.; Yu, Z.; Li, M.; Raj, B.; Song, L. Sphereface: Deep hypersphere embedding for face recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 212–220. [Google Scholar]
  9. Wang, H.; Wang, Y.; Zhou, Z.; Ji, X.; Gong, D.; Zhou, J.; Li, Z.; Liu, W. Cosface: Large margin cosine loss for deep face recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 5265–5274. [Google Scholar]
  10. Deng, J.; Guo, J.; Xue, N.; Zafeiriou, S. Arcface: Additive angular margin loss for deep face recognition. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 4690–4699. [Google Scholar]
  11. Meden, B.; Rot, P.; Terhörst, P.; Damer, N.; Kuijper, A.; Scheirer, W.J.; Ross, A.; Peer, P.; Štruc, V. Privacy–enhancing face biometrics: A comprehensive survey. IEEE Trans. Inf. Forensics Secur. 2021, 16, 4147–4183. [Google Scholar] [CrossRef]
  12. Aljeraisy, A.; Barati, M.; Rana, O.; Perera, C. Privacy laws and privacy by design schemes for the internet of things: A developer’s perspective. ACM Comput. Surv. (Csur) 2021, 54, 1–38. [Google Scholar] [CrossRef]
  13. You, Z.; Li, S.; Qian, Z.; Zhang, X. Reversible privacy-preserving recognition. In Proceedings of the 2021 IEEE International Conference on Multimedia and Expo (ICME), Shenzhen, China, 5–9 July 2021; pp. 1–6. [Google Scholar]
  14. Goodfellow, I.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative adversarial networks. Commun. ACM 2020, 63, 139–144. [Google Scholar] [CrossRef]
  15. Korshunova, I.; Shi, W.; Dambre, J.; Theis, L. Fast face-swap using convolutional neural networks. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 3677–3685. [Google Scholar]
  16. Yang, Y.; Huang, Y.; Shi, M.; Chen, K.; Zhang, W. Invertible mask network for face privacy preservation. Inf. Sci. 2023, 629, 566–579. [Google Scholar] [CrossRef]
  17. Mirjalili, V.; Raschka, S.; Ross, A. PrivacyNet: Semi-adversarial networks for multi-attribute face privacy. IEEE Trans. Image Process. 2020, 29, 9400–9412. [Google Scholar] [CrossRef] [PubMed]
  18. Liu, J.; Zhao, Z.; Li, P.; Min, G.; Li, H. Enhanced Embedded AutoEncoders: An Attribute-Preserving Face De-identification Framework. IEEE Internet Things J. 2023, 10, 9438–9452. [Google Scholar] [CrossRef]
  19. Vanitha, C.; Malathy, S.; Anitha, K.; Suwathika, S. Enhanced Security using Advanced Encryption Standards in Face Recognition. In Proceedings of the IEEE 2021 2nd International Conference on Communication, Computing and Industry 4.0 (C2I4), Bangalore, India, 16–17 December 2021; pp. 1–5. [Google Scholar]
  20. Erkin, Z.; Franz, M.; Guajardo, J.; Katzenbeisser, S.; Lagendijk, I.; Toft, T. Privacy-preserving face recognition. In Proceedings of the Privacy Enhancing Technologies: 9th International Symposium, PETS 2009, Seattle, WA, USA, 5–7 August 2009; Proceedings 9. Springer: Berlin/Heidelberg, Germany, 2009; pp. 235–253. [Google Scholar]
  21. Bai, J.; Zhang, X.; Song, X.; Shao, H.; Wang, Q.; Cui, S.; Russello, G. CryptoMask: Privacy-preserving Face Recognition. In Proceedings of the International Conference on Information and Communications Security, Tianjin, China, 18–20 November 2023; Springer: Singapore, 2023; pp. 333–350. [Google Scholar]
  22. Ma, Z.; Liu, Y.; Liu, X.; Ma, J.; Ren, K. Lightweight privacy-preserving ensemble classification for face recognition. IEEE Internet Things J. 2019, 6, 5778–5790. [Google Scholar] [CrossRef]
  23. Boragule, A.; Yow, K.C.; Jeon, M. On-device Face Authentication System for ATMs and Privacy Preservation. In Proceedings of the 2023 IEEE International Conference on Consumer Electronics (ICCE), Las Vegas, NV, USA, 6–8 January 2023; pp. 1–4. [Google Scholar]
  24. Im, J.H.; Jeon, S.Y.; Lee, M.K. Practical privacy-preserving face authentication for smartphones secure against malicious clients. IEEE Trans. Inf. Forensics Secur. 2020, 15, 2386–2401. [Google Scholar] [CrossRef]
  25. Zhang, Y.; Xiao, X.; Yang, L.X.; Xiang, Y.; Zhong, S. Secure and efficient outsourcing of PCA-based face recognition. IEEE Trans. Inf. Forensics Secur. 2019, 15, 1683–1695. [Google Scholar] [CrossRef]
  26. Lei, J.; Pei, Q.; Wang, Y.; Sun, W.; Liu, X. PRIVFACE: Fast privacy-preserving face authentication with revocable and reusable biometric credentials. IEEE Trans. Dependable Secur. Comput. 2021, 19, 3101–3112. [Google Scholar] [CrossRef]
  27. Mai, G.; Cao, K.; Lan, X.; Yuen, P.C. Secureface: Face template protection. IEEE Trans. Inf. Forensics Secur. 2020, 16, 262–277. [Google Scholar] [CrossRef]
  28. Walia, G.S.; Jain, G.; Bansal, N.; Singh, K. Adaptive weighted graph approach to generate multimodal cancelable biometric templates. IEEE Trans. Inf. Forensics Secur. 2019, 15, 1945–1958. [Google Scholar] [CrossRef]
  29. Morales, A.; Fierrez, J.; Vera-Rodriguez, R.; Tolosana, R. SensitiveNets: Learning agnostic representations with application to face images. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 43, 2158–2164. [Google Scholar] [CrossRef]
  30. Zhang, Y.; Wang, T.; Zhao, R.; Wen, W.; Zhu, Y. RAPP: Reversible Privacy Preservation for Various Face Attributes. IEEE Trans. Inf. Forensics Secur. 2023, 18, 3074–3087. [Google Scholar] [CrossRef]
  31. Wu, Z.; Kang, J.; Jiang, Q. Semantic key generation based on natural language. Int. J. Intell. Syst. 2022, 37, 4041–4064. [Google Scholar] [CrossRef]
  32. Wu, Z.; Lv, Z.; Kang, J.; Ding, W.; Zhang, J. Fingerprint bio-key generation based on a deep neural network. Int. J. Intell. Syst. 2022, 37, 4329–4358. [Google Scholar] [CrossRef]
  33. Chamikara, M.A.P.; Bertok, P.; Khalil, I.; Liu, D.; Camtepe, S. Privacy preserving face recognition utilizing differential privacy. Comput. Secur. 2020, 97, 101951. [Google Scholar] [CrossRef]
  34. Mi, Y.; Huang, Y.; Ji, J.; Liu, H.; Xu, X.; Ding, S.; Zhou, S. Duetface: Collaborative privacy-preserving face recognition via channel splitting in the frequency domain. In Proceedings of the 30th ACM International Conference on Multimedia, Lisboa, Portugal, 10–14 October 2022; pp. 6755–6764. [Google Scholar]
  35. Mi, Y.; Huang, Y.; Ji, J.; Zhao, M.; Wu, J.; Xu, X.; Ding, S.; Zhou, S. Privacy-Preserving Face Recognition Using Random Frequency Components. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Paris, France, 2–3 October 2023; pp. 19673–19684. [Google Scholar]
  36. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  37. Badrinarayanan, V.; Kendall, A.; Cipolla, R. Segnet: A deep convolutional encoder-decoder architecture for image segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 2481–2495. [Google Scholar] [CrossRef]
  38. Huang, G.B.; Mattar, M.; Berg, T.; Learned-Miller, E. Labeled faces in the wild: A database forstudying face recognition in unconstrained environments. In Workshop on Faces in ‘Real-Life’ Images: Detection, Alignment, and Recognition; HAL: Lyon, France, 2008. [Google Scholar]
  39. Zhang, K.; Zhang, Z.; Li, Z.; Qiao, Y. Joint face detection and alignment using multitask cascaded convolutional networks. IEEE Signal Process. Lett. 2016, 23, 1499–1503. [Google Scholar] [CrossRef]
  40. Turk, M.; Pentland, A. Eigenfaces for recognition. J. Cogn. Neurosci. 1991, 3, 71–86. [Google Scholar] [CrossRef] [PubMed]
  41. Wang, Z.; Bovik, A.C.; Sheikh, H.R.; Simoncelli, E.P. Image quality assessment: From error visibility to structural similarity. IEEE Trans. Image Process. 2004, 13, 600–612. [Google Scholar] [CrossRef] [PubMed]
  42. Huang, G.; Liu, Z.; Van Der Maaten, L.; Weinberger, K.Q. Densely connected convolutional networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 4700–4708. [Google Scholar]
Figure 1. Privacy-preserving face recognition based on randomization and local feature learning (RLLFPR) structure.
Figure 1. Privacy-preserving face recognition based on randomization and local feature learning (RLLFPR) structure.
Jimaging 10 00059 g001
Figure 2. Training process of privacy-preserving face recognition based on randomization and local feature learning (RLLFPR).
Figure 2. Training process of privacy-preserving face recognition based on randomization and local feature learning (RLLFPR).
Jimaging 10 00059 g002
Figure 3. The original face is passed through the SN network to produce the privacy-preserved face, the SN_PrivacyFace (SNPF).
Figure 3. The original face is passed through the SN network to produce the privacy-preserved face, the SN_PrivacyFace (SNPF).
Jimaging 10 00059 g003
Figure 4. Randomized convolution operation.
Figure 4. Randomized convolution operation.
Jimaging 10 00059 g004
Figure 5. Convolutional layer and batch normalization layer in SN network.
Figure 5. Convolutional layer and batch normalization layer in SN network.
Jimaging 10 00059 g005
Figure 6. (a) Facial image after the convolutional layer processing; (b) facial image after the convolutional and batch normalization layer processing.
Figure 6. (a) Facial image after the convolutional layer processing; (b) facial image after the convolutional and batch normalization layer processing.
Jimaging 10 00059 g006
Figure 7. Privacy-preserved faces produced by RLLFPR on different datasets: (a,b) from the LFW dataset, (c,d) from the Celeba dataset, and (e) from the HDU dataset.
Figure 7. Privacy-preserved faces produced by RLLFPR on different datasets: (a,b) from the LFW dataset, (c,d) from the Celeba dataset, and (e) from the HDU dataset.
Jimaging 10 00059 g007
Figure 8. Gray histograms of two different face images from different methods: (a,e) derived from the original face images, (b,f) derived from RLLFPR after processing, (c,g) derived from Arnold transformed images, and (d,h) derived from AES encrypted images. (ad) each represent a human face and (eh) represent a different face from (ad).
Figure 8. Gray histograms of two different face images from different methods: (a,e) derived from the original face images, (b,f) derived from RLLFPR after processing, (c,g) derived from Arnold transformed images, and (d,h) derived from AES encrypted images. (ad) each represent a human face and (eh) represent a different face from (ad).
Jimaging 10 00059 g008
Figure 9. Correlation analysis of adjacent pixels of original face, AES, Arnold, and RLLFPR images: (a) original face, (b) AES, (c) Arnold, and (d) RLLFPR.
Figure 9. Correlation analysis of adjacent pixels of original face, AES, Arnold, and RLLFPR images: (a) original face, (b) AES, (c) Arnold, and (d) RLLFPR.
Jimaging 10 00059 g009
Figure 10. Composition of questionnaire questions. (a) the direct similarity between the privacy-protected image and the original image; (b) the direct matching degree between the privacy-protected image and the original image.
Figure 10. Composition of questionnaire questions. (a) the direct similarity between the privacy-protected image and the original image; (b) the direct matching degree between the privacy-protected image and the original image.
Jimaging 10 00059 g010
Figure 11. Different privacy-preserved images produced by SN network after self-learning and SN with loss function backpropagation.
Figure 11. Different privacy-preserved images produced by SN network after self-learning and SN with loss function backpropagation.
Jimaging 10 00059 g011
Table 1. Comparison of accuracy, misidentification rate, and F1 score of RLLFPR with currently available face privacy-preserving recognition methods on three datasets (Unit/%).
Table 1. Comparison of accuracy, misidentification rate, and F1 score of RLLFPR with currently available face privacy-preserving recognition methods on three datasets (Unit/%).
Database LFW Celeba HDU
Methods Privacy
Protection
Accuracy Misidentification F1 Accuracy Misidentification F1 Accuracy Misidentification F1
Original faceNo99.980.1099.8399.690.0899.7399.970.1099.75
AESYes000
Eigenface [40]No98.930.4598.8898.410.5298.4180.3918.9079.61
Arnold [2]Yes99.820.1099.6698.330.1598.1390.579.0689.74
DifferentialYes83.738.7783.6181.689.3181.98<50>50<50
privacy [33]
PartialFace [35]Yes99.800.1599.6798.730.2497.9199.340.7898.94
RLLFPRYes99.930.1399.6798.770.1198.6799.580.5999.23
Table 2. The average correlation of RGB channels of the original face, AES, Arnold, and RLLFPR images in the horizontal, vertical, and diagonal directions.
Table 2. The average correlation of RGB channels of the original face, AES, Arnold, and RLLFPR images in the horizontal, vertical, and diagonal directions.
MethodHorizontalVerticalDiagonal
Original face0.99230.99160.9829
Arnold0.86080.75480.9279
AES−0.0039−0.0032−0.0007
RLLFPR−0.09290.0326−0.0795
Table 3. Comparison of accuracy and error rates of different methods.
Table 3. Comparison of accuracy and error rates of different methods.
MethodPSNRUACI
Noise14.1636.24
Arnold11.5353.93
AES8.7775.27
RLLFPR11.8050.61
Table 4. Questionnaire results.
Table 4. Questionnaire results.
Purpose of the QuestionTypes of Design Problems (Two Groups)Experimental GroupControl Group (AES)
Fuzziness testGroup 1 dissimilarity3.513.55
(0–4, 0 clear, 4 no similarity)
Group 2 match20.90%25.35%
(choose 1 from 4, measure accuracy)
Table 5. Fuzziness and safety analysis of RLLFPR, AES, Arnold, and noise methods.
Table 5. Fuzziness and safety analysis of RLLFPR, AES, Arnold, and noise methods.
MethodPrivacy ProtectionIdentification after ProtectionFuzziness2Fuzziness3KeyReversibilityRevocability
ArnoldYesYes0.40–0.700.996NoReversibleRevocable
AESYesNo0.60–0.900.999YesReversibleRevocable
NoiseYesYes0.20–0.400.992NoIrreversibleIrrevocable
Original faceNo00NoIrrevocable
RLLFPRYesYes0.80–0.950.997NoIrreversibleRevocable
Table 6. Comparison of three blurriness tests for privacy-preserved images produced by SN after self-learning and SN network with loss function backpropagation.
Table 6. Comparison of three blurriness tests for privacy-preserved images produced by SN after self-learning and SN network with loss function backpropagation.
Fuzziness1Fuzziness2Fuzziness3
SN (loss function backpropagation)2.110.75–0.950.983
SN (self-learning)3.510.80–0.950.997
Table 7. Performance comparison of ResNet50 and DenseNet121 as backbone networks in RLLFPR (Unit/%).
Table 7. Performance comparison of ResNet50 and DenseNet121 as backbone networks in RLLFPR (Unit/%).
AccuracyMisidentification RateF1
ResNet5099.930.1399.67
DenseNet12199.910.0799.67
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Huang, Y.; Wu, Z.; Chen, J.; Xiang, H. Privacy-Preserving Face Recognition Method Based on Randomization and Local Feature Learning. J. Imaging 2024, 10, 59. https://doi.org/10.3390/jimaging10030059

AMA Style

Huang Y, Wu Z, Chen J, Xiang H. Privacy-Preserving Face Recognition Method Based on Randomization and Local Feature Learning. Journal of Imaging. 2024; 10(3):59. https://doi.org/10.3390/jimaging10030059

Chicago/Turabian Style

Huang, Yanhua, Zhendong Wu, Juan Chen, and Hui Xiang. 2024. "Privacy-Preserving Face Recognition Method Based on Randomization and Local Feature Learning" Journal of Imaging 10, no. 3: 59. https://doi.org/10.3390/jimaging10030059

APA Style

Huang, Y., Wu, Z., Chen, J., & Xiang, H. (2024). Privacy-Preserving Face Recognition Method Based on Randomization and Local Feature Learning. Journal of Imaging, 10(3), 59. https://doi.org/10.3390/jimaging10030059

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop