Next Article in Journal
Parametric Analysis of Outrigger Systems for High-Rise Buildings with Different Geometric Shapes
Previous Article in Journal
Exposure to Oxycide™ Disinfectant in Dental Practices: Occupational Symptoms and Risk Factors
Previous Article in Special Issue
A Deep Learning Framework for High-Frequency Signal Forecasting Based on Graph and Temporal-Macro Fusion
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Optical Coherence Tomography-Based Fingerprint Presentation Attack Detection via Multi-Tissue Features Contribution Analysis

College of Computer Science and Technology, Zhejiang University of Technology, Hangzhou 310023, China
*
Author to whom correspondence should be addressed.
Appl. Sci. 2025, 15(10), 5642; https://doi.org/10.3390/app15105642
Submission received: 26 April 2025 / Revised: 14 May 2025 / Accepted: 16 May 2025 / Published: 19 May 2025
(This article belongs to the Special Issue Advances in Neural Networks and Deep Learning)

Abstract

:
Currently, unsupervised anti-spoofing methods suffer from low accuracy and susceptibility to irrelevant factors. In order to solve these problems, this paper analyzes the contribution of different fingertip tissue structures to anti-spoofing tasks and proposes an unsupervised anti-spoofing method based on the weighted contribution of tissue structures. Unsupervised OCT anti-spoofing methods suffer from weak robustness and lack comprehensive exploration of sub-dermal structure features. The proposed method introduces quantified weights of fingertip tissue contributions with self-attention through the Shapley value. The module can amplify crucial fingertip features and extracts more key fingertip information, thereby improving the anti-spoofing performance.

1. Introduction

Fingerprints, as one of the most common physiological features of the human body, are widely used in determining individual identity information [1]. With the development of technology, fingerprint recognition technology has been applied in various fields such as fingerprint payment, fingerprint unlocking, and Internet of Things applications [2,3,4]. However, the collection process of epidermis fingerprints is easily affected by the external environment, and it is difficult to ensure the imaging quality of epidermis fingerprints, which leads to a decrease in the accuracy of fingerprint recognition [5]. In addition, epidermis fingerprint traces can easily be replicated and the cost of forgery is low.
The emergence of 3D fingerprint collection technology can alleviate the impacts brought by the above problems. Finger skin is a layered tissue, which is divided into the stratum corneum, the epidermis layer, and the dermis layer; at the junction of the dermis and the epidermis, there exists a papillary layer [6], which is the source of the fingerprint structure. In the epidermis layer, there are spiral ducts that connect the sweat glands in the dermis to the sweat pores on the surface [7]. Therefore, sub-dermal fingerprints have strong resistance to deformation and damage, and can make up for defects in the imaging quality of epidermis fingerprints. Meanwhile, by jointly considering complex features such as sweat glands and the layered structure of finger skin, the difficulty of forging samples can be greatly increased.
Optical coherence tomography (OCT) technology [7] can achieve collection of the three-dimensional structure of the finger. It can not only detect the epidermis fingerprints on the finger but also detect the micron-level complex structures of multiple tissue layers within the skin. The 3D fingerprint dataset obtained through OCT technology provides a solid data foundation for the development of more efficient and accurate fingerprint anti-spoofing methods.
Due to the diversity of materials used to make spoof samples, anti-spoofing methods use zero-shot [8] or one-class sample [9,10,11] approaches to construct deep autoencoders. Only training on bona fide fingerprints can prevent attacks from unknown spoof materials.
Currently, OCT fingerprint anti-spoofing methods can be divided into feature-based and learning-based methods. In the early days, for bona fide and spoof samples, the method of manually annotating image features was usually chosen for anti-spoofing tasks. The annotated features include sweat pores [12], physiological features [13], etc. Bossen et al. [7] carried out anti-spoofing by judging the presence or absence of the papillary layer. Liu et al. [14] used sweat gland annotations to distinguish spoof samples. Gutierrez et al. [15] used the difference in the deformation amount between internal and external fingerprints to mark bona fide fingerprints. Ding et al. [16] proposed an anti-spoofing method based on one-class support vector machines. Darlow et al. first proposed a fully automatic detection method based on highly separable double-layer features, which detected spoof fingerprints with 100% accuracy [17]. The above methods are based on manual feature selection, meaning they require a large amount of work and are inefficient when dealing with large-scale datasets.
Learning-based methods can automatically extract the features of input samples, significantly improving the efficiency and generalization ability of anti-spoofing methods. They are mainly divided into two categories: supervised and unsupervised learning. Supervised learning methods learn relevant features through the labels of bona fide and spoof samples for judgment. Chugh et al. [6] cut OCT images into image patches and input them into a convolutional binary classification network for in-depth detection to judge the authenticity of samples. Nogueira et al. [18] verified the superiority of CNN in supervised fingerprint anti-spoofing methods. Sun et al. [19] proposed the ISAPAD dual-branch network to detect spoof samples by learning the internal structure and overall features of fingerprints. However, in the existing methods, the samples used for training and testing come from the same data domain, and the detection performance for spoof samples with unknown data distributions remains unknown. It can be seen that these supervised methods rely on the limited attack samples in the training set and still have the problem of insufficient generalization ability when facing unknown attack samples. In contrast, in view of the diversity and unknown nature of spoof samples, unsupervised methods that only train on bona fide samples are more valuable in practical applications. Liu et al. [9] proposed an autoencoder method based on activation map reconstruction. Without any spoof samples for reference, it not only has better generalization ability but also alleviates the data dependence problem caused by the significant differences between different spoof samples. This method can achieve a true positive rate of 96.59% without using spoof fingerprints. Zhang et al. [20] found that there is noise interference in OCT scanned images. Therefore, by extracting the frequency information in the OCT spatio-temporal domain, decomposing the images into four different groups of frequency-domain information, and using wavelet decomposition to remove irrelevant interference information, they improved the performance by 81.89%. Liu et al. [21] introduced a prototype reconstruction module into the denoising autoencoder, used the sparse addressing method to find the prototype most similar to the input image in the memory module, and proposed a representation consistency constraint to make the reconstructed features closer to bona fide fingerprints, thereby increasing the reconstruction distance of spoof samples and reducing the error rate by 40.07%. Zhang et al. [22] proposed a brand-new unified representation model. By introducing an anti-spoofing task during the fingerprint reconstruction process, the intermediate features are connected and interact across layers, enabling both the reconstruction task and the anti-spoofing method to achieve good performance.
However, through the study of attention heat maps, it can be found that for fingerprints, the key information for judging the authenticity of samples is often concentrated in specific areas of the fingerprints. Most existing methods are based on the overall image. Therefore, there is still an overfitting phenomenon in the anti-spoofing task, a large number of sample features need to be annotated, and the reconstruction distance based on images cannot distinguish between bona fide and spoof samples.
In view of the diversity and unknown nature of spoof materials, a more robust anti-spoofing method is needed to ensure the security of fingerprint recognition systems. In this paper, a self-built 1310 nm OCT system is used to collect three-dimensional structural information on fingertips, and an unsupervised fingerprint anti-spoofing method that only requires bona fide samples is proposed, which can more accurately judge the authenticity of samples. Subsequently, a CNN is used to conduct ablation comparison experiments on different sub-dermal tissue structures of the fingers, and the contribution of these structures in the fingerprint anti-spoofing task is analyzed, providing strong evidence for the method proposed in this paper. The specific research contents are as follows:
In order to further analyze the importance of different tissue structures of bona fide samples for the anti-spoofing task, starting from the characteristics of multiple local features and overall features of OCT fingerprints that can be used for anti-spoofing, this paper explores the correlation between the internal features of fingerprints and their contribution to the anti-spoofing task, and improves the performance of the anti-spoofing method by formulating relevant analysis criteria. Through the correlation study of all local 3D features and relevant anti-spoofing tasks, the degree of dependence of each task on fingerprint features is quantitatively analyzed. The model training process not only relies on a single 3D feature of bona fide samples but also comprehensively considers the relationships among the overall features, local features, and anti-spoofing tasks, so as to improve the anti-spoofing performance of the model. Finally, based on the visualization segmentation results of bona fide samples and spoof samples, starting from the data distribution of the input samples, the spoof score criteria are established by using the mean and variance of the Gaussian distribution of features, which are used for the detection of spoof samples.
The remainder of this article is organized as follows. In Section 2, we review some related work about OCT technology and deep learning visual model architectures. We introduce our proposed unsupervised anti-spoofing method in Section 3. Section 4 shows related experimental results and compares the performance of our method with other unsupervised models. In Section 5, we summarize the research methods of this paper, describe deficiencies and improvement directions related to the current research, and explain the next steps pertaining to our work.

2. Related Work

This section first introduces the OCT collection system and related technologies adopted in the experiment. Subsequently, it describes the characteristics of OCT images of bona fide samples and those of spoof samples, and provides an overview of the existing OCT fingerprint anti-spoofing methods and their performance evaluation metrics. Finally, it describes the deep learning theory involved in the method proposed in this paper.

2.1. OCT Fingerprint Collection Device System

OCT technology can achieve high-resolution acquisition of three-dimensional volume data of fingertips. This technology utilizes the principle of low-coherence interference of near-infrared light (780 nm–2526 nm). By measuring the optical reflection of internal tissues, it realizes non-invasive and high-resolution cross-sectional imaging and has been widely applied in the biomedical field [7]. As shown in Figure 1, in an OCT scanner, the light beam is divided into a reference arm and a sample arm, each containing a mirror unit. If the reflected light from the two arms is within the coherence distance, an interference pattern will be generated. This interference pattern represents the depth profile at a point, which is called an A-scan. A 2D image at a fixed depth is called an enface slice; laterally combining a series of A-scans along a line can provide a cross-sectional pattern, which is called a B-scan image; stacking multiple B-scan images together can represent the volume data structure of the scanned fingerprint. In the data acquisition stage, the finger must be placed tightly on a thin glass sheet and kept still to ensure the stability of the system during data collection.
Our laboratory has independently developed an OCT scanning system to collect the tissue information from fingertips, as shown in Figure 1. It uses a broadband source with a central wavelength of 1310 nm and a bandwidth of 85.6 nm. The light can penetrate through the stratum corneum and the epidermis layer until reaching the dermis layer, and the maximum scanning depth can reach 4.64 mm beneath the skin. Each collection of fingertip volume data contains 1400 or 700 B-scan images. The size of each B-scan image is 1800 × 500, as shown in Figure 2.

2.2. OCT Dataset

The corresponding relationship between the B-scan images acquired by OCT technology and the real sub-dermal structures of fingertips is shown in Figure 3. The outermost layer, which comes into daily contact with the external environment, is called the epidermis. On the outer surface of the epidermis, the tough part with friction ridge structures is known as the stratum corneum. The part beneath the epidermis is the dermis, which provides most of the skin’s elasticity and strength. Sweat glands are distributed between the epidermis and the dermis, usually in a spiral or oval shape, and are responsible for secreting sweat to regulate body temperature. The special structure located between the stratum corneum and the dermis is the viable epidermis junction layer. This layer has unique undulating features, called the internal fingerprint, which provides a basis for identifying an individual’s identity.
The bona fide samples used in the experiment were collected from 160 white-collar and blue-collar workers aged between 20 and 60, with a ratio of approximately 1:4. All the volunteers who participated in the fingerprint collection were Chinese. A total of 2,039,800 slices were collected. Due to the issue of the slicing angle during scanning, the first 400 B-scan slices of each volume of data do not have a relatively complete tissue structure. Therefore, in the experiment, only the last 1000 scanned slices of each volume of data were selected as the source of the bona fide sample dataset.
In this paper, 21 kinds of materials are used to make the spoof sample dataset, with a total of 525 groups of spoof sample data. The spoof sample dataset is roughly divided into two categories. The first category uses a single material, which can only ensure that the surface threads of the samples are similar to those of bona fide fingerprints. The materials used include epoxy resin, Elmer’s glue, handmade white glue, conductive silica gel, etc. The other category of spoof samples has a double-layer tissue structure similar to that of the OCT images of bona fide samples, which increases the difficulty of the anti-spoofing task. This type of spoof sample is composed of eight kinds of materials, including transparent silica gel, flesh-colored silica gel, etc.
As shown in Table 1, for spoof samples made from fingerprints of different materials, the OCT scanning results also vary. Ordinary single-layer silica gel samples have a simple organizational structure. For the double-layer film fingerprint structure, as different materials are added during the production of spoof samples, the refractive index of light during OCT acquisition is different in the two materials. Therefore, an obvious double-layer structure will appear, posing a great challenge to current anti-spoofing methods. Polystyrene microspheres and conductive silver particles in the fingerprint film can simulate the sub-dermal sweat gland tissue of bona fide fingerprints, which also makes it difficult to distinguish between bona fide and spoof samples.
In response to the need for feature extraction of the organizational structure of OCT fingerprints, this paper establishes an OCT layer segmentation dataset. A total of 1050 B-scan images are randomly selected from bona fide samples to form the layer segmentation dataset, and manual annotations are carried out for the glass layer, stratum corneum layer, and viable epidermis junction layer of the OCT fingerprint samples.
In this paper, each B-scan image in the dataset is divided into three regions, which are parts (c), (d), and (e) of Figure 4, respectively. The open-source image annotation software used is Labelme v4.5.12. As an image annotation tool, it is often used in the field of image segmentation. After using marking points of different colors to encircle the contours of each layer of the tissue structure in the B-scan image, Labelme will output the final annotation result. Among them, the area surrounded by the green markings is regarded as the glass layer, and the viable epidermis junction layer is marked as the red area. When the stratum corneum is not attached to the glass layer, the upper edge of the green area is regarded as the glass layer, as shown by the upper boundary of the green area in the figure. The part between the green and red areas is regarded as the stratum corneum.

2.3. Evaluation Metrics for the Anti-Spoofing Performance of OCT Fingerprints

The evaluation metrics for OCT fingerprint anti-spoofing include the ofllowing: the area under the curve (AUC), the equal error rate (EER), and the bona fide presentation classification error rate (BPCER).
AUC calculates the area under the curve. The curve usually refers to the receiver operator characteristic curve (ROC) or the precision–recall curve (PRC). In this paper, the area under the ROC curve is used. The value of AUC ranges from 0 to 1. The closer it is to 1, the better the model performance. When the AUC value is 1, it means that the model’s prediction is completely correct. The ROC curve provides a graphical analysis interface. The abscissa represents the false positive rate (FPR), and the ordinate represents the true positive rate (TPR), which is also called the recall. FPR indicates the proportion of spoof samples that are wrongly judged as bona fide samples among all spoof samples, and TPR indicates the proportion of bona fide samples that are correctly judged as bona fide samples among all bona fide samples. As shown in Figure 5, the shaded part represents the AUC, and the intersection point of the green dotted line and the blue ROC curve represents the value when FPR is equal to TPR. Therefore, the performance of the classification method can be intuitively measured through the ROC curve. The closer the ROC curve is to the upper-left corner, the better the model performance. In this case, the true positive rate is higher and the false positive rate is lower. Therefore, the AUC curve is an important metric for evaluating the performance of binary classification models, which can reflect the model’s prediction ability and accuracy.
The EER is one of the important performance metrics in binary classification problems, especially in the fields of biometric recognition, security verification, etc. Specifically, in the ROC, the EER refers to the error rate corresponding to the point where the FPR is equal to the FNR, that is, when the probabilities are equal. The lower the EER, the better the performance of the classifier, indicating that under the same error rate, the classifier can correctly identify more positive samples. The value range of EER is [0, 1], and the closer it is to 0, the better the performance.
The BPCER is an evaluation metric for measuring the performance of fingerprint recognition methods. It represents the proportion of bona fide identities that are wrongly rejected, that is, the proportion of cases where the system wrongly identifies a user’s bona fide biometric feature as an illegal feature when the user provides a bona fide biometric feature. It is usually used in conjunction with the APCER (Attack Presentation Classification Error Rate) to comprehensively evaluate the performance of the system. Corresponding to the APCER, the lower the BPCER, the better, indicating that the proportion of the system wrongly rejecting legitimate users is lower—that is, the system has a higher recognition accuracy for legitimate users.

2.4. Deep Learning Visual Model Architecture

Deep learning achieves efficient representation and learning of large-scale data by constructing multi-layer neural network models and has achieved breakthrough results in fields such as image recognition, natural language processing, and speech recognition. The hierarchical structure enables the model to extract features of the data level by level, thus realizing the modeling and prediction of complex data. Its advantage lies in the ability to automatically learn more advanced and abstract feature representations from the raw data without the need for manual feature design. Through optimization methods such as backpropagation algorithms and gradient descent, the automatic adjustment and learning of the parameters of the neural network model are realized. These optimization methods enable the model to be trained efficiently on large-scale datasets and gradually optimize the model parameters during the training process, continuously improving the prediction performance of the model. This end-to-end training method enables deep learning models to adapt to various complex tasks. With the optimization of deep learning models by researchers, a series of innovative model architectures have emerged, such as convolutional neural network (CNN), recurrent neural network (RNN), generative adversarial network (GAN), etc.

2.4.1. Convolutional Neural Network (CNN)

In the field of computer vision, image feature extraction and classification have always been the main focus of research attention. Convolutional Neural Network (CNN) is one of the mainstream methods. It updates network parameters through the backpropagation algorithm. During the training process, it can automatically extract effective features from images and perform classification. Therefore, it is widely used in tasks such as image classification, object detection, and segmentation. Compared with the perceptron, the uniqueness of CNN lies in its processing of local regions of images through convolution kernels and the sharing of weights at different levels. The network can learn more abstract and high-level feature representations from the data. This structure makes CNN more suitable for processing image data and better able to capture the spatial locality and hierarchical structure in images.
The earliest CNN structure was the LeNet model proposed by LeCun, which is mainly used for handwritten character recognition [23]. Limited by computing resources, its performance was still not ideal, so it did not attract much attention. The AlexNet network proposed by Alex et al. in 2012 achieved amazing results in the ImageNet image classification competition, successfully demonstrating the powerful ability of CNN in image recognition tasks [24]. The VGGNet network [25] improved by the University of Oxford based on AlexNet not only reduced the number of parameters but also greatly enhanced the model’s performance. The residual network proposed by He et al. [26] significantly improved the problems of gradient vanishing and explosion. The object detection algorithm RCNN proposed by Ross Girshick et al. [27] combined deep learning with traditional object detection algorithms for the first time, achieving a high detection accuracy. CNN has achieved remarkable results in fields such as image segmentation [28], object tracking [29,30], and vehicle driving [31,32]. CNN is mainly composed of the following parts:
  • Convolutional layer. This is an important hierarchical structure in the CNN model, mainly used to process gridded data such as images, voices, and text. The core operation is the convolution operation, which extracts features from the input data by sliding the convolution kernel. In the convolution operation, the convolution kernel slides over the input data with a certain stride. Each time, it performs a dot-product operation with a local region of the input data and sums the results to obtain a new value as an element of the output feature map. Through multiple convolution operations, the high-level features of the input data can be gradually extracted. In the convolutional layer, if the input tensor feature map is X, the convolution kernel required for the convolution operation is W, m and n represent the row and column indices of the convolution kernel, respectively, and the feature map S obtained after convolution can be obtained through Equation (1).
    S ( i , j ) = ( X W ) ( i , j ) = m n X ( i + m , j + n ) · W ( m , n )
    The convolutional layer uses multiple different convolution kernels, and each convolution kernel is used to extract different features. In addition, the convolutional layer also includes a bias term and an activation function, which are used to introduce non-linearity and increase the expressive ability of the model. Parameters such as stride and padding control the process of the convolution operation. The stride determines the distance that the convolution kernel slides on the input data, and padding adds additional values (usually 0) at the boundaries of the input data, so as to keep the size of the output feature map consistent with that of the input during the convolution operation.
  • Pooling layer. The pooling layer is placed after the convolutional layer and is used to reduce the size of the feature map, lower the model complexity, and increase the translation invariance of the model. It is used for downsampling the feature map. Its main functions include the following: (a) Size reduction: The pooling operation aggregates (takes the maximum or average value of) the local regions of the feature map and compresses the information within the regions, thus reducing the size of the feature map. This helps to reduce the number of model parameters and the amount of computation, and improves the computational efficiency of the model. (b) Translation invariance: The pooling operation aggregates local regions, making the feature map have certain invariance under transformations such as translation and rotation. This helps the model learn more robust feature representations and improves the generalization ability of the model for input data. Common pooling operations include max-pooling and average-pooling. Max-pooling selects the maximum value within the local region as the aggregation result, while average-pooling selects the average value within the local region as the aggregation result.
  • Activation function. Its main role is to convert the input of neurons into output and introduce non-linearity, enabling the neural network to learn and represent complex non-linear relationships. If the activation function is missing in the model, the output of the network will be limited to the category of linear functions. This means that the complexity of the network will be greatly restricted, making it difficult for the network to effectively learn and simulate more complex and diverse data types, such as images, videos, and voices. The activation function allows the neural network to learn and map the complex non-linear functional relationships in reality, greatly enhancing the expressive ability of CNNs, and thus enabling more advanced learning and reasoning tasks. Commonly used activation functions include Sigmoid, Tanh, Relu, etc.
  • Fully connected layer. Also known as the dense-connected layer, it is located in the last few layers of the CNN and is used to map the features extracted by the convolutional layer into the final output results. In classification tasks, the fully connected layer is usually used to convert the feature map into a class probability distribution or class labels. Each neuron in the fully connected layer is connected to all neurons in the previous layer, and each neuron has weight and bias parameters, which are used to adjust the weights and biases of the input features. The fully connected layer performs linear combinations and non-linear transformations on the features extracted by the convolutional layer, thus obtaining higher-level abstract feature representations for better distinguishing different classes. Therefore, the Softmax function can be used in the last layer for classification to ensure that the sum of the probabilities of all classes in the obtained output results is 1.

2.4.2. Transformer

The Transformer was proposed by Vaswani et al. [33] and has achieved great success in the field of natural language processing. Especially in machine translation tasks, it has become an important benchmark model. The model is based on the self-attention mechanism. By weighted aggregation of information at different positions in the input sequence, it realizes the modeling of long-distance dependencies, thus better capturing the semantic information in the sequence. This model has excellent global feature extraction ability. It can extract features for upstream tasks through pre-training and be applied to different transfer learning tasks. In the field of computer vision, research has found that the Transformer can achieve performance comparable to that of CNN. At the ICLR 2021 conference, Vision Transformer (ViT) proposed by Alexey et al. [34] successfully applied the Transformer model to image classification tasks. The ViT model adopts a brand-new image representation method. It divides the input image into small image patches, converts each image patch into a sequence form, and then inputs these sequences into the Transformer encoder for processing. Currently, the Transformer has achieved remarkable results in both natural language processing and visual classification [35,36], detection [37,38], and segmentation [39,40].
As shown in Figure 6, on the left is the encoder of the Transformer. The encoder is stacked by multiple encoder layers with the same structure. Each encoder layer includes a multi-head attention mechanism and a feed-forward neural network. The encoder is responsible for mapping the input sequence to a high-dimensional feature representation space. On the right is the decoder architecture. The decoder is stacked by multiple decoder layers with the same structure. Compared with the encoder, in addition to the multi-head attention mechanism and the feed-forward neural network, each decoder layer also has an additional encoder–decoder attention mechanism. The decoder is responsible for mapping the feature representation output by the encoder to the target sequence. The main components of the Transformer model include the following:
  • Attention mechanism. The Transformer model mainly adopts the multi-head self-attention mechanism. The calculation of self-attention requires Query, Key, and Value. After obtaining the values of QKV, the output can be calculated:
    A t t e n t i o n ( Q , K , V ) = S o f t m a x Q K T d k V ,
    where d k represents the vector dimension, that is, the number of columns of Q and K. To prevent issues such as gradient vanishing or gradient explosion caused by overly large inner products, the result of the dot-product between the query and the key is divided by d k .
    The multi-head attention mechanism is composed of multiple self-attention mechanisms. The input sequence is first mapped to the representations of query, key, and value through linear transformation, and then the calculations of multiple attention heads are performed in parallel, respectively.
  • Feed-forward neural network. It usually consists of two linear transformation layers and an activation function. These two linear transformation layers are called fully connected layers or hidden layers, and there is no limit to the number of hidden layers. An activation function is usually added between each hidden layer, and the most commonly used activation function is the ReLU function. The role of the feed-forward neural network is to perform non-linear transformation and feature extraction on the representation of each position, thereby helping the model better understand and process the information of the input sequence and improving the generalization ability of the model. The calculation process of the feed-forward neural network [33] is as follows.
    F F d ( x ) = W 2 ( R e L U ( W 1 x + b 1 ) ) + b 2
    where x is the output of the hidden layer, W 1 and W 2 are the learning weight matrices, and b 1 and b 2 are trainable bias parameters.
  • Residual connections and layer normalization. Residual connections are part of the network structure ResNet proposed by Kaiming He et al. [26]. They are used to alleviate the problems of gradient vanishing or explosion during the training of multi-layer networks. By introducing the residual connection structure, the performance of the model can be prevented from decreasing due to the deepening of the network. Meanwhile, to mitigate the gradient vanishing problem, improve the model’s optimization efficiency, and enable the model to converge quickly, layer normalization is applied to the output results. Layer normalization normalizes each sample independently in each layer, rather than normalizing the entire batch.

2.4.3. Unsupervised Neural Network Model

Unsupervised neural networks are a machine learning paradigm that extracts useful features and information from input data without labels or target outputs. They mainly extract relevant attributes by discovering the structural features of the data themselves and use these features to establish non-linear relationships among parameters. Unsupervised learning can find patterns in datasets with unlabeled inputs, such as statistical features, correlations, or categories. These patterns are then used for classification, decision-making, or generating new data. Unsupervised learning plays a significant role in tasks such as data clustering, anomaly detection, dimensionality reduction, and data generation.
Unsupervised neural network models include those based on image feature learning, clustering algorithms, contrastive learning, etc. Autoencoders are one of the most commonly used unsupervised neural networks in image feature learning. They extract information from the original data through the encoder and use these feature maps to reconstruct the original data, thereby learning the features of the images. This type of network has a wide range of applications in areas such as image denoising [40,41], feature learning [41], and style transfer. Generative Adversarial Network (GAN) is a neural network generation model proposed by Goodfellow et al. [42], which consists of two parts: a generator and a discriminator. Zhou et al [43] use generator to generate bona fide samples, while the discriminator is trained to distinguish between generated samples and bona fide samples. During training, these two modules compete to capture the semantic features of bona fide samples. Chen et al. [44] proposed a way of generating labels with outlier label smoothing regularization, assigning uniformly distributed labels to unlabeled images to improve the performance of the base model. Kim et al. [45] separately mixed the reference features with the original features in the discriminator and then performed regularization to improve the image quality. In view of the fact that the distribution of traditional features is more complex than the Gaussian distribution, Zhang et al. [46] matched real-world data by precisely matching the empirical cumulative distribution functions of image features, and this has been effectively verified in various style transfer tasks.
In summary, unsupervised methods can avoid time-consuming and labor-intensive manual data annotation. They have strong generalization ability, can make full use of unlabeled data, and explore undiscovered data feature distributions and correlations. Thus, they have broad application prospects in the field of computer vision.

2.5. Summary of This Section

This section introduces the relevant theoretical and technical background. Firstly, the OCT fingerprint acquisition system and the OCT bona fide and spoof sample dataset are introduced. Secondly, the evaluation metrics for the anti-spoofing performance of OCT are presented. Finally, the deep learning models used in this paper are introduced, including convolutional neural networks, Transformers, and unsupervised neural network models.

3. Proposed Method

The OCT images of sub-dermal fingertip tissue have a distinct layering structure, with texture shape and other features between layers forming the basis for the uniqueness and reliability of each person’s fingerprint identity security. However, different regions in the OCT fingerprint data may exhibit different features, and their contribution to the task of anti-spoofing needs further exploration. By analyzing these differences, we can optimize the performance of anti-spoofing methods. Based on the analysis of the contribution of the anti-spoofing features of the sub-dermal fingertip tissue, this section proposes an anti-spoofing method based on quantified contribution weights and introduces a variational attention module into the reconstruction module to detect spoof samples using the distribution of test data. The experimental results show that this method can improve the anti-spoofing performance of the model and achieve high levels of performance across multiple metrics.

3.1. Basic Architecture

Through reconstruction experiments on OCT images, the contributions of different parts to the anti-spoofing model are explored. Then, according to the experimental results, the segmentation network fuses the anti-spoofing contribution of the quantized feature vectors with high-dimensional vectors. As shown in Figure 7, in this section, U-Net is used to separate the regions between key layers of the image from the background region, thus extracting the high-dimensional features of each layer structure. These features are then combined with the reconstruction model based on the contribution.
In this section, three segmentation networks with different structures are used to separate the OCT fingertip images of different tissue structures. Through pre-training, the training parameters of different segmentation networks are obtained. When the pre-processed data are input, the images are, respectively, input into the two branches of the segmentation network and the reconstruction network. By fixing the parameters of the segmentation network, it can be free from the interference of the parameters of the reconstruction network. Finally, the high-dimensional semantic features of the model are obtained by reconstructing the image with an auto-encoder added with a variational attention module. The weights obtained from the analysis are fused with the tissue features in the segmentation model to obtain a feature map, and then reconstruction is carried out to obtain the final output result.

3.2. Experiments on Fingertip Tissue Feature Mask

In OCT fingerprint scan images, whether it is the stratum corneum or the viable epidermis junction layer, their structures exhibit obvious continuity. However, sampling with a convolutional neural network (CNN) can easily disrupt this structure. To explore the relationship between each tissue structure in OCT fingerprint images and pixel reconstruction, this subsection uses the mask autoencoder (MAE) to perform the image reconstruction task. MAE is an expandable self-supervised learning model based on Vision Transformer (ViT) [38]. Its design concept is derived from the cloze-learning strategy in the natural language model BERT, and it is trained using a masking mechanism. The encoder randomly masks a sequence of image patches obtained by randomly cutting the input image, and the model is trained on the remaining unmasked image patches to extract key feature representations. The masking strategy prompts the encoder to learn the overall context information of the image, rather than being limited to local pixel features. As shown in Figure 8, the decoder uses these random high-dimensional features to restore the entire image, including the masked parts.
This model can successfully restore the general outline of the overall image, and the reconstructed masked parts also have good coherence. However, the performance of MAE is interfered with by both the randomness of masking and the large amount of black background in the images, which results in a large amount of information failing to be effectively recovered when restoring the sample tissue regions, thus losing crucial semantic information. Especially for the glass layer and the viable epidermis layer, only parts with less pixel information such as the stratum corneum are reconstructed well. The contours between different tissue structure layers in OCT images are unique, and the semantic information of the images is not continuous in the vertical direction. As can be seen in the figure, the reconstructed layers do not have obvious curved contours but are straight lines. Therefore, although the MAE model performs excellently in extracting the structural information from the dataset, its effect is not ideal for extracting the key semantic information in the tissues.

3.3. Ablation Experiment on Tissue Texture of Bona Fide and Spoof Samples

It is inferred from the MAE masking experiment that both structural features and pixel features have an impact on the fingerprint reconstruction results. However, the contribution of each part still cannot be determined. In this subsection, the methods of mixing and ablation of the texture structures of bona fide and spoof samples are adopted to clarify the specific contributions of each tissue structure in terms of anti-spoofing. Since the supervised method has a high degree of accuracy and can effectively reveal the internal characteristics of the data through bona fide and spoof labels, the supervised method is combined for the ablation experiment. The core of the experimental design lies in gradually removing or adding specific tissue structures to observe their impacts on the final anti-spoofing effect.
Given that the high accuracy rate of supervised anti-spoofing methods can often reach 100%, this paper adopts a binary classification network constructed by ordinary convolutional and fully connected layer modules as the baseline network for pre-training. The experimental training set uses 5000 bona fide samples and 5000 spoof samples for pre-training. For the test set, this paper randomly selects 100 bona fide samples after tissue segmentation and fuses them with 2000 spoof samples. Moreover, the segmented tissue structures are randomly selected from the 100 bona fide fingerprints and integrated into the 2000 spoof sample images. The fusion results are shown in Figure 9. During the entire training process, the training epoch is set to 100, the batchsize of images is 15, the learning rate is 0.001, the optimizer is Adam, and the experimental results are shown in Table 2.
In the table, the anti-spoofing accuracy is the correct rate of the model in judging authenticity after fusing the labels. It can be seen that different regions contribute differently to the anti-spoofing method. Compared with the stratum corneum and the glass layer, which can only provide partial anti-spoofing contributions, the dermal tissue often provides more information. In order to convert the experimental results into quantifiable weights for the model, this subsection proposes using the Shapley value to calculate the contribution weights and thereby determine the impact of different tissue regions on the anti-spoofing results.
The Shapley value is used in game theory to distribute profits or costs, and it was proposed by the American economist Lloyd Shapley in 1953. The Shapley value is used in evaluating the contribution of each participant to the overall outcome and distributing profits or costs on this basis [47]. For a multi-person cooperative project, the Shapley value measures the average contribution of each participant in all possible combinations, as well as the marginal contribution when joining or withdrawing from the cooperation. This is obtained by considering all possible permutations and combinations of participants and calculating the impact of each participant’s joining or withdrawing on the final result. Suppose there are n key regional features τ = ( τ 1 , τ 2 , τ 3 , , τ n ) in the anti-spoofing task, and its Shapley value is obtained through Equation (4):
S i ( v ) = S N , i S ( | S | 1 ) ! ( n | S | ) ! n ! v ( S ) v ( S { i } )
where S i ( v ) is the Shapley value of the fingertip tissue, N represents the set of all tissue features, S is a subset of the tissue features, and v ( S ) represents the value provided in the cooperation, that is, the accuracy obtained in Table 2. Finally, in this paper, the proportion of the contribution of each key feature is roughly the following: viable epidermis layer:glass layer:stratum corneum = 4:1:0.2.
To quantitatively measure the contribution of different tissue layers to anti-spoofing, the Shapley value was employed as a statistical attribution mechanism. The calculation consists of two steps: Supervised ablation experiments were performed using the segmentation masks of bona fide samples to selectively remove specific tissue structures and measure the resulting anti-spoofing accuracy on a validation set, as shown in Table 2. Based on these accuracy scores, the Shapley values of each tissue region were calculated to reflect their respective contribution weights. The resulting ratio is viable epidermis layer:glass layer:stratum corneum = 4:1:0.2, which was then used to guide feature weighting in the subsequent unsupervised model. It should be noted that in our framework, the Shapley value serves as a statistical estimate of tissue contribution and is not directly tied to spatial feature attribution like in gradient-based saliency or integrated gradients. While gradient-based methods compute pixel- or region-level attention maps in visual models, our use of Shapley values focuses on quantifying the contribution of tissue types rather than locations, offering a global and interpretable foundation for weight assignment.

3.4. Weighted Module Based on Contribution Weight

Through the above experiments, it can be concluded that in the anti-spoofing task, the contributions of the features of different tissue regions inside the fingertip to the anti-spoofing effect show significant differences. In order to quantify this difference, this subsection proposes a Shapley-weighted contribution module (SWCM). By combining the Shapley value with the ablation experiment of the test accuracy of the supervised anti-spoofing task, the final contribution of each tissue structure of the fingertip in the anti-spoofing task can be calculated and converted into a corresponding weight, that is, the weighting coefficient.
In order to apply the obtained experimental results to optimize the anti-spoofing method, this paper uses three independent U-Net networks for pre-training and fixes the parameters of the segmentation network. U-Net is a convolutional neural network architecture used for semantic segmentation tasks, which was proposed by Ronneberger et al. in 2015 [48]. Its main feature stems from its U-shaped network structure, which has a symmetric encoder and decoder. The middle part is a downsampling path composed of pooling layers and convolutional layers, which is used to extract high-level features of the image. The upsampling module is composed of transposed convolutional layers and skip connections, which are used to restore the feature maps extracted by the encoder to the segmentation results at the original resolution. Through U-Net networks with different parameters, the structural information of each region can be separated. Subsequently, the feature maps of each tissue part obtained by the U-Net segmentation network are multiplied by the previously calculated weighting coefficients, and the outputs of all fingertip tissues are accumulated and finally divided by the number of tissues, so as to obtain the structural and semantic information extracted from the segmentation task. The specific weighting method can be expressed as
X k = F e n + F 1 w 1 + F 2 w 2 + F 3 w 3 N k .
This method uses the Swin Transformer as the backbone network for feature extraction. F e n is the multi-scale feature task map of the image obtained through the Swin Transformer. w 1 , w 2 , and w 3 are the weights converted from the proportions of the task contribution degrees of different fingertip sample tissues obtained previously. As shown in Figure 10, the obtained multi-scale feature task map and the feature map after fusing the segmentation fusion features are further processed for the fused features through a self-attention residual network. Through the attention mechanism, the dependencies between the fused features can be captured, and the problem of gradient vanishing can be effectively alleviated.
In this subsection, the key features are amplified according to the contribution ratio, and a contribution-based feature map is obtained accordingly. In this way, the model can focus more on the features with higher contribution, especially those in the dermal region.

3.5. Variational Self-Attention Module

Both the stratum corneum and the viable epidermis connection layer in the OCT images of bona fide samples have good continuity. However, for spoof samples, there are edge pixel breaks and many areas with scattered pixel disorder. Therefore, this subsection proposes a variational self-attention module (VAM). Since traditional autoencoders do not have a continuous latent space distribution but a fixed latent vector, they are relatively limited in data generation and representation learning. Therefore, in this paper, by introducing a Bayesian network, the fused vector X k of the latent space after extracting bona fide samples is mapped to a data probability distribution, allowing the model to learn the continuous representation of bona fide samples. At the same time, through the re-parameterization method, the obtained latent space vector X k passes through a fully connected layer to obtain a mean vector μ and a variance vector σ 2 , and a random variable ϵ is sampled in N ( μ , σ 2 ) within the range [ 0 , 1 ] , as shown in Equation (6). As shown in Figure 11, in this subsection, the hidden vector z is reconstructed through the mean vector μ and the variance vector σ 2 , and then long-range attention dependencies are established to solve the problem that the receptive field of the convolutional kernel of ordinary autoencoders is limited.
z = μ + σ ϵ ,   ϵ N [ 0 , 1 ]
Since z is obtained through re-parameterization based on the distribution of bona fide samples, after model training, the higher the similarity between the distribution of the sample data input during testing and the distribution of bona fide samples, the higher the probability that the scanned image is bona fide. Assume that the obtained weight-fused vector X k conforms to the prior distribution P ( X k ) :
P ( X k ) = z P ( X k | z ) P ( z )   d z .
The parameter ϑ in the distribution of the latent space P ( z ) can be solved using the maximum-likelihood method of logarithmic functions:
ϑ X k = arg max ϑ i = 1 n log P ( X k ) .
However, if we want to generate a high-dimensional vector X k similar to the weight-fused features, it is necessary to enumerate and sum all z variables. This requires an excessive amount of computation and is difficult to implement. Therefore, the KL divergence is introduced to calculate the posterior probability distribution Ψ ( z | X k ) :
  K L ( Ψ ( z | X k ) | | P ( z | X k ) ) =   log P ( X k ) + K L ( Ψ ( z | X k ) | | P ( z ) ) E z Ψ ( z | X k ) .
By maximizing the lower bound of the maximum likelihood of the logarithmic function and simultaneously minimizing the approximate data distribution Ψ ( z | X k ) , the true posterior probability P ( z | X k ) of X k can be approximated. The calculation formula for the KL divergence is
  K L ( N ( μ , σ 2 ) | | N ( 0 , 1 ) ) =   1 2 π σ 2 e ( X k μ ) 2 / 2 σ 2 log 1 2 π σ 2 e ( X k μ ) 2 / 2 σ 2 1 2 π e X k 2 / 2 d x .
As can be seen from Equation (10), the value of the KL divergence must be greater than zero. By using the KL divergence value as the loss function of the high-dimensional fused vector, the approximate distribution Ψ ( z | X k ) can be continuously made to approach the bona fide data input distribution. After obtaining the mean μ and variance σ after training is completed, the deception score of the input sample is calculated, thus enabling the detection of spoof samples.
Ultimately, leveraging the contribution-based weighted features in conjunction with the learnable weight matrices W Q , W K , and W V , three feature projections, namely, Query (Q), Key (K), and Value (V), are computed:
  M H S A ( Q , K , V ) =   S o f t m a x Q ( L N ( X k ) ) K ( L N ( X k ) ) T d k V ( L N ( X k ) )
where Q ( X k ) , K ( X k ) , and V ( X k ) are the self-attention matrix projections of X k , LN represents the normalization layer, and MHSA represents multi-head self-attention. A residual connection is applied to the obtained results X o u t and X k to prevent the model from overfitting. Then, through the upsampling part, the feature map obtained from the residual connection is restored layer by layer to the same resolution as the original image, resulting in a reconstructed image X o u t p u t :
X o u t p u t = F D ( X W )
where X W is the output of the self-attention module, and F D is the upsampling encoder, which is composed of a CNN, a ReLU activation layer, and layer normalization. It is worth noting that most spoof samples often only exhibit a single-layer structure and lack key information in the viable epidermis layer and the stratum corneum part. Therefore, by enhancing the attention to the viable epidermis layer area and introducing reconstruction loss, the model can identify spoof samples more accurately. In this subsection, the Euclidean distance is used as the criterion for judging the reconstruction distance:
d = 1 n i = 1 n ( y ^ i y i ) 2 .

3.6. Model Training

The dual-branch method proposed in this section requires simultaneously training the U-Net segmentation network and the reconstruction network, making the training process relatively complex. This subsection will introduce the entire training process in detail. When training the segmentation network, only bona fide images and labels are used to train the three-way segmentation network. At this time, the parameters of the reconstruction network are frozen, and its loss function is
l o s s C E = i = 1 n P ( x t r ) log ( P ( x p r e d ) )
where P ( x t r u e ) represents the true probability distribution of the input image, and P ( x a r r ) is the probability distribution predicted by the model. After the segmentation network is trained, the image needs to be input into the reconstruction network again. At this time, the parameters of the segmentation network are frozen. The semantic information obtained by the encoder and the spatial structure information of the segmentation network are used to learn the bona fide data distribution. Finally, the self-attention layer is used to establish the attention distribution weights in the reconstructed latent space feature map.
Since the network uses the KL divergence loss as a constraint, the reconstructed data distribution conforms to a Gaussian distribution. Finally, the final reconstructed input is obtained through the decoder, and its loss functions are l o s s k l and l o s s a e , respectively. The overall loss function of the entire process is shown in Equation (15).
  l o s s k l = ( μ 2 + σ 2 log σ 2 1 ) / 2   l o s s a e = X o u t p u t X i n p u t 2   l o s s t o t a l = l o s s k l + l o s s a e
During the training process, the segmentation network and the reconstruction network are trained alternately. To ensure that the reconstruction network can accurately capture structural information, before each training of the reconstruction network, the segmentation network needs to be trained multiple times to ensure that the weighting module can correctly amplify the features of key areas.

3.7. The Reasoning Process and Spoof Scores

OCT images have rich details. Using only the reconstruction image error as the criterion for abnormal judgment often leads to relatively low accuracy. To accurately identify the input images, the reconstruction distances of the fused features in the high-dimensional space and those of the input images themselves are comprehensively considered. By accumulating the reconstruction distances, the differences in reconstruction distances among different samples are further enlarged, and the overlapping areas of the reconstruction distances between bona fide samples and spoof samples are effectively reduced, thus obtaining a clearer decision-making boundary. The anti-spoofing judgment result, SpoofScore, is expressed as
S p o o f S c o r e = ( F D i s > s 1 ) · ( R D i s > s 2 )
where FDis and RDis represent the reconstruction distance of the fused features in the high-dimensional space and the reconstruction distance of the image, respectively. During the training phase, this method only uses bona fide samples for training. Therefore, in the model inference phase, the boundary of the reconstruction distance of bona fide samples, s2, and the reconstruction distance of the fused features in the high-dimensional space, s1, can be obtained. According to the requirements of specific datasets, the threshold of the reconstruction distance can be flexibly adjusted. When both the reconstruction distance of the input image and the reconstruction distance of the features exceed the preset thresholds, it is determined as a bona fide sample; conversely, if either of the distances is lower than the threshold, the image will be regarded as a spoof sample. To ensure a fair and consistent evaluation of the unsupervised method, thresholds s 1 and s 2 in Equation (16) were calibrated using a validation set. Given the inherent randomness and susceptibility to overfitting in unsupervised training, the validation set serves as a reference to select optimal parameters. Specifically, multiple threshold combinations were evaluated, and the ones achieving the best anti-spoofing performance on the validation set were chosen to serve as s 1 and s 2 during inference.
By constraining the reconstruction scores of the autoencoder, the model can learn the features of bona fide samples, and then distinguish between bona fide and spoof samples. As shown in the density graph in Figure 12, when the overlapping area of the reconstruction distances between bona fide samples and spoof samples is too large, the obtained reconstruction scores cannot effectively distinguish between bona fide and spoof samples. Therefore, by introducing a new reconstruction distance score in the high-dimensional space, the model can obtain more high-dimensional information about bona fide samples, further expanding the partition intervals of bona fide samples and spoof samples.
In this subsection, the data distribution of bona fide input images is introduced, and the Gaussian probability density function in the data distribution, along with the mean and variance of the distribution, are intended to be used to comprehensively consider the spoof probability of input samples.
Given a training sample x, after segmenting the sample, it is cropped and reshaped into N feature maps x i with the same resolution. Through the training of the reconstruction module, the mean μ and variance σ of the data distribution of the bona fide sample dataset can be obtained. According to the density graph of the reconstruction distance of the fused features in Figure 12, the reconstruction distance d r can be regarded as conforming to a Gaussian distribution. Therefore, in this subsection, the probability density function G d r of the bona fide sample is expressed as
G d r = 1 2 π σ 2 e ( x i μ ) 2 / 2 σ 2 .
The closer the test sample is to the bona fide sample, the higher the similarity between the high-dimensional data distribution of the test sample and G d r . In this subsection, using the similarity between the distribution mean μ and variance σ of the test sample and those of the bona fide sample, two types of anomaly scores are defined to determine the authenticity of the test sample. Equation (18) represents the determination of the spoof score based on the probability density function of the bona fide sample,
S p o o f S c o r e G = 1 G d r ( 1 N R t e s t )
where SpoofScoreG represents the probability that the test sample is a bona fide sample and R t e s t represents the input of the test sample. The smaller the value of SpoofScoreG, the greater the probability that the test sample is a bona fide sample. Therefore, when the probability is higher than a predetermined threshold, the model is detected as a spoof sample; otherwise, it is judged as a bona fide sample. Regarding the distribution mean μ and variance σ , another type of spoof score detection is designed in this section:
μ S c o r e = | μ t e s t μ m i n | | μ m a x μ m i n | σ S c o r e = | σ t e s t σ m i n | | σ m a x σ m i n |
where μ t e s t and σ t e s t are the mean and variance of the data distribution of the test samples, and they are compared with the predetermined thresholds. Samples exceeding the thresholds are judged as spoof samples. To normalize the spoof scores, this section uses a normalization method to convert the scores into probabilities and combines the mean and variance to detect spoof samples,
S p o o f S c o r e μ σ = 1 S ( μ S c o r e < ϕ 1 ) · S ( σ S c o r e < ϕ 2 )
where S ( · ) is the sign function. ϕ 1 and ϕ 2 are predefined thresholds. When both μ S c o r e and σ S c o r e are less than the thresholds, the test sample is judged as a spoof sample.
This paper proposes a new classification constraint method. As shown in Figure 13a, represents the existing unsupervised anti-spoofing boundary discrimination method, which usually distinguishes spoof samples by constraining the reconstruction distance of bona fide samples. However, this method may cause some spoof samples with similar structures and too small reconstruction distances to be wrongly aligned with bona fide samples, and it is unable to constrain spoof samples, thus interfering with the anti-spoofing method. Therefore, as shown in Figure 13b, this chapter combines the spoof scores of SpoofScore μ σ and SpoofScoreG. By imposing data distribution constraints on the high-dimensional features of input samples, it can make the scattered spoof samples and bona fide samples align with different data distributions, respectively, thereby generating a clearer discrimination boundary in the input samples.

4. Experimental Results

In this subsection, AUC, EER, and B P C E R 10 are selected as metrics to evaluate the performance of the anti-spoofing method that incorporates the contribution weights. A total of 10,000 B-scans from blue-collar and white-collar workers are randomly selected for the training set. In total, 1000 bona fide samples and 1000 spoof samples are selected for the validation set and 50,000 bona fide samples and 50,000 spoof samples are selected for the test set. The algorithm model proposed in this paper is trained and tested using the Pytorch framework. During the entire training process, the training epoch is set to 100, the batchsize of images is 15, the learning rate is 0.001, and the optimizer is Adam. All experiments are run on the Linux operating system, and the GPU used is the NVIDIA RTX3090.

4.1. Comparison Experiment of Unsupervised Models

In this subsection, the most advanced, currently available unsupervised models are compared. The experimental results are shown in Table 3. The following is a brief introduction to each of the comparative models.
AE [44] is an unsupervised learning algorithm that learns a compressed representation of the input data and then reconstructs the original data. It consists of two parts: an encoder and a decoder. VAE [45] is a variant of AE. By introducing a regularization term, the encoded latent variables are made to follow a given prior distribution, and then the reconstruction of data samples is achieved by learning the latent distribution of the data. MemAE [46], on the other hand, is another variant of AE. It introduces a memory module to store the effective information of the input images, and in the anomaly detection task, this module is used to store the features of bona fide samples. OCPAD [9] is the first to use AE for fingerprint anti-spoofing. Through the representation of the latent space and the calculation of the image reconstruction loss, it can still achieve a relatively high anti-spoofing accuracy without relying on attack samples. PPAD [20] is a method proposed based on the MemAE model. Through a sparse addressing memory network and representational consistency constraints, it obtains a more compact boundary for distinguishing bona fide and spoof samples. U_OCFR [21] unifies the anti-spoofing task and the segmentation task. By introducing intermediate features during the fingerprint reconstruction process for the anti-spoofing task, it enables both the segmentation task and the anti-spoofing task to achieve good performance.
Compared with judging the reconstruction distance of fixed latent space feature maps, learning the average data distribution of bona fide samples is relatively stable. Since most spoof samples contain obvious glass plate information and a small amount of fingerprint film information, the pixel values in most areas are relatively small, and the features of the viable epidermis layer are not obvious. Therefore, the image reconstruction error of spoof samples is much smaller than that of bona fide samples. By magnifying the key areas in the image and establishing a Gaussian distribution of bona fide sample data, due to the rich semantic information and high pixel value distribution, the analysis of high-dimensional data becomes more effective.
In terms of the AUC value, the method proposed in this paper is 0.9688, which exceeds that of the state-of-the-art U_OCFR model. The EER is 0.0668, showing a slight decrease compared with the most novel method, and there is a relatively obvious decrease in B P C E R 10 . Overall, the model designed in this paper has improvements in various key metrics, demonstrating that introducing the anti-spoofing contribution information of fingerprint tissues is beneficial to the improvement of the performance of anti-spoofing tasks. The proposed method exhibits a relatively higher standard deviation in AUC compared to other models. This is primarily due to the more complex spoof score computation strategy employed, which increases the sensitivity of the model to threshold selection. Since AUC measures model performance across a range of threshold values, this sensitivity results in greater variability. In contrast, metrics such as EER and BPCER are computed using fixed thresholds—hence, they show much lower standard deviations. Despite this higher variance, our method achieves a higher upper bound of AUC, indicating strong potential. In real-world deployments, performance can be further optimized by carefully calibrating the thresholds.

4.2. Ablation Experiment

The ablation experiments in this subsection include ablation experiments on the weighted module of the segmentation network, the variational attention module, and the spoof score to demonstrate the effectiveness of the proposed method.

4.2.1. Ablation Experiment for Weighted Module of Segmentation Network

To quantify the contribution of the contribution weights to the recognition of spoof samples, this experiment tests the anti-spoofing performance of the fusion module with the introduced contribution weights w i , the fusion module without weights, the single-path fusion module, and the single-reconstruction branch. Among them, the single-path fusion module refers to placing the segmentation results of different levels of the input bona fide samples on the same segmentation label for training. The fusion module based on the contribution weights w i proposed in this paper uses three segmentation networks to segment different tissue structures of the samples, respectively. In this experiment, the SpoofScore μ σ is used as the anomaly detection score.
The experiment shows the experimental results of the weighted fusion in Table 4. From the results, it can be found that when the dual-branch network is used, since the segmentation task can prompt the reconstruction task to learn the data distribution of the training samples, the error rate of the model is reduced. Separately training the labels can reduce the interference factors brought by other regions, enabling the reconstruction branch to learn features more robustly. Research shows that fusing the contribution weights into the model can play the role of an attention prior, and introducing this prior condition helps the model to better judge spoof samples. In the single-branch segmentation experiment, an additional semantic segmentation branch was introduced to segment the OCT images into three target tissue regions. The segmentation results in the testing phase were used as spatial masks to extract region-specific representations, which were then passed to the unsupervised reconstruction branch for spoof score calculation. However, when the segmentation quality during testing is suboptimal, the generated spatial masks may fail to accurately localize tissue-specific areas. This results in the feature weighting module amplifying features unrelated to the target tissues. Consequently, it may distort the representations of bona fide samples and misclassify them as spoof samples, leading to a higher BPCER value.

4.2.2. Ablation Experiment for Variational Attention Module

In order to analyze the effectiveness of introducing the variational attention module, this subsection conducts ablation experiments on the reconstruction branch network, and takes the multi-scale feature reconstruction network of the weighted segmentation network, including three branches as the baseline model.
As can be seen from Table 5, simply adding the self-attention module does not significantly improve the anti-spoofing performance of the model. The area under the AUC curve only increases by 2.4% (compared with 0.9028). However, adding the variational structure can greatly enhance the model’s performance, and the equal error rate decreases by 37% (compared with 0.1865). This may be because the changes in the key information of the input samples are within a certain range rather than a fixed value, and the variational structure can separate the bona fide samples from spoof samples by introducing the data distribution of bona fide samples. In addition, this subsection also finds that by combining residual connections and the self-attention module, long-distance dependencies can be established, and the continuous features between the structural contours of bona fide samples can be well utilized.

4.2.3. Ablation Experiment for Anomaly Score Judgment

In Table 4 and Table 5, SpoofScore μ σ is used as the anomaly judgment score, which improves the anomaly judgment performance of the model. This subsection conducts ablation experiments on the previously designed different anti-spoofing scores, as shown in Table 6. It can be seen from the table that σ S c o r e has poor performance as an anomaly judgment score. However, by introducing μ S c o r e to complement σ S c o r e and jointly conduct anomaly judgment on the input images of the model, it can be observed that the performance of SpoofScore μ σ is better than simply calculating the distribution mean μ and variance σ of the test samples. When the attack presentation classification error rate (APCER) is 10%, SpoofScore μ σ is 2% higher than the single anomaly score μ S c o r e and 4.5% higher than σ S c o r e in terms of the AUC. Compared with the judgment based on the reconstruction distance, the spoof score proposed in this section shows a stronger discriminative ability in the anti-spoofing task.

4.3. Model Complexity and Inference Cost

To evaluate the practical deployability of the proposed method, we conducted a comparative analysis of model complexity and computation. As shown in Table 7, our method requires 5.624 million parameters and 9.31 GFLOPs, which is efficient than PPAD (6.15 M parameters, 25.6 GFLOPs) and moderately more complex than UOCFR (3.87 M parameters, 8.65 GFLOPs). The results indicate that our model provides a favorable trade-off between computational cost and detection performance, making it suitable for real-time OCT applications.

4.4. Statistical Significance Analysis

To validate the statistical significance of our improvements over baseline models, we conducted paired t-tests on AUC, EER, and BPCER10 across 10 repeated trials. As shown in Table 8, our method achieves significantly better performance than U-OCFR in all three metrics (p < 0.05), and outperforms PPAD in BPCER (p = 0.0420). These results support the robustness and consistency of the proposed method.

4.5. Generalization to Unseen Spoof Materials

To further assess the model’s ability to handle unseen spoof types, several representative out-of-distribution spoof materials were tested individually. The AUC and EER results are presented in Figure 14.

5. Conclusions

Due to the unique structural and pixel features of sub-dermal fingertip tissue, if all the effective regions in the OCT image area are simply used for overall feature extraction, important and secondary features will often be treated equally, which has an adverse impact on anti-spoofing performance. Therefore, this paper proposes an anti-spoofing method based on the contribution of sub-dermal tissue structure at the fingertips. Firstly, it analyzes the relationship between structure and pixels as well as existing problems during MAE reconstruction. Then, through ablation experiments on different fingertip tissues using the Shapley value, the quantitative contribution of different tissues to anti-spoofing are obtained and transformed into weights for anti-spoofing tasks. Finally, by combining the variational structure with the attention mechanism and using the high-dimensional fused data distribution as the criterion for anomaly detection, the anti-spoofing performance of the model is effectively improved.
This study investigated unsupervised anti-spoofing methods of OCT images based on fingertip tissue structures. Although it achieved good results in terms of test performance on existing fingerprint datasets, in-depth research is still needed concerning the following aspects:
  • This study explored the impact of the sub-dermal tissue structure segmentation task and the semantic feature reconstruction task on anti-spoofing performance from the perspective of correlation. In the future, tasks related to speckle noise suppression and image style learning can be introduced to obtain more efficient OCT fingerprint feature representations and establish a correlation analysis standard.
  • The experimental data in this study were trained and tested within a single domain, and research and discussion on cross-domain datasets was not included. Future research can explore the task correlation between data from different domains through transfer learning methods to further improve the generalization of the method.
In the future, OCT fingerprint data from diverse populations and devices will be collected to expand the coverage of spoof samples and enhance the model’s domain generalization and adaptability.

Author Contributions

Conceptualization, Z.L. and Y.L.; data curation, J.Q.; formal analysis, J.J.; funding acquisition, Y.L.; methodology, Z.L. and H.F.; project administration, Y.L.; resources, J.Q.; software, Z.L. and J.Q.; supervision, Y.L.; validation, J.J. and H.F.; visualization, J.J. and H.F.; writing—original draft, Z.L., J.J., J.Q., H.F. and Y.L. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Natural Science Foundation of China (62376251, 62076220) and the Medical Health Science and Technology Project of Zhejiang Province (2024KY881).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are available on request from the corresponding author.

Acknowledgments

We would like to express our deepest gratitude to all those who have contributed to the completion of this research and the writing of this paper.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Cappelli, R.; Ferrara, M.; Maltoni, D. Minutia cylinder-code: A new representation and matching technique for fingerprint recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2010, 32, 2128–2141. [Google Scholar] [CrossRef] [PubMed]
  2. Hemalatha, S. A systematic review on Fingerprint based Biometric Authentication System. In Proceedings of the 2020 International Conference on Emerging Trends in Information Technology and Engineering (ic-ETITE), Vellore, India, 24–25 February 2020; pp. 1–4. [Google Scholar]
  3. Ali, G.; Dida, M.; Sam, A. A Secure and Efficient Multi-Factor Authentication Algorithm for Mobile Money Applications. Future Internet 2021, 13, 299. [Google Scholar] [CrossRef]
  4. Kanchana, S. Fingerprint based biometric authentication in IoT for resolving security challenges. Int. J. Res. Anal. Rev 2018, 5, 1000–1003. [Google Scholar]
  5. Bontrager, P.; Roy, A.; Togelius, J.; Memon, N.; Ross, A. Deepmasterprints: Generating masterprints for dictionary attacks via latent variable evolution. In Proceedings of the 2018 IEEE 9th International Conference on Biometrics Theory, Applications and Systems (BTAS), Redondo Beach, CA, USA, 22–25 October 2018; pp. 1–9. [Google Scholar]
  6. Chugh, T.; Jain, A.K. OCT fingerprints: Resilience to presentation attacks. arXiv 2019, arXiv:1908.00102. [Google Scholar]
  7. Bossen, A.; Lehmann, R.; Meier, C. Internal fingerprint identification with optical coherence tomography. IEEE Photonics Technol. Lett. 2010, 22, 507–509. [Google Scholar] [CrossRef]
  8. Liu, Y.; Stehouwer, J.; Jourabloo, A.; Liu, X. Deep tree learning for zero-shot face anti-spoofing. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 4680–4689. [Google Scholar]
  9. Liu, H.; Zhang, W.; Liu, F.; Wu, H.; Shen, L. Fingerprint presentation attack detector using global-local model. IEEE Trans. Cybern. 2021, 52, 12315–12328. [Google Scholar] [CrossRef]
  10. Engelsma, J.J.; Jain, A.K. Generalizing fingerprint spoof detector: Learning a one-class classifier. In Proceedings of the 2019 International Conference on Biometrics (ICB), Crete, Greece, 4–7 June 2019; pp. 1–8. [Google Scholar]
  11. Santoro, A.; Bartunov, S.; Botvinick, M.; Wierstra, D.; Lillicrap, T. Meta-learning with memory-augmented neural networks. In Proceedings of the International Conference on Machine Learning, PMLR, New York, NY, USA, 19–24 June 2016; pp. 1842–1850. [Google Scholar]
  12. Marcialis, G.L.; Roli, F.; Tidu, A. Analysis of fingerprint pores for vitality detection. In Proceedings of the 2010 20th International Conference on Pattern Recognition, Istanbul, Turkey, 23–26 August 2010; pp. 1289–1292. [Google Scholar]
  13. Marasco, E.; Sansone, C. Combining perspiration-and morphology-based static features for fingerprint liveness detection. Pattern Recognit. Lett. 2012, 33, 1148–1156. [Google Scholar] [CrossRef]
  14. Liu, M.; Buma, T. Biometric mapping of fingertip eccrine glands with optical coherence tomography. IEEE Photonics Technol. Lett. 2010, 22, 1677–1679. [Google Scholar] [CrossRef]
  15. da Costa, H.S.G.; Maxey, J.R.; Silva, L.; Ellerbee, A.K. Evaluation of fingerprint deformation using optical coherence tomography. In Proceedings of the Optical Elastography and Tissue Biomechanics, SPIE, San Francisco, CA, USA, 1–2 February 2014; Volume 8946, pp. 71–78. [Google Scholar]
  16. Ding, Y.; Ross, A. An ensemble of one-class SVMs for fingerprint spoof detection across different fabrication materials. In Proceedings of the 2016 IEEE International Workshop on Information Forensics and Security (WIFS), Abu Dhabi, United Arab Emirates, 4–7 December 2016; pp. 1–6. [Google Scholar]
  17. Darlow, L.N.; Webb, L.; Botha, N. Automated spoof-detection for fingerprints using optical coherence tomography. Appl. Opt. 2016, 55, 3387–3396. [Google Scholar] [CrossRef]
  18. Nogueira, R.F.; de Alencar Lotufo, R.; Machado, R.C. Fingerprint liveness detection using convolutional neural networks. IEEE Trans. Inf. Forensics Secur. 2016, 11, 1206–1213. [Google Scholar] [CrossRef]
  19. Sun, H.; Zhang, Y.; Chen, P.; Wang, H.; Liang, R. Internal structure attention network for fingerprint presentation attack detection from optical coherence tomography. IEEE Trans. Biom. Behav. Identity Sci. 2023, 5, 524–537. [Google Scholar] [CrossRef]
  20. Zhang, W.; Liu, H.; Liu, F. Fingerprint presentation attack detection by learning in frequency domain. In Proceedings of the 2021 IEEE 2nd International Conference on Pattern Recognition and Machine Learning (PRML), Chengdu, China, 16–18 July 2021; pp. 183–189. [Google Scholar]
  21. Liu, Y.P.; Zuo, W.; Liang, R.; Sun, H.; Li, Z. Prototype-guided autoencoder for OCT-based fingerprint presentation attack detection. IEEE Trans. Inf. Forensics Secur. 2023, 18, 3461–3475. [Google Scholar] [CrossRef]
  22. Zhang, W.; Liu, H.; Liu, F.; Ramachandra, R. A Uniform Representation Learning Method for OCT-based Fingerprint Presentation Attack Detection and Reconstruction. arXiv 2022, arXiv:2209.12208. [Google Scholar]
  23. LeCun, Y.; Boser, B.; Denker, J.; Henderson, D.; Howard, R.; Hubbard, W.; Jackel, L. Handwritten digit recognition with a back-propagation network. Adv. Neural Inf. Process. Syst. 1989, 2, 396–404. [Google Scholar]
  24. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. Adv. Neural Inf. Process. Syst. 2012, 25. [Google Scholar] [CrossRef]
  25. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  26. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  27. Girshick, R.; Donahue, J.; Darrell, T.; Malik, J. Rich feature hierarchies for accurate object detection and semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 24–27 June 2014; pp. 580–587. [Google Scholar]
  28. Long, J.; Shelhamer, E.; Darrell, T. Fully convolutional networks for semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 3431–3440. [Google Scholar]
  29. Zhu, T.; Hiller, M.; Ehsanpour, M.; Ma, R.; Drummond, T.; Reid, I.; Rezatofighi, H. Looking beyond two frames: End-to-end multi-object tracking using spatial and temporal transformers. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 45, 12783–12797. [Google Scholar] [CrossRef]
  30. Gu, F.; Lu, J.; Cai, C. RPformer: A robust parallel transformer for visual tracking in complex scenes. IEEE Trans. Instrum. Meas. 2022, 71, 1–14. [Google Scholar] [CrossRef]
  31. Zheng, T.; Huang, Y.; Liu, Y.; Tang, W.; Yang, Z.; Cai, D.; He, X. Clrnet: Cross layer refinement network for lane detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 898–907. [Google Scholar]
  32. Li, P.; Wang, P.; Berntorp, K.; Liu, H. Exploiting temporal relations on radar perception for autonomous driving. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 17071–17080. [Google Scholar]
  33. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is all you need. Adv. Neural Inf. Process. Syst. 2017, 30, 6000–6010. [Google Scholar]
  34. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An image is worth 16x16 words: Transformers for image recognition at scale. arXiv 2020, arXiv:2010.11929. [Google Scholar]
  35. Wang, W.; Xie, E.; Li, X.; Fan, D.P.; Song, K.; Liang, D.; Lu, T.; Luo, P.; Shao, L. Pyramid vision transformer: A versatile backbone for dense prediction without convolutions. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 568–578. [Google Scholar]
  36. Wang, W.; Xie, E.; Li, X.; Fan, D.-P.; Song, K.; Liang, D.; Lu, T.; Luo, P.; Shao, L. Pvt v2: Improved baselines with pyramid vision transformer. Comput. Vis. Media 2022, 8, 415–424. [Google Scholar] [CrossRef]
  37. Dai, Z.; Cai, B.; Lin, Y.; Chen, J. Up-detr: Unsupervised pre-training for object detection with transformers. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 19–25 June 2021; pp. 1601–1610. [Google Scholar]
  38. Wang, W.; Cao, Y.; Zhang, J.; Tao, D. Fp-detr: Detection transformer advanced by fully pre-training. In Proceedings of the International Conference on Learning Representations, Vienna, Austria, 3–7 May 2021. [Google Scholar]
  39. Yuan, Y.; Fu, R.; Huang, L.; Lin, W.; Zhang, C.; Chen, X.; Wang, J. Hrformer: High-resolution vision transformer for dense predict. Adv. Neural Inf. Process. Syst. 2021, 34, 7281–7293. [Google Scholar]
  40. Gu, J.; Kwon, H.; Wang, D.; Ye, W.; Li, M.; Chen, Y.H.; Lai, L.; Chandra, V.; Pan, D.Z. Multi-scale high-resolution vision transformer for semantic segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 12094–12103. [Google Scholar]
  41. Zheng, S.; Lu, J.; Zhao, H.; Zhu, X.; Luo, Z.; Wang, Y.; Fu, Y.; Feng, J.; Xiang, T.; Torr, P.H.; et al. Rethinking semantic segmentation from a sequence-to-sequence perspective with transformers. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 19–25 June 2021; pp. 6881–6890. [Google Scholar]
  42. Goodfellow, I.J.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative adversarial nets. Adv. Neural Inf. Process. Syst. 2014, 27. [Google Scholar]
  43. Zhou, C.; Paffenroth, R.C. Anomaly detection with robust deep autoencoders. In Proceedings of the 23rd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, Halifax, NS, Canada, 13–17 August 2017; pp. 665–674. [Google Scholar]
  44. Chen, X.; Ding, M.; Wang, X.; Xin, Y.; Mo, S.; Wang, Y.; Han, S.; Luo, P.; Zeng, G.; Wang, J. Context autoencoder for self-supervised representation learning. Int. J. Comput. Vis. 2024, 132, 208–223. [Google Scholar] [CrossRef]
  45. Kim, J.; Choi, Y.; Uh, Y. Feature statistics mixing regularization for generative adversarial networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 11294–11303. [Google Scholar]
  46. Zhang, Y.; Li, M.; Li, R.; Jia, K.; Zhang, L. Exact feature distribution matching for arbitrary style transfer and domain generalization. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 8035–8045. [Google Scholar]
  47. Li, J.; Kuang, K.; Li, L.; Chen, L.; Zhang, S.; Shao, J.; Xiao, J. Instance-wise or class-wise? A tale of neighbor shapley for concept-based explanation. In Proceedings of the 29th ACM International Conference on Multimedia, Chengdu, China, 20–24 October 2021; pp. 3664–3672. [Google Scholar]
  48. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In Proceedings of the Medical Image Computing and Computer-Assisted Intervention–MICCAI 2015: 18th International Conference, Munich, Germany, 5–9 October 2015; Proceedings, part III 18. Springer: Cham, Switzerland, 2015; pp. 234–241. [Google Scholar]
  49. Lee, G.; Jung, M.; Song, M.; Choo, J. Unsupervised anomaly detection of the gas turbine operation via convolutional auto-encoder. In Proceedings of the 2020 IEEE International Conference on Prognostics and Health Management (ICPHM), Detroit, MI, USA, 8–10 June 2020; pp. 1–6. [Google Scholar]
  50. Kingma, D.P.; Welling, M. Auto-ENCODING Variational Bayes. arXiv 2022, arXiv:1312.6114. [Google Scholar]
  51. Gong, D.; Liu, L.; Le, V.; Saha, B.; Mansour, M.R.; Venkatesh, S.; Hengel, A.v.d. Memorizing normality to detect anomaly: Memory-augmented deep autoencoder for unsupervised anomaly detection. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 1705–1714. [Google Scholar]
Figure 1. Image of 1310 nm OCT fingerprint acquisition system.
Figure 1. Image of 1310 nm OCT fingerprint acquisition system.
Applsci 15 05642 g001
Figure 2. Diagram of 1310 nm OCT fingerprint dataset.
Figure 2. Diagram of 1310 nm OCT fingerprint dataset.
Applsci 15 05642 g002
Figure 3. Correspondence between B-scan and fingertip tissue structure.
Figure 3. Correspondence between B-scan and fingertip tissue structure.
Applsci 15 05642 g003
Figure 4. Fingerprint segmentation examples. (a) Bona fide fingerprint B-scan image. (b) Labeled image. (c) Glass layer label. (d) Stratum corneum label. (e) Viable epidermis label. The colored lines in (b) represents the layer surfaces.
Figure 4. Fingerprint segmentation examples. (a) Bona fide fingerprint B-scan image. (b) Labeled image. (c) Glass layer label. (d) Stratum corneum label. (e) Viable epidermis label. The colored lines in (b) represents the layer surfaces.
Applsci 15 05642 g004
Figure 5. ROC curve and AUC area example. The blue solid line represent the EER curve.
Figure 5. ROC curve and AUC area example. The blue solid line represent the EER curve.
Applsci 15 05642 g005
Figure 6. Transformer structure [33].
Figure 6. Transformer structure [33].
Applsci 15 05642 g006
Figure 7. Reconstruction anti-spoofing method based on contribution weights.
Figure 7. Reconstruction anti-spoofing method based on contribution weights.
Applsci 15 05642 g007
Figure 8. Reconstruction for pixel block mask. (a) Original OCT image. (b) Masked image. (c) Reconstructed image.
Figure 8. Reconstruction for pixel block mask. (a) Original OCT image. (b) Masked image. (c) Reconstructed image.
Applsci 15 05642 g008
Figure 9. Ablation study on tissue texture. (a) Fuse the labels of the viable epidermis layer. (b) Fuse the labels of the glass layer. (c) Fuse the labels of the stratum corneum.
Figure 9. Ablation study on tissue texture. (a) Fuse the labels of the viable epidermis layer. (b) Fuse the labels of the glass layer. (c) Fuse the labels of the stratum corneum.
Applsci 15 05642 g009
Figure 10. Tissue feature weighted structure based on segmentation network.
Figure 10. Tissue feature weighted structure based on segmentation network.
Applsci 15 05642 g010
Figure 11. Variational self-attention module.
Figure 11. Variational self-attention module.
Applsci 15 05642 g011
Figure 12. Density graph of pixel reconstruction distance and density graph of fused-feature reconstruction distance after Gaussian fitting.
Figure 12. Density graph of pixel reconstruction distance and density graph of fused-feature reconstruction distance after Gaussian fitting.
Applsci 15 05642 g012
Figure 13. Spoofness score determination constraint. (a) Reconstruction distance constraint. (b) Data distribution constraint.
Figure 13. Spoofness score determination constraint. (a) Reconstruction distance constraint. (b) Data distribution constraint.
Applsci 15 05642 g013
Figure 14. Anti-spoofing performance and visual representations of different out-of-distribution spoof materials.
Figure 14. Anti-spoofing performance and visual representations of different out-of-distribution spoof materials.
Applsci 15 05642 g014
Table 1. Fake fingerprint example.
Table 1. Fake fingerprint example.
Fingerprint ExampleActual ImageOCT Scan Image
Bona fide fingerprintApplsci 15 05642 i001Applsci 15 05642 i002
Single-layer structure, flesh-colored silica gelApplsci 15 05642 i003Applsci 15 05642 i004
Single-layer structure, fingerprint filmApplsci 15 05642 i005Applsci 15 05642 i006
Ultra-thin fingerprint film, transparent silica gelApplsci 15 05642 i007Applsci 15 05642 i008
Double-layer fingerprint film, flesh-colored silica gelApplsci 15 05642 i009Applsci 15 05642 i010
Double-layer ultra-thin fingerprint film, transparent silica gelApplsci 15 05642 i011Applsci 15 05642 i012
Sweat gland simulated fingerprint film, transparent silica gelApplsci 15 05642 i013Applsci 15 05642 i014
Table 2. Ablation study on tissues texture of bona fide and spoof samples.
Table 2. Ablation study on tissues texture of bona fide and spoof samples.
ImageAnti-Spoofing Accuracy
Only the labels of the glass layer0.4
Only the labels of the viable epidermis layer1
Only the labels of the stratum corneum0.05
Glass layer + viable epidermis layer labels1
Viable epidermis layer + stratum corneum labels1
Glass layer + stratum corneum labels0.54
Glass layer + stratum corneum + viable epidermis layer labels1
Table 3. Performance comparison with anti-spoofing methods (data in brackets are standard deviation values).
Table 3. Performance comparison with anti-spoofing methods (data in brackets are standard deviation values).
ModelAUCEERBPCER10
AE [49]0.8431 (0.0140)0.2172 (0.0129)0.3342 (0.0304)
VAE [50]0.8674 (0.0189)0.2158 (0.0153)0.3253 (0.0206)
MemAE [51]0.8981 (0.0135)0.1429 (0.0138)0.1590 (0.0200)
OCPAD [9]0.9181 (0.0114)0.1224 (0.0072)0.1368 (0.0179)
PPAD [21]0.9574 (0.0044)0.0705 (0.0106)0.0672 (0.0131)
U_OCFR [22]0.9681 (0.0073)0.0777 (0.0121)0.0678 (0.0175)
Ours0.9688 (0.0610)0.0668 (0.0112)0.0592 (0.0074)
Table 4. Ablation study on weighted module in segmentation network.
Table 4. Ablation study on weighted module in segmentation network.
ModelContribution WeightEERAUCBPCER10
Single-path reconstruction×0.16760.90340.1781
+ Single-branch segmentation×0.15520.93260.1224
+ Single-branch segmentation0.12270.93810.1678
+ Three-branch segmentation×0.10200.95790.0964
+ Three-branch segmentation0.06680.97480.0562
Table 5. Ablation study on variational attention module.
Table 5. Ablation study on variational attention module.
ModelEERAUCBPCER10
Transformer baseline0.18650.90280.2267
CNN baseline0.21580.86740.3253
+ Variational structure0.11760.95880.0938
+ Self-attention0.15330.92510.1786
+ VAM0.06680.97480.0562
Table 6. Ablation study on anomaly score determination.
Table 6. Ablation study on anomaly score determination.
Spoof ScoreEERAUCBPCER10
σScore0.07510.95680.0882
μScore0.09650.93290.1128
SpoofScore μ σ 0.06680.97480.0562
SpoofScoreG0.08520.96370.0711
Table 7. Comparison of model complexity and computation cost.
Table 7. Comparison of model complexity and computation cost.
ModelFLOPs (G)Parameters (M)
AE2.473.14
VAE1.6453.5
MAE1.593.14
OCPAD1.593.14
PPAD6.1510.87
UOCFR3.878.65
Ours5.629.31
Table 8. Paired t-test results comparing our method with PPAD and U-OCFR.
Table 8. Paired t-test results comparing our method with PPAD and U-OCFR.
ComparisonMetricOurs (Mean ± STD)Baseline (Mean ± STD)p-Value
Ours vs. PPADAUC0.9731 ± 0.00710.9553 ± 0.0034<0.0001
EER0.0654 ± 0.01020.0749 ± 0.01120.0759
BPCER100.0571 ± 0.00960.0673 ± 0.01010.0420
Ours vs. U-OCFRAUC0.9731 ± 0.00710.9665 ± 0.00660.0495
EER0.0654 ± 0.01020.0790 ± 0.01190.0180
BPCER100.0571 ± 0.00960.0702 ± 0.01480.0401
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Li, Z.; Ji, J.; Qi, J.; Fang, H.; Liu, Y. Optical Coherence Tomography-Based Fingerprint Presentation Attack Detection via Multi-Tissue Features Contribution Analysis. Appl. Sci. 2025, 15, 5642. https://doi.org/10.3390/app15105642

AMA Style

Li Z, Ji J, Qi J, Fang H, Liu Y. Optical Coherence Tomography-Based Fingerprint Presentation Attack Detection via Multi-Tissue Features Contribution Analysis. Applied Sciences. 2025; 15(10):5642. https://doi.org/10.3390/app15105642

Chicago/Turabian Style

Li, Zhanqing, Jie Ji, Jiajin Qi, Haonan Fang, and Yipeng Liu. 2025. "Optical Coherence Tomography-Based Fingerprint Presentation Attack Detection via Multi-Tissue Features Contribution Analysis" Applied Sciences 15, no. 10: 5642. https://doi.org/10.3390/app15105642

APA Style

Li, Z., Ji, J., Qi, J., Fang, H., & Liu, Y. (2025). Optical Coherence Tomography-Based Fingerprint Presentation Attack Detection via Multi-Tissue Features Contribution Analysis. Applied Sciences, 15(10), 5642. https://doi.org/10.3390/app15105642

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop