Next Article in Journal
Improving Pyroelectric Energy Harvesting Using a Sandblast Etching Technique
Previous Article in Journal
Strontium-Doped Hematite as a Possible Humidity Sensing Material for Soil Water Content Determination
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Finger Vein Recognition Based on Personalized Weight Maps

School of Computer Science and Technology, Shandong University, Jinan 250101, China
*
Author to whom correspondence should be addressed.
Sensors 2013, 13(9), 12093-12112; https://doi.org/10.3390/s130912093
Submission received: 18 July 2013 / Revised: 21 August 2013 / Accepted: 3 September 2013 / Published: 10 September 2013
(This article belongs to the Section Physical Sensors)

Abstract

: Finger vein recognition is a promising biometric recognition technology, which verifies identities via the vein patterns in the fingers. Binary pattern based methods were thoroughly studied in order to cope with the difficulties of extracting the blood vessel network. However, current binary pattern based finger vein matching methods treat every bit of feature codes derived from different image of various individuals as equally important and assign the same weight value to them. In this paper, we propose a finger vein recognition method based on personalized weight maps (PWMs). The different bits have different weight values according to their stabilities in a certain number of training samples from an individual. Firstly we present the concept of PWM, and then propose the finger vein recognition framework, which mainly consists of preprocessing, feature extraction, and matching. Finally, we design extensive experiments to evaluate the effectiveness of our proposal. Experimental results show that PWM achieves not only better performance, but also high robustness and reliability. In addition, PWM can be used as a general framework for binary pattern based recognition.

1. Introduction

Biometric recognition refers to the use of distinctive physiological and behavioral characteristics (e.g., fingerprints, face, hand geometry, iris, gait, signature), called biometric identifiers or simply biometrics, for automatically recognizing a person [1,2]. Since it is difficult to misplace, forge or share biometric identifiers, biometric recognition is more reliable than traditional token-based identification methods (e.g., keys or ID cards) or knowledge-based methods (e.g., passwords or PINs). Besides, biometric recognition also promises better security, higher efficiency, and better user experience in many cases. Among various kinds of biometric identifiers, finger vein recognition is a newly emerging biometrics technology. Medical research proves that each finger has a unique vein pattern that can be used for personal verification [3]. Generally, the finger vein recognition demonstrates some advantages over other biometrics methods [4,5]: (1) non-contact: finger vein patterns are not influenced by surface conditions, and it is more acceptable for the users; (2) live body identification: finger vein patterns can only be identified on a live body without fake finger attacks in fingerprint recognition; (3) higher security: finger vein patterns are internal features that are difficult to forge; (4) smaller device size: most finger vein capturing devices are smaller in size as compared to palm vein based verification devices [6]. Like other traits, finger vein recognition is composed of four main steps, including image capturing, pre-processing, feature extraction and matching. In order to better utilize the features from the segmented blood vessel network for recognition, the authors of [7] extract the finger vein pattern from the unclear image with line tracking. By means of fingerprint features, the minutiae features including bifurcation points and ending points are extracted from finger vein images [8], and a modified Hausdorff distance algorithm is provided to evaluate the identification abilities. In [9] a mean curvature method is proposed, which regards the vein image as a geometric shape, reporting an equal error rate of 0.25% with test images from 125 fingers. Due to the optical blurring and skin scattering problems, the finger vein images are not always clear and may show irregular shadings [10]. However, the image enhancement effect of finger vein is not very good when using traditional image enhancement methods such as histogram equalization and median value filtering [11,12]. To restrain the noises and emphasize the finger-vein linear pattern in the finger-vein image, Zhang et al. originally performed a multiscale self-adaptive enhancement transform based on the curvelets to the finger-vein image [13]. Yang et al. use the Multi-Channel Gabor Filter, Circular Gabor Filter and the Frangi Filter to enhance the quality of the vein image, respectively, and obtain better results [1416]. Li et al. analyze the shortages of Wavelet transformation and propose an image enhancement algorithm based on the Ridgelet transformation [17].

Multi-biometric systems are provided to further boost the performance of finger vein recognition. The work described in [18] exploits finger vein features in local moments, topological structure and statistics, respectively, and a fusion scheme is adopted for decision making. In [19] first a 2-D Gabor filter is used to filter the original finger vein images, and then the phase and direction texture features for combination in a feature level fusion are extracted. A recognition method based on the score-level fusion of finger veins, fingerprints, and finger geometry features is proposed in [20]. Other similar schemes can be seen in [21,22]. Extracting a clear segmented blood vessel network is very important for finger vein recognition [23], and improperly segmented networks may degrade the recognition accuracy to some extent. In order to avoid this problem, some binary pattern based methods are proposed [2426], whereby the captured finger vein images are enhanced by modified Gaussian high-pass filter and then LBP, LDP [26] or LLBP [25] are applied to extract the binary codes from the enhanced images. The similarity between the extracted and enrolled binary codes is measured by Hamming distance. Based on the binary pattern methods, the vein images are finally changed into different binary code sequences of the same length. The similarity between two vein images is equivalent to the similarity between two corresponding binary code sequences.

Current binary pattern based finger vein matching methods treat every bit of feature binary codes derived from different images of various individuals as equally important and assign the same weight value to them. Taking the LBP feature as an example, generally speaking, there are many samples captured from the same finger and LBP binary codes (abbreviated in LBPCodes) extracted from these samples, respectively. Considering each bit with same location of these LBPCode series, we find some of them are stable, i.e., the bit values are all 1's or 0's. On the contrary, some other bits are unstable, where some LBPCode series take the value of 1's, and the others a value of 0's. Furthermore, the stable bits of one finger vein are different from the others in terms of location and bit value. Motivated by these phenomena, in our previous work [27], a personalized best bit map (PBBM) method was proposed, in which only the most stable bits called the ‘best bits’ are used for recognition. However, in the PBBM method, the ‘unstable bits’ are all discarded without being used. Can these ‘unstable bits’ contribute some useful information, and how to make the most of these bits? These questions remind us to reconsider the binary pattern methods. We claim that the bits with different degrees of stability should have different contributions to the final matching result. Inspired by [28,29], in this paper, we propose a new finger vein recognition method based on a personalized weight map (PWM). Extensive experiments show that PWM can significantly improve recognition performance as well as robustness.

The rest of this paper is organized as follows: Section 2 presents the principles of binary pattern- based finger vein recognition. Section 3 introduces the definition of the PWM. Section 4 describes the framework of finger vein recognition based on a PWM. Section 5 reports the experimental results to verify the proposed method. Finally, Section 6 concludes this paper.

2. Binary Pattern Based Finger Vein Recognition

Binary pattern based finger vein recognition methods aim to solve the low quality images matching problem. Firstly, the captured finger vein images are preprocessed, mainly by ROI extraction, image enhancement, and size normalization. Then certain binary patterns such as LBP, LLBP, or LDP etc. are applied to extract features (called binary codes) from the preprocessed images. Finally, the similarity between the extracted and enrolled binary codes is measured by Hamming distance.

2.1. Binary Codes by LBP Operator

Local binary pattern (LBP) is a popular technique used for image representation and classification. LBP has been widely applied in various applications due to its high discriminative power and tolerance to illumination changes such as texture analysis and object recognition. In [30,31] a LBP operator as a nonparametric 3 × 3 kernel for texture classification was proposed. In [26] a LBP-based finger vein recognition method was proposed. A LBP can be defined as an ordered set of binary values determined by comparing the gray values of a center pixel and its neighboring pixels. In other words, first we compare the gray values of the center pixel and its neighboring pixels in a 3 × 3 kernel. If the gray value of a neighboring pixel is bigger than it of the center pixel, a 1 will be given to the corresponding bit in binary code, otherwise, 0 will be given. Then we order these binary codes from random bits clockwise or anticlockwise. We give an example in Figure 1. The binary values can be expressed in decimal form as shown in Equation (1) [26]:

L B P ( x c , y c ) = n = 0 n = 7 s ( i n i c ) 2 n
where ic and in in Equation (1) denote the gray value of the center pixel (xc,yc) and its eight neighboring pixels, respectively. The function s(x) is defined as [30]:
s ( x ) = { 1 i f x 0 0 i f x < 0

2.2. Binary Codes by LLBP Operator

Motivated by LBP, Petpon and Srisuk [32] proposed an LLBP operator for face recognition. The LLBP algorithm first obtains the line binary code along with horizontal and vertical direction separately and its magnitude, which characterizes the change in image intensity such as edges and corners, is then computed. It can be mathematically expressed as in Equation (3)(5):

LLBP h N , c ( x , y ) = n = 1 c 1 s ( h n h c ) 2 c n 1 + n = c + 1 N s ( h n h c ) 2 n c 1
LLBP v N , c ( x , y ) = n = 1 c 1 s ( v n v c ) 2 c n 1 + n = c + 1 N s ( v n v c ) 2 n c 1
LLBP m = LLBP h 2 + LLBP v 2
where LLBPh, LLBPv, LLBPm are LLBP on horizontal direction, vertical direction, and its magnitude respectively, N is the length of the line in pixel, c = (N/2) is the position of the center pixel hc on the horizontal line and vc on the vertical line, hn is the pixel along with the horizontal line and vn is the pixel along with the vertical line, and s function is same as in LBP (2). The illustration of LLBP operator is shown in Figure 2.

Employing (2) and (3), the horizontal component of LLBP (LLBPh) extracts a binary code of N-1 bits for each pixel. The same numbers of bits are extracted by the vertical component of LLBP (LLBPv) using (2) and (4). Consequently, by concatenating the binary codes from LLBPh and LLBPv, the total binary code of LLBP for each pixel is 2(N-1) bits.

2.3. Matching for Binary Codes

Denoting the binary feature of enrolled image by BinaryCodeA and the binary feature of input image by BinaryCodeB, the Hamming Distance (HD) is generally adopted to measure dissimilarities between the two binary patterns, which are represented as follows:

H D = BinaryCodeA BinaryCodeB LengthofbinaryCode

In Equation (6), is a Boolean exclusive-OR operator between two binary patterns.

3. Personalized Weight Map

If many samples are captured from a specific finger, we can get the corresponding BinaryCodes by extracting binary features from each sample. As previously mentioned, the values of the bits in the same location of BinaryCodes have different traits. The values of some bits are consistent, either 1's or 0's; the values of some bits have a majority of 1's or 0's; the values of some bits interlace 1's and 0's. Figure 3 gives an example of BinaryCodes of six samples from a certain individual. Apparently, bit 0, bit 2, bit 3, bit 5, bit 6 are very stable, the values of bit 1 are mostly 1's, the values of bit 7 are mainly 0's, and bit4 has interlaced 1's and 0's values.

These investigations provide a valuable clue to improve finger vein recognition performance: the stable bits should play a more important role in matching than the unstable bits, and the stability of a certain bit should be measured to use as a weight for matching. Suppose there are k training finger vein images in a finger vein class, and their binary codes are denoted by code1codek, we make k × k times of intraclass matching using these k codes and obtain the average matching result by Equation (7):

P = 1 k × k a = 1 k b = 1 k code a code b , P = { p 1 , p n }
where P is a vector with the same length as the binary codes, and ⊙ means logical NOT of ⊕in Boolean operations. For the ith bit, pi represents the probability of the matching result being 1. Every pi is of different value and bigger pi indicates that bit i is more stable. Suppose there are k binary codes in the same class, and the value of a certain bit i is 1's for m1 times and 0's for m0 times (m1 + m0 = k), so the average result of bit i is P i = m 1 2 + m 0 2 ( m 1 + m 0 ) 2 by Equation (7). It is obvious that pi is always between 0.5 and 1. For convenience, we normalize it to [0, 1] by Equation (8).

W = 2 P 1
where W (W = {w1; … ; wn}) is called the “weight map”. Because the weight map is different from individual to individual, we call the weight map for each individual the Personalized Weight Map (PWM).

With PWM, our improvement to the traditional Hamming distance is that a “weight map” WA of the registered finger vein class A, which is denoted by an n-dimensional vector {w1; w2; …; wn}T, is introduced to the matching function as follows:

Similarity A B = 1 ( BinarycodeA BinarycodeB ) × W A W A

If SimilarityAB is bigger more than a given threshold, the two finger vein images are considered from the same individual.

4. The Proposed Method

Based on the mentioned PWM definition, we propose a finger vein recognition method which mainly involves two stages: a training stage and a recognition stage. The training stage aims to generate the PWM for each individual, which includes preprocessing, feature extraction and finally generating the PWM using certain number of training samples. In the recognition stage, we first preprocess the test sample, and then extract the corresponding binary code, and next compute the similarity of binary codes between this test sample and the template of a certain individual in the enrolled database, finally obtain a recognition result with a given threshold. The framework of the proposed method is demonstrated in Figure 4.

4.1. Preprocessing

Like in reference [27], the preprocessing operation mainly refers to ROI extraction, size normalization and gray normalization. Figure 5 shows the results in each step.

4.2. Feature Extraction

The feature extraction process mainly involves how to extract the binary code of a certain finger vein image by a given binary pattern operator such as LBP, LLBP or LDP, etc. It should be pointed out that the PWM method can be regarded as a general framework, and then we call a certain binary pattern such as LBP the base feature of the PWM.

4.3. Training Process

Since each individual has his/her own PWM, so we need to train the corresponding PWM for each individual, and apparently a certain number of training samples from each individual is needed. First we get the corresponding BinaryCodes by extracting binary features from each training sample of one individual. Then the PWM for this individual will be computed by Equations (7) and (8). With different numbers of training samples, the corresponding PWMs are different too. Although the number of training samples is the same, the corresponding PWM is also different for different individuals. Considering the above situation, the main objective of the training process is to generate the optimal PWM for each individual. Several factors should be explored such as: (1) how many samples are suitable to train the PWM? (2) The robustness of the training method. Given the condition that the number of training samples is fixed, if the samples are different, can the recognition performance be kept stable? (3) With an increasing number of training samples, can the recognition performance also be boosted?

4.4. Recognition Process

In the recognition stage, we first extract the corresponding binary code of the test sample, and then compute the similarity of this test sample and the enrolled sample of a certain individual. The recognition process can work in two modes: verification mode and identification mode. In the verification mode, the class of the input finger vein (test sample) is known, and each of the samples is matched with all the other samples in the database. A successful matching is called intra-class matching or genuine, if the two samples are from the same class. Otherwise, the unsuccessful matching is called interclass matching or imposter. In the identification mode, we do not know the class of the input finger vein images. The samples of each class in the enrolled database can be classified into two categories: templates and probes. The probes were matched with all the template models, and for each probe, the matching results were ordered according to the matching scores.

5. Experimental Results and Analysis

5.1. The Experimental Database

The experiments were conducted on our new finger vein database which is collected from 34 individuals, including 20 males and 14 females; they are students, professors and staff in our school. The age of the participants is between 19 and 48 years. We collected the images in two separate sessions separated by 20 days. In each session, the subject was asked to provide 10 images for each of the left index finger, the left middle finger, the right index finger and the right middle finger. The capture device was manufactured by the Joint Lab for Intelligent Computing and Intelligent System of Wuhan University, China. The capture device mainly consists of a near-infrared light source, lens, light filter, and image capture equipment. Vein patterns can be viewed through an image sensor which is sensitive to near-infrared light, because near-infrared light passes through human body tissues and is blocked by pigments such as hemoglobin or melanin. A groove in the shell of the device is used to guide the finger orientation, and the capture device is illustrated in Figure 6.

5.2. The Experimental Settings

The original spatial resolution of the finger vein image is 320 × 240, After ROI extraction and size normalization, the size of the region used for feature extraction is reduced to 96 × 64. We set the parameter N of LLBP as 21 in our experiment. In order to verify the proposed method comprehensively, we design four experiments: (a) in Experiment 1, we take LBP as base feature to generate the PWM, and evaluate performance of the proposed method in verification mode and identification mode respectively. (b) Experiment 2 reports the recognition result of the PWM, which takes LLBP as base feature. (c) Experiment 3 is conducted to compare the performances of the PWM and the PBBM method which was proposed in our previous work [27]. (d) Experiment 4 discusses the computation time of the PWM method.

5.3. Experiment 1—Comparison between LBP and PWM

5.3.1. Verification Mode

In the experiments, we use the first ten samples of each class in the database to generate the PWM and use the other ten as testing samples. In order to avoid class imbalances, we use all the ten testing vein images in intra-class matching meanwhile the first two of the ten testing vein images in interclass matching. Consequently, there are 6,120 (136 × 45) intra-class matching and 73,440 (136 × 135 × 4) interclass matches in total. In this paper, we evaluate the recognition performance by the equal error rate (EER), the false rejection rate (FRR) at 0.1% false acceptance rate (FAR) and the FAR at 0.1% FRR, respectively. The EER is mainly used to measure the overall performance of biometric systems because the FRR and FAR are treated equally. On the other hand, the FRR at 0.1%-FAR is suited for high security systems, as in those systems, false acceptance errors are much more critical than false rejection errors. On the contrary, the FAR at 0.1%-FRR shows the acceptance rate of impostors when no genuine rejections are desired.

We compare the proposed method with the LBP based method described in [26]. Genuine and imposter matching score distributions of these two methods are shown in Figures 7 and 8, respectively. From Figure 7, we can see that the genuine and imposter matching scores by LBP are overlapped between 0.6 and 0.7. Otherwise, the genuine matching scores by PWM are mainly between 0.75 and 0.95, and the imposter matching scores are mainly between 0.40 and 0.75.

The ROC (receiver operating characteristic) curves are shown in Figure 9. Table 1 lists the ERR, FRR at 0.1%-FAR and FAR at 0.1%-FRR values. From Figure 9 and Table 1, we can see that the proposed method achieves a much lower EER than the LBP-based method. This indicates that the stability of a certain bit contributes to the final matching with a different degree. Traditional binary pattern-based finger vein matching methods treat every bit of feature binary codes derived from different image of various individuals as equally important and assign the same weight value to them. Such a uniform matching strategy definitely is not optimal in terms of finger vein recognition accuracy because it ignores significant differences of finger vein texture patterns among individuals and variations of image structures from the same individual.

To verify the robustness of the PWM trained by samples acquired in different sessions, we randomly select ten samples of each class in the database to generate PWM and used the other ten as test samples. We repeat this process ten times, and the statistical data of the verification performance (EER) is shown in Table 2. From Table 2, we can see that the PWM is robust for training and testing samples acquired from different sessions and always has better performance than traditional LBP based method.

5.3.2. Identification Mode

The experiments of identification are also reported. We use the first ten samples of each class in the database to generate the PWM and randomly select one from the other ten samples as template; the rest nine samples are taken as probes. Therefore, there are 136 templates, and 1,224 (136 × 9) probes in total. For each probe, there is a corresponding gallery of the same person, the gallery is sorted by decreasing similarity for each probe, and the probe is said to be correctly recognized at rank k if the gallery of the same person is among the first k images in the sorted gallery. Then, we can get the cumulative match curves as shown in Figure 10. The cumulative matching performance, rank-one recognition rate, and lowest rank of perfect recognition (i.e., the lowest rank when the recognition rate reaches 100%) are listed in Table 3. From the experimental results, we can see that the performance of the proposed method is much better than that of the LBP-based method.

5.3.3. On Different Number of Training Samples

According to the PWM definition, the proposed method needs a certain number of samples to generate the PWM for each individual. This experiment is conducted to show the effect on recognition performance when using different number of training samples to generate a PWM. For the sake of justice, we measure the performance by fixing the number of testing samples and adjusting the number of training samples to generate the corresponding PWMs. We set the last six as testing samples of each class in the established database, so that the number of training samples is between 2 and 14 (here we neglect the number of training samples is equal to 1, due to its meaninglessness ).

The ROC curves of PWMs trained by different number of samples are shown in Figure 11, The ERR, FRR at-0.1%-FAR and FAR at 0.1%-FRR values are listed in Table 4. From Figure 11 and Table 4, we can see that better performance is achieved with an increasing number of training samples; it also indicates that we can only use a small number of training samples of each class to generate a competitive PWM. For example, capturing the ten samples for each finger is not a very tedious task.

We also conduct experiments on the identification mode. Table 5 and Figure 12 show the experimental results. Apparently, when the number of training samples reaches 11, the rank-one recognition rate can achieve 100%, it means that all the probes can be identified correctly.

5.4. Experiment 2—Comparison between LLBP and PWM

With an identical experimental setting in Experiment 1 such as training number and experimental steps, this experiment reports the recognition result by using LLBP as base feature to generate the PWM. The verification results are reported in Table 6, and the ROC curves are illustrated in Figure 13. From Table 6, we can see that the PWMs have better performance than the traditional LLBPs method.

The identification results are shown in Table 7, and the cumulative match curves are illustrated in Figure 14. From the experimental results, we can see that the performance of the proposed method is also much better than that of the LLBP-based methods in identification mode. In addition, PWM-LLBP has overall better performance than PWM-LLBPh and PWM-LLBPv due to its combination of vertical and horizontal direction information at same time.

5.5. Experiment 3—Comparison between PBBM and PWM

According to the definition of PBBM and PWM, only the ‘best bits’ are used for matching in the PBBM method, whereas in the PWM method all the bits which contribute different weight values are used for matching. The difference between the PWM and PBBM can be described as follows: (1) In essence, the PBBM is a template; the binary code of a test sample matches with the PBBM directly. In addition, one class has only one template; it is difficult to conduct multi-template matching strategy. The PWM is just a weight map; it combines with two binary codes for matching. (2) The PWM has more generality, and the PBBM can be taken as a special case of the PWM. If the best bits of PBBM are very few, the recognition result may be not credible. The PWM utilizes not only the best bits (stable bits) but also other unstable bits, and assigns to these bits different weight values according to their stability.

This experiment is conducted to compare the recognition performance of PBBM and PWM when using different number of training samples. Using the same experimental setting of the experiment described in section 5.3.3, the last six samples of each class in the established database are fixed as the testing samples, and the number of training samples is changing from 2 to 14. The ERR values of PBBM and PWM are listed in Table 8. The contrast of ROC curves between PBBM and PWM trained by different samples are shown in Figure 15 (in Figure 15, ts is an acronym of training samples). From Table 8 we can see that as increasing the training samples from 3 to 14, the EER values of PWM are always lower than those of PBBM, and the superiority of PWM is more pronounced when the training samples is up to 10. It indicates that the ‘unstable bits’ which are discarded by the PBBM method can contribute to the final matching.

5.6. Experiment 4—Measurement of Time Complexity

All the above experiments are implemented in MATLAB, and run on a PC with a 2.4 GHz CPU and 2 G Memory. Average computation times for key operations in our system are listed in Table 9. Similar to the above experimental setting, we use 10 training samples to evaluate the PWM training time which includes the time interval from input of training images to getting the PWM. The matching time interval is from the input of two images to getting the final matching result. Although the average training time is somewhat time-consuming, the training process can be done off-line. In a word, from Table 9, we can see that the proposed method can be used in real time.

5.7. Discussion

According to the definition of PWM, given the fixed number of training samples, the possible weight value of a bit in the PWM is also decided. For example, if the number of training samples is 8, then the possible weight values of a certain bit are 0.00, 0.0625, 0.25, 0.5625, or 1.00 respectively. If the number of training samples is only 2, then the possible weight values of a certain bit are 0.00, or 1.00. Apparently we can conclude that the more number of training samples, the better of recognition performance. This fact also was verified in Experiment 1.

There is another open finger vein image database from Hong Kong Polytechnic University [33], which consists of 3,132 finger vein images taken from 156 volunteers over an average interval of 66.8 days. Among these 156 volunteers, only 105 subjects turned up for the imaging during the second session, other subjects contribute six images in the first session. Like in the analysis above, it's a limitation of the performance advantage of the PWM method with fewer training images, but in order to further assess the performance of the proposed method, we also conduct a simple experiment on this database. In this part, we use the first four samples of each class in the database to generate the PWM and the last two as testing samples. The verification results are reported in Table 10 and the ROC curves are illustrated in Figure 16. We get the region of interesting (ROI) images via the method proposed in [34] and extract features such as LBP and LLBP directly, without performing further image enhancement, so the EERS gained in this database are all higher than that in our database. From Table 10, we can see that the PWM-method has better performance than the Base-method, and we ascertain that a much better performance will gained as increasing the training samples.

6. Conclusions

This paper presents a novel finger vein recognition method based on the concept of a personalized weight map. The experimental results show the superior performance of our method in comparison with other traditional binary pattern based methods. The advantages of PWM can be summarized as follows: (1) PWM can effectively reduce the adverse effect of unstable bits. Through the training process, different bits are assigned different weight values which are used for final matching. The stable bit has a higher weight value and vice versa. (2) PWM must be trained for every individual respectively, so PWMs are different from individual to individual, thus PWMs can be regarded as a kind of personalized feature that reflects the differences between each individuals remarkably. (3) PWM is highly robust and reliable; this has been verified in the aforementioned experiments. (4) PWM can be used as a general framework for binary pattern based recognition. Besides LBP and LLBP, we can use other binary code to generate the corresponding PWM, and this will be the subject of our future work. At last, we expect that the PWM will display better adaptability to finger vein images with low quality.

Acknowledgments

This work is supported by National Natural Science Foundation of China under Grant No. 61173069, 61070097, Program for New Century Excellent Talents in University of Ministry of Education of China under Grant No. NCET-11-0315 and Shandong Natural Science Funds for Distinguished Young Scholar. The authors would like to thank Chunming Li for his helpful comments and constructive advice on structuring the paper. In addition, the authors would particularly like to thank the anonymous reviewers for their helpful suggestions.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Maltoni, D.; Maio, D.; Jain, A.K.; Prabhakar, S. Handbook of Fingerprint Recognition, 2nd ed.; Springer-Verlag: Berlin, Germany, 2009. [Google Scholar]
  2. Ross, A.A.; Nandakumar, K.; Jain, A.K. Handbook of Multibiometrics, 1st ed.; Springer-Verlag: Berlin, Germany, 2006. [Google Scholar]
  3. Yanagawa, T.; Aoki, S.; Ohyama, T. Human finger vein images are diverse and its patterns are useful for personal identification. MHF Prepr. Ser. 2007, 12, 1–7. [Google Scholar]
  4. Liu, Z.; Yin, Y.L.; Wang, H.J.; Song, S.L.; Li, Q.L. Finger vein recognition with manifold learning. J. Netw. Comput. Appl. 2010, 33, 275–282. [Google Scholar]
  5. Wu, J.D.; Ye, S.H. Driver identification using finger-vein patterns with radon transform and neural network. Expert Syst. Appl. 2009, 36, 5793–5799. [Google Scholar]
  6. Zhang, Y.B.; Li, Q.; You, J.; Bhattacharya, P. Palm Vein Extraction and Matching for Personal Authentication. Proceedings of the 9th International Conference on Advances in Visual Information Systems, Shanghai, China, 28–29 June 2007; pp. 154–164.
  7. Miura, N.; Nagasaka, A.; Miyatake, T. Feature extraction of finger-vein patterns based on repeated line tracking and its application to personal identification. Mach. Vis. Appl. 2004, 15, 194–203. [Google Scholar]
  8. Yu, C.B.; Qin, H.F.; Zhang, L.; Cui, Y.Z. Finger-vein image recognition combining modified hausdorff distance with minutiae feature matching. J. Biomed. Sci. Eng. 2009, 1, 280–289. [Google Scholar]
  9. Song, W.; Kim, T.; Kim, H.C.; Choi, J.H.; Kong, H.J.; Lee, S.R. A finger-vein verification system using mean curvature. Pattern Recognit. Lett. 2011, 32, 1541–1547. [Google Scholar]
  10. Lee, E.C.; Park, K.R. Image restoration of skin scattering and optical blurring for finger vein recognition. Opt. Lasers Eng. 2011, 49, 816–828. [Google Scholar]
  11. Hoshyar, A.N.; Sulaiman, R.; Houshyar, A.N. Smart access control with finger vein authentication and neural network. J. Am. Sci. 2011, 7, 192–200. [Google Scholar]
  12. Huang, B.N.; Dai, Y.G.; Li, R.F. Finger-Vein Authentication Based on Wide Line Detector and Pattern Normalization. Proceedings of the 20th International Conference on Pattern Recognition, Istanbul, Turkey, 23–26 August 2010; pp. 1269–1272.
  13. Zhang, Z.B.; Ma, S.L.; Han, X. Multiscale Feature Extraction of Finger-Vein Patterns Based on Curvelets and Local Interconnection Structure Neural Network. Proceedings of the 18th International Conference on Pattern Recognition, Hong Kong, China, 20–24 August 2006; pp. 145–148.
  14. Yang, J.F.; Yang, J.L. Multi-Channel Gabor Filter Design for Finger-Vein Image Enhancement. Proceedings of the Fifth International Conference on Image and Graphics, Xi’an, China, 20–23 September 2009; pp. 87–91.
  15. Yang, J.F.; Yan, M.F. An Improved Method for Finger-Vein Image Enhancement. Proceedings of the 2010 IEEE 10th International Conference on Signal Processing, Beijing, China, 24–28 October 2010; pp. 1706–1709.
  16. Yang, J.F.; Yang, J.L.; Shi, Y.H. Combination of Gabor Wavelets and Circular Gabor Filter for Finger-Vein Extraction. Proceedings of the 5th International Conference on Emerging Intelligent Computing Technology and Applications, Ulsan, Korea, 16–19 September 2009; pp. 346–354.
  17. Li, H.B.; Yu, C.B.; Zhang, D.M. Study on finger vein image enhancement based on ridgelet transformation. J. Chongqing Univ. Posts Telecommun. Nat. Sci. Ed. 2011, 23, 224–230. [Google Scholar]
  18. Yang, J.F.; Shi, Y.H.; Yang, J.L.; Jiang, L.H. A Novel Finger-Vein Recognition Method with Feature Combination. Proceedings of the 16th IEEE International Conference on Image Processing, Cairo, Egypt, 7–10 November 2009; pp. 2709–2712.
  19. Wang, K.J.; Liu, J.Y.; Popoola Oluwatoyin, P.; Feng, W.X. Finger Vein Identification Based on 2-D Gabor Filter. Proceedings of the 2nd International Conference on Industrial Mechatronics and Automation, Wuhan, China, 30–31 May 2010; pp. 10–13.
  20. Kang, B.J.; Park, K.R.; Yoo, J.H.; Kim, J.N. Multimodal biometric method that combines veins, prints, and shape of a finger. Opt. Eng. 2011, 50. [Google Scholar] [CrossRef]
  21. Lin, K.M.; Han, F.L.; Yang, Y.M. Feature level fusion of fingerprint and finger vein biometrics. Adv. Swarm Intell. 2011, 6729, 348–355. [Google Scholar]
  22. Yang, W.M.; Yu, X.; Liao, Q.M. Personal Authentication Using Finger Vein Pattern and Finger-Dorsa Texture Fusion. Proceedings of the 17th ACM International Conference on Multimedia, Beijing, China, 19–24 October 2009; pp. 905–908.
  23. Miyura, N.; Nagasaka, A.; Miyatake, T. Extraction of finger-vein patterns using maximum curvature points in image profiles. IEICE Trans. Inf. Syst. 2007, 8, 1185–1194. [Google Scholar]
  24. Lee, H.C.; Kang, B.J.; Lee, E.C.; Park, K.R. Finger vein recognition using weighted local binary pattern code based on a support vector machine. J. Zhejiang Univ. Sci. C 2010, 11, 514–524. [Google Scholar]
  25. Rosdi, B.A.; Shing, C.W.; Suandi, S.A. Finger vein recognition using local line binary pattern. Sensors 2011, 11, 11357–11371. [Google Scholar]
  26. Lee, E.C.; Jung, H.; Kim, D. New finger biometric method using near infrared imaging. Sensors 2011, 11, 2319–2333. [Google Scholar]
  27. Yang, G.; Xi, X.; Yin, Y. Finger vein recognition based on a personalized best bit map. Sensors 2012, 12, 1738–1757. [Google Scholar]
  28. Hollingsworth, K.P.; Bowyer, K.W.; Flynn, P.J. The best bits in an iris code. IEEE Trans. Pattern Anal. Mach. Int. 2009, 31, 964–973. [Google Scholar]
  29. Dong, W.B.; Sun, Z.; Tan, T.N. Iris matching based on personalized weight map. IEEE Trans. Pattern Anal. Mach. Int. 2011, 33, 1744–1757. [Google Scholar]
  30. Ojala, T.; Pietikainen, M.; Harwood, D. A comparative study of texture measures with classification based on feature distributions. Pattern Recognit. 1996, 29, 51–59. [Google Scholar]
  31. Ojala, T.; Pietikainen, M.; Maenpaa, T. Multiresolution gray-scale and rotation invariant texture classification with Local Binary Patterns. IEEE Trans. Pattern Anal. Mach. Int. 2002, 24, 971–987. [Google Scholar]
  32. Petpon, A.; Sisuk, S. Face Recognition with Local Line Binary Pattern. Proceedings of the Fifth International Conference on Image and Graphics, Xi’an, China, 20–23 September 2009; pp. 533–539.
  33. Kumar, A.; Zhou, Y. Human identification using finger images. IEEE Trans. Image Process. 2012, 21, 2228–2244. [Google Scholar]
  34. Yang, L.; Yang, G.; Yin, Y.; Xiao, R. Sliding window-based region of interest extraction for finger vein images. Sensors 2013, 13, 3799–3815. [Google Scholar]
Figure 1. Example of an LBP operator.
Figure 1. Example of an LBP operator.
Sensors 13 12093f1 1024
Figure 2. Example of an LLBP operator.
Figure 2. Example of an LLBP operator.
Sensors 13 12093f2 1024
Figure 3. Examples of binary code.
Figure 3. Examples of binary code.
Sensors 13 12093f3 1024
Figure 4. Framework of the proposed method.
Figure 4. Framework of the proposed method.
Sensors 13 12093f4 1024
Figure 5. Examples of preprocessing. (a) Original image. (b) ROI extraction. (c) Size normalization. (d) Gray normalization.
Figure 5. Examples of preprocessing. (a) Original image. (b) ROI extraction. (c) Size normalization. (d) Gray normalization.
Sensors 13 12093f5 1024
Figure 6. The finger vein image capture device.
Figure 6. The finger vein image capture device.
Sensors 13 12093f6 1024
Figure 7. Genuine and imposter matching score distributions by LBP.
Figure 7. Genuine and imposter matching score distributions by LBP.
Sensors 13 12093f7 1024
Figure 8. Genuine and imposter matching score distributions by PWM.
Figure 8. Genuine and imposter matching score distributions by PWM.
Sensors 13 12093f8 1024
Figure 9. ROC curves by different methods.
Figure 9. ROC curves by different methods.
Sensors 13 12093f9 1024
Figure 10. Cumulative match curves by different methods.
Figure 10. Cumulative match curves by different methods.
Sensors 13 12093f10 1024
Figure 11. ROC curves by different number of training samples.
Figure 11. ROC curves by different number of training samples.
Sensors 13 12093f11 1024
Figure 12. Cumulative match curves by different number of training samples.
Figure 12. Cumulative match curves by different number of training samples.
Sensors 13 12093f12 1024
Figure 13. ROC curves by different methods.
Figure 13. ROC curves by different methods.
Sensors 13 12093f13 1024
Figure 14. Cumulative match curves by different methods.
Figure 14. Cumulative match curves by different methods.
Sensors 13 12093f14 1024
Figure 15. ROC curves by different methods.
Figure 15. ROC curves by different methods.
Sensors 13 12093f15 1024
Figure 16. ROC curves by different methods.
Figure 16. ROC curves by different methods.
Sensors 13 12093f16 1024
Table 1. Verification performance by different methods.
Table 1. Verification performance by different methods.
EERFRR at-0.1%-FARFAR at-0.1%-FRR
LBP method0.03160.12500.5157
PWM method0.00560.02160.1150
Table 2. The statistical data of EER by random sampling.
Table 2. The statistical data of EER by random sampling.
MinimumMaximumMeanVariance
LBP method0.02370.03060.02735.9489e-06
PWM method0.00410.00650.00537.1540e-07
Table 3. Identification performance by different methods.
Table 3. Identification performance by different methods.
Rank-one Recognition RateLowest Rank of Perfect Recognition
LBP method94.93%46
PWM method98.61%5
Table 4. Verificaion of performance with different numbers of training samples.
Table 4. Verificaion of performance with different numbers of training samples.
EERFRR at-0.1%-FARFAR at-0.1%-FRR
2 training samples0.02750.12210.3175
3 training samples0.02450.10690.2526
4 training samples0.02070.08770.2663
5 training samples0.01960.07600.2468
6 training samples0.01820.06670.2410
7 training samples0.01660.05390.2414
8 training samples0.01520.05290.1604
9 training samples0.01280.04260.0702
10 training samples0.00780.03380.0185
11 training samples0.00450.02450.0098
12 training samples0.00300.01180.0056
13 training samples0.00250.00640.0048
14 training samples0.00250.00540.0048
Table 5. Identification of performance with different numbers of training samples.
Table 5. Identification of performance with different numbers of training samples.
Rank-one Recognition RateLowest Rank of Perfect Recognition Rate
2 training samples96.76%14
3 training samples96.62%14
4 training samples97.21%12
5 training samples97.79%7
6 training samples98.53%8
7 training samples98.53%8
8 training samples98.97%6
9 training samples99.26%5
10 training samples99.85%2
11 training samples100%1
12 training samples100%1
13 training samples100%1
14 training samples100%1
Table 6. Verification performance by different methods.
Table 6. Verification performance by different methods.
EERFRR at 0.1%-FARFAR at 0.1%-FRR
LLBP0.04020.14260.4937
PWM-LLBP0.00520.01760.1983

LLBPv0.04190.19460.4851
PWM-LLBPv0.00880.03190.2013

LLBPh0.05920.17500.8041
PWM-LLBPh0.00850.04170.1452
Table 7. Identification performance by different methods.
Table 7. Identification performance by different methods.
Rank-one Recognition RateLowest Rank of Perfect Recognition
LLBP92.97%50
PWM-LLBP99.67%5

LLBPv93.46%76
PWM-LLBPv98.13%44

LLBPh89.87%123
PWM-LLBPh99.02%37
Table 8. EER contrast with different training samples.
Table 8. EER contrast with different training samples.
EER of PBBMEER of PWM
2 training samples0.02700.0275
3 training samples0.02600.0245
4 training samples0.02210.0207
5 training samples0.02200.0196
6 training samples0.02080.0182
7 training samples0.02080.0166
8 training samples0.01960.0152
9 training samples0.01840.0128
10 training samples0.01590.0078
11 training samples0.01480.0045
12 training samples0.01340.0030
13 training samples0.01110.0025
14 training samples0.01220.0025
Table 9. The average computation times.
Table 9. The average computation times.
Base FeaturePreprocessingFeature ExtractionTraining PWMMatching
LBP43.7 ms1.3 ms363.9 ms90.5 ms
LLBP43.7 ms9.2 ms435.7 ms106.3 ms
LLBPv43.7 ms4.1 ms412.6 ms95.7 ms
LLBPh43.7 ms3.5 ms398.5 ms93.2 ms
Table 10. Verification performance by different methods.
Table 10. Verification performance by different methods.
LBPLLBPLLBPVLLBPH
Base-EER0.10220.12590.09630.1891
PWM-EER0.07010.07050.05710.1232

Share and Cite

MDPI and ACS Style

Yang, G.; Xiao, R.; Yin, Y.; Yang, L. Finger Vein Recognition Based on Personalized Weight Maps. Sensors 2013, 13, 12093-12112. https://doi.org/10.3390/s130912093

AMA Style

Yang G, Xiao R, Yin Y, Yang L. Finger Vein Recognition Based on Personalized Weight Maps. Sensors. 2013; 13(9):12093-12112. https://doi.org/10.3390/s130912093

Chicago/Turabian Style

Yang, Gongping, Rongyang Xiao, Yilong Yin, and Lu Yang. 2013. "Finger Vein Recognition Based on Personalized Weight Maps" Sensors 13, no. 9: 12093-12112. https://doi.org/10.3390/s130912093

Article Metrics

Back to TopTop