Next Article in Journal
Utilization of CD and DVD Pick-Up Heads for Scratch Inspection of Magnetic Disk in Dynamic State Using Microcontroller
Next Article in Special Issue
Application of Stockwell Transform and Shannon Energy for Pace Pulses Detection in a Single-Lead ECG Corrupted by EMG Artifacts
Previous Article in Journal
Imaging of Cell Structures Using Optimized Soft X-ray Contact Microscopy
Previous Article in Special Issue
ECG-Signal Multi-Classification Model Based on Squeeze-and-Excitation Residual Neural Networks
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

QRS Differentiation to Improve ECG Biometrics under Different Physical Scenarios Using Multilayer Perceptron

by
Paloma Tirado-Martin
*,
Judith Liu-Jimenez
,
Jorge Sanchez-Casanova
and
Raul Sanchez-Reillo
University Group for Identification Technologies (GUTI), Carlos III University of Madrid, 28911 Leganés (Madrid), Spain
*
Author to whom correspondence should be addressed.
Appl. Sci. 2020, 10(19), 6896; https://doi.org/10.3390/app10196896
Submission received: 30 August 2020 / Revised: 24 September 2020 / Accepted: 27 September 2020 / Published: 1 October 2020
(This article belongs to the Special Issue Electrocardiogram (ECG) Signal and Its Applications)

Abstract

:
Currently, machine learning techniques are successfully applied in biometrics and Electrocardiogram (ECG) biometrics specifically. However, not many works deal with different physiological states in the user, which can provide significant heart rate variations, being these a key matter when working with ECG biometrics. Techniques in machine learning simplify the feature extraction process, where sometimes it can be reduced to a fixed segmentation. The applied database includes visits taken in two different days and three different conditions (sitting down, standing up after exercise), which is not common in current public databases. These characteristics allow studying differences among users under different scenarios, which may affect the pattern in the acquired data. Multilayer Perceptron (MLP) is used as a classifier to form a baseline, as it has a simple structure that has provided good results in the state-of-the-art. This work studies its behavior in ECG verification by using QRS complexes, finding its best hyperparameter configuration through tuning. The final performance is calculated considering different visits for enrolling and verification. Differentiation in the QRS complexes is also tested, as it is already required for detection, proving that applying a simple first differentiation gives a good result in comparison to state-of-the-art similar works. Moreover, it also improves the computational cost by avoiding complex transformations and using only one type of signal. When applying different numbers of complexes, the best results are obtained when 100 and 187 complexes in enrolment, obtaining Equal Error Rates (EER) that range between 2.79–4.95% and 2.69–4.71%, respectively.

1. Introduction

In the last years, improvements in sensors have produced more affordable and faster biometric acquisition systems, with higher processing capabilities. Besides, evolution in mathematics has led to faster and more accurate pattern recognition algorithms, such as those covered in machine learning.
Electrocardiogram (ECG) biometrics is one of the biometric modalities that has been noticeably improved thanks to machine learning algorithms. This modality is based on the electrical activity in the heart related to time represented by ECGs. ECGs are measured with electrocardiographs through different parts and angles of the body, allowing them to obtain the signal from several perspectives. Its standard acquisition results in 12 different types of signals or leads that provide different information about the heart’s performance [1]. These signals get frequently classified in two types [2]:
  • Limb leads: require four sensors, placed on the right arm, left arm, left leg, and right leg. The latter functions as a ground. Limb leads are also divided into two categories:
    Standard bipolar leads (I, II, and III): measure the voltages between left arm-right arm, left leg-right arm, and left arm-left leg, respectively.
    Augmented unipolar leads (aVR, aVL, and aVF): measure the relative voltages concerning the extremities, instead of using ground references. They allow us to observe the previous signals from different angles.
  • Chest (precordial) leads (V1, V2, V3, V4, V5, and V6): are calculated considering ground as a reference and require of six sensors placed carefully on different parts of the chest.
Figure 1 represents the general sinus rhythm signal, formed by the P waveform followed by the QRS complex and the T waveform. The different waveforms provide information about how the different areas of the heart perform. P represents the depolarization of the atria. The QRS complex is a consequence of the depolarization of the ventricles, and the T wave shows their repolarization [3].
ECGs can be used as a biometric signal because they are universal, unique, invariant enough with respect to its template and quantitative [5]. Every alive person can provide an ECG and it can be measured with commercial electrocardiographs. This biological signal also depends heavily on the morphology of the heart, which is unique even among healthy individuals. The addition of variations due to physiological conditions related to skin properties, gender, or age, makes ECGs unique for every person [6]. They are also considered stable in the long term and provide good data separation between genuine users and imposters [7]. To our knowledge, the first approaches using ECG signals for human identification were released in 2001 [8,9].
These types of signals are interesting for biometrics because of their nature. They are difficult to access without the user’s cooperation and susceptible to the specific scenario the user is in, due to variations such as those caused by heart rate o amplitude changes. These characteristics make them difficult to counterfeit while providing them with a lot of potential in biometrics.
The present work provides a study of the potential of Multilayer Perceptron (MLP) Neural Network structure for ECG biometric verification, extending the achieved results in Reference [10]. Achieving good performances with a classifier such as MLP results in relevant information, due to the challenging properties of the applied database. In addition, the feature extraction process is simplified by re-using the calculated data in segmentation, reducing it to a simple window selection. This simplification allows us to reduce computational processing time and cost. Finally, we provide the optimal configuration for the classifier, including the type of features and data size, as well as a discussion about the information these results provide about the dataset.

Related Work

Many different steps take part in a biometric system: the sensor should be comfortable to use and provide a good quality signal. The pre-processing needs to improve the signal quality to facilitate further feature extraction and requires to be done wisely, to avoid dealing with non-relevant data. Pre-processing aims to avoid slowdowns and lower performances in the final stage, where the classification takes place. The chosen algorithm must generalize properly with new data while only using limited available data. If one of these steps is performed poorly, it can affect noticeably to the system’s performance.
Out of the different steps, signal acquisition is probably one of the most problematic in ECG biometrics. Commercial sensors are developed for medical use, that is, electrocardiographs, providing good quality signals but they are not realistic to use in a biometric environment. Biometrics require easy-to-use sensors, as they contribute to a faster recognition process as users usually do not have previous knowledge. Besides, medical sensors are inconvenient due to sensor placing, which is time-consuming. Moreover, depending on the required lead, the sensors have more uncomfortable and complicated placing, such as those in the chest. Some portable ECG acquisition devices are available for the public, such as AliveCor’s KardiaMobile [11] and Nymi’s band [12]. They are focused on the type I lead as they only require the arms to be involved. The first is focused on health; the latter is the only one, to our knowledge, that uses ECG for recognition purposes. However, neither of them allows us to obtain raw samples. This fact limits the development of biometric systems with reliable and user-friendly sensors. However, working with medical sensors is common to set a baseline. It allows us to start from the most ideal case, due to the precision these devices provide and work towards more complex scenarios. As an alternative, other researchers develop their prototypes, adding more challenges in the system’s implementation [13,14,15].
Even if sensors collect signals with high signal to noise ratio (SNR), data needs to be prepared accordingly to facilitate the feature extraction process. The noise in ECGs usually comes from three main sources: baseline wanders and drift, power-line interference and muscle artifacts [16]. Baseline wander is usually in the 0.2–0.5 Hz range and power-line interferences are found from 50–60 Hz, so both can be removed with a band-pass filter [17]. The simultaneous removal of both types of noise is also achieved with alternatives such as Discrete Cosine Transform (DCT) [18] and Discrete Wavelet Transform (DWT), which allows decomposing the signal, however, it is not as frequent as band-pass filters [19]. In recent years, deep learning techniques such as Convolutional Neural Networks (CNN) have also been applied to this problem [20]. Muscle artifacts also provide high-frequency noise around 100 Hz and can get removed with a low-pass filter, but techniques such as Moving Average (MA) filters are also applied [21].
Fiducial point detection is usually the next step to filtering. These points are those such as the ones that take part in the sinus rhythm wave in Figure 1, but not limited to the ones represented. They behave as reference points to help in the following feature extraction process. The fiducial points to detect depending on the selected feature extraction approach. Some works only need to spot the QRS complex, and most of them apply the Pan-Tompkins algorithm [22] or modify it slightly [13]. Once the reference points are calculated, signals can be segmented accordingly if needed. It is common to select fixed ranges using the detected points as a reference [23].
The signal conditioning usually leads to the feature detection process. Literature divides this process into two main approaches—fiducial features, which are based on time and amplitude related metrics; and non-fiducial features, based on the shape of the whole signal or selected segments. The use of fiducial features is not a trivial task to achieve, because their accuracy relies on the performance of the fiducial point detection algorithm [24]. Non-fiducial features usually apply Fourier or Wavelet transforms [25,26]. Fiducial features tend to work better in databases with lower variability (i.e., databases with a low number of subjects), because it avoids the use of unnecessary data, deleting noise as the selected signal characteristics belong to specific, narrower regions. However, in databases with higher variability, non-fiducial features work better. They deal more efficiently with the higher chances of noise in a greater number of samples [27].
One of the most common techniques for ECG classification used to be Support Vector Machines (SVMs) [28] and the k-Nearest Neighbour classifier (k-NN) [7]. Although they are still applied techniques, Artificial Neural Networks (ANNs) have increased their popularity since they starting being used at the beginning of the last decade [29]. Concerning this, future approaches are expected to use ANNs and Deep Neural Networks, due to their potential to solve SVM and k-NN problems, while even improving the results of more conventional classifiers [19].

2. Materials and Methods

2.1. System’s Description

The system in this work follows a common approach in biometrics. The raw signal, in this case, ECG, goes through a pre-processing stage that improves the quality of the information. The data can go through transformations or detection algorithms for relevant reference points. After preparing the data, the feature extraction segments the most important information to go through the final classification.

2.1.1. Pre-Processing and Feature Extraction

The acquired ECG data follows the scheme in Figure 2. As referenced in Reference [10], to delete noise, the ECG signals pass through a fifth-grade Butterworth bandpass filter from 1 to 35 Hz. Once the signal is filtered; the goal is detecting the R peaks to help in further signal segmentation. The filtered signal gets differentiated, obtaining first and second differentiations, called Non-Differentiation (ND), First Differentiation (FD), and Second Differentiation (SD) respectively. FD and SD help in the R peak detection because they provide information about changes in ND. R peaks are the local maxima in ND and translate into local minima in FD. These local minima are easier to detect than the original R peaks because they are more prominent. In the case there are outliers, two adjacent segments get compared by correlation. SD finally helps to check that the final segments have the corresponding shape.
After the R peak references are calculated, the feature extraction consists of segmenting the QRS complex in ND, FD, and SD independently, obtaining 3 different types of signals, as seen in Figure 2. The segmentation is given by two variables: rng1 determines the number of signal points to the left of the reference point and rng2 the samples to the right including the reference point itself. This implies that the segments have a length of r n g 1 + r n g 2 points, which are the selected feature points. N peaks specifies the number of peaks selected to form the matrix, selected by appearing order, which is the samples.

2.1.2. Classification

The chosen algorithm for classification is the MLP neural network. These structures have already been used as a classifier in ECG biometrics, achieving good performances [29,30]. Despite of the simplicity of this artificial neural network, it is considered a promising algorithm in the context of ECG biometrics [31]. MLP networks are applied in supervised learning and they have three main parts: input, output, and hidden layers, as represented in Figure 3. The input layer is formed by nodes or neurons that represent the different input features { x i | x 1 , x 2 , , x n } . Every feature is labelled with its correspondent class, { y i | y 1 , y 2 , , y n } . In the case of only having one hidden layer, the output layer gives the function in Equation (1) [32]:
f ( x ) = W 2 g ( W 1 T x + b 1 ) + b 2 .
Where W 1 represents the sets of weights applied to every feature in the input layer. These weights vary between them, in the way that every feature x i has m different weights: one per node in the following hidden layer. On the same way, W 2 represents the weights applied in the hidden layer, at nodes { a j | a 1 , a 2 , , a m } . Value b 1 is the bias in the hidden layer while b 2 is the bias on the output layer. The activation function is represented by g ( · ) . The most common functions are identity (or no activation function), logistic, hyperbolic tangent (tanh) and rectified linear unit function (ReLU). All the corresponding functions are represented in Table 1.
As this structure works for supervised learning, the weights need to change in every connection after the data is processed to decrease the processed error. In this case, it is done by backpropagation, which comes from the Least Mean Squares (LMS) algorithm. These weights can be updated differently, depending on the approach for their optimization. The most common solver is of Stochastic Gradient Descent (SGD). Its formula depends on a factor called learning rate, which ensures the weights converge quickly.

2.2. Algorithm Modelling

As referred to in the previous section, the feature extraction depends on several values that need to be fixed for segmentation. The amount of data to introduce to the MLP classifier is determined based on these parameters, which rely on the database characteristics. At the same time, the MLP classifier depends on other parameters, called model hyperparameters, which vary depending on the implementation. Their values are set through a process called tuning, whose goal is achieving the most suitable values for correct classification.

2.2.1. Database

As ECG signals provide sensible information, releasing databases to the public is complicated due to privacy policies, leaving only a few databases to work with. Several databases are public in Physionet [33] being MIT-BIH Normal Sinus Rhythm (MIT-BIH NSR) and PTB [34] the most common ones in the literature. However, none of them are thought for human recognition but for helping in pathology classification, thus the acquisitions are done with commercial electrocardiographs. These databases provide long signals that are minutes or even hours long, while having the user in limited physiological conditions, usually resting. The ECG-ID database [35] is the only public one that aims for human recognition, providing 20-s signals. They are acquired by wrist electrodes and under two different conditions: sitting and free movement, making ECG-ID the closest to a biometric scenario.
According to ISO 19795 [36], data collection needs to be representative of the target application. This is the reason why in this case we need to consider that heart rates fluctuate constantly under different situations like stress and exercise. This parameter can be controlled in enrolment by helping the subject to relax. However, that is not something doable in recognition, where the subject must be independent. Adding this extra step would make the system less user-friendly and inconvenient. With these considerations, the previous databases focused on medical purposes do not provide the required data for biometrics. These databases tend to have long acquisitions and/or do not usually provide visits under different scenarios and/or days. In the case of ECG-ID, the free movement scenario is not specific enough about the activity, so the heart rate increase is not pursued, not making it useful for the study under more specific physiological conditions.
In terms of the capture device, working with professional medical sensors provides higher signal quality. This approach is chosen to remove as much noise as possible, maximizing the isolation of the signal behavior, even if it does not have the required usability for biometrics. The database acquisition is done using a Biopac MP150 system with a 1000 Hz sample frequency, as described in Reference [10]. The sensors obtain Type I lead, which is convenient in biometrics because it only involves the arms, measuring the voltage difference between left and right. Moreover, the placement of the limb sensors requires little expertise so their placement is similar between the different visits, in the pursuit of reducing this type of noise. The present work uses the second part of the extended database, as it is the only one that provides different physiological states. The collection is done with 55 healthy subjects in 2 different days, with two visits per day distributed as shown in Table 2. Each visit records 5 signals of 70 s duration per signal with a 15 s posture adjustment between every recoding.
The applied database remains private, due to the General Data Protection Regulation (GDPR) by the European Government. The law started its implementation in 2018 and considers that biometric data is sensitive. The GDPR takes into consideration the potential need to use this data in research but demanding some specific conditions. As this database was collected before the GDPR implementation, it does not fulfill the legal criteria to get published.

2.2.2. Implementation

Classification

The classifier implementation is done using Python 3 and using scikit-learn [37]. Every user provides 5 signals per visit, from which ND, FD, and SD are calculated. The first and last 5 s are discarded from every signal, obtaining a 60 s signal. From every one of them, 50 R peaks are detected and segmented accordingly to extract the QRS complex. The complex has the R peak on the 101th position and is formed by 200 points. All complexes get concatenated, summarizing the database into three types of matrixes for every visit: one per type of signal differentiation. Train and test sets are extracted from this data, selecting the differentiation and visit differently based on the type of experiment to carry out. The training matrix, X t r a i n , has dimensions [ r t r a i n × c ] where c is the number of columns or points in the QRS complex and r t r a i n is the number of rows, which corresponds to the number selected complexes. Similarly, testing data, X t e s t , has [ r t e s t × c ] dimensions.
Some of the required hyperparameters for the MLP classifier need to be set to a specific value due to previous knowledge and considerations. Fixing some of the hyperparameters make the hyperparameter tuning easier. The number of hidden layers is set to one because it usually achieves proper results and avoids extra slowdown [38]. Regarding the type of solvers, not only SGD is an option, but also another SGD-based optimizer, adam [39], and L-BFGS, which is a quasi-Newton optimizer. However, among these three solvers, only adam is used due to our preliminary results where SGD and L-BFGS provide very low-performance results. This decision discards the option of applying a specific learning rate formula because it is only applied to SGD solvers. In this case, we only need the size of the step that updates the weights, which is fixed to the default value, 0.0001. Convergence is assumed when the result does not improve significantly in a specific number of iterations, and this hyperparameter is also given by its default value which is 10. Table 3 summarizes the given values to the discussed parameters and their functionality.
After fixing the previous values, the remaining hyperparameters need to be fixed to reach an optimal system’s performance. Even though there is one hidden layer, the number of nodes in the hidden layer needs to be assessed. The same happens with the activation function which can be any of those in Table 1. This implementation solver optimizes the loss function in Equation (2). This formula has two components: the first summation corresponds to Cross-Entropy, where t indicates whether a class is positive or not, p ( s ) is the softmax output implemented by the library in a multiclass case and C is the number of classes; the latter belongs to L2-regularization, where α or alpha is the penalty term and | | W | | 2 represent the Euclidean norm of the weight hyperparameters. The value alpha avoids underfitting when it gets a positive value lower than 1 and it needs to be specified.
L o s s ( W ; t , p ( s ) ) = i C t i l o g ( p ( s ) i ) + α | | W | | 2 2 .
Finally, closely related to the convergence assumption, is the tolerance hyperparameter. The tolerance provides the quantity the loss function needs to be improved to keep iterating.

Performance Measurement

To evaluate the performance of a verification system, the ISO 19795 [36] does not encourage the use of single values such as percentages, because they miss relevant information. The Detection Error Trade-off (DET) graphs are the main tool instead, as they show the False Non-Match Rate (FNMR) and False Match Rate (FMR) evolution with respect to different thresholds. The point where both are equal represents the lower possible error rate and is called the Equal Error Rate (EER) [5].
The decision on whether the user is whom he/she says is based on the scores retrieved by the classifier. This implies that depending on the applied criteria, the system could not verify genuine users (FNMR) when doing otherwise with impostors (FMR). Some biometric applications, such as the entrance of a gym, would rather have higher FMR as a trade-off for a fast authentication process. On the other hand, if the authentication is for a bank account, the system would rather have high FNMR. This could increase the probability of asking the user for their data more than once, making the process more inconvenient. The trade-off between convenience and security must be determined in design, depending on the system’s target application.
In the case of the selected database, classes for the test samples are known. With this information, the model’s performance can be assessed from the probabilities retrieved by the MLP classifier. Typically, scorers or loss functions are applied for this purpose. However, in this case, we implement a different approach and apply a custom function that implements the EER calculation.
To do so, when several samples from the same class are used in the test set, all the probabilities are unified applying the mean. This provides only one probability vector for that given class in the test, containing the mean probability for every comparison combination. By paying attention to the reference class in every comparison, we can divide these results into two types: mated that is, when reference and comparison classes are the same; and non-mated if otherwise. From them, we can obtain the False FNMR and FMR. As they represent both possible types of error, the crossing point of both lines gives the EER, acting as the score for further optimization.
Despite ISO guidelines, performance metrics are still heterogeneous in the state-of-the-art. To be able to compare this work’s results with similar ones in the literature, accuracy calculation was also implemented. In this case, a threshold is applied to the scores retrieved by the classifier, giving values 0 (if discarded) or 1 (if verified). The accuracy is calculated by doing a simple division of total correct classifications by the total number of samples.

2.2.3. Hyperparameter Tuning

The tuning process is done independently for ND, FD, and SD signals, to observe which one provides better performance results. The type of signal is not specified through this section, but it is repeated similarly for the three of them. The only difference is on the model hyperparameters and their performance results.
The goal of tuning is determining the most optimal set of values to model our data. This work achieves these results by first applying a Random Search, as it has been proven a suitable technique that requires less computational time [40]. The exhaustive Grid Search is performed afterward, as represented in Figure 4. The whole tuning process is done based on the data in one of the visits, which behaves as the enrolment. In this case, it is the data in R1, which is expected to be the most stable set because the user’s relaxed and sitting down. This process is repeated individually for ND, FD, and SD, obtaining 3 different configurations, one per signal.
As for testing the model after the tuning process, the test data needs to be new, we need to split R1 into two parts: one is used for training (development matrix), and the other one for validating (validation matrix) the trained models. We decided to set the train size to 50% to obtain the same sizes for training and validating, which results in 125 QRS complexes per user in both matrixes.
The division is done using a method called Stratified Shuffle Split, which selects the samples randomly but keeping the class proportion to avoid having an unbalanced model. The random shuffling is replicable, so results are not biased by different shuffling orders.

Random Search

The Random Search method sets a number of random hyperparameter combinations, training a new model for each one of them. The activation function can be all four types in Table 1. The rest of the hyperparameters are numerical, and the selected range of values tries to represent extreme values, with equidistant middle values to observe if performances are correlated to little changes in any of them, as seen in Table 4. Tolerance is still relatively low to try to achieve more accurate results.
In this case, we obtain 50 random value combinations by training with the development set. The MLP training is done using cross-validation to observe how well the model performs under different train and test sets. In this case, the cross-validation process is done 5 times, dividing the development set into 80% training and 20% test. This division is also done by using Stratified Shuffle Split. The final performance results for every combination contains the mean EER through all the splits. Table 5 shows the best 3 results for ND, FD, and SD in descending order.

Grid Search

Selecting only the present values in Table 5, Grid Search performs training similarly to Random Search, but doing it exhaustively that is, doing it for every possible combination of hyperparameter values instead of for a limited number. This can be done due to the value reduction Random Grid has provided, because it allows us to discard a huge range of potentially bad combinations, decreasing the processing time significantly. The results are summarized in Table 6. This combination gets tested with the validation data, to confirm it performs properly under new data.

3. Results

By only using half of the R1 data for training, we observe that best verification performances are obtained through ND signals, followed by FD and SD. However, this does not imply that is going to happen when applying different train and test sets. The tuning has allowed us to fix those hyperparameters, ensuring the result will be as good as it can get when using data in R1 as training and testing. However, we need to check if the selected structures for ND, FD, and SD behave correctly in other cases.
The different combinations of the train and test sets provide different information. Good performances using the same visit for training and testing means that the applied data has low intra-class variability and high inter-class variability. This translates into the same user signals being similar enough among them, but also significantly different from the rest of the users. When using different visits in training and testing, we can observe how intra-class and inter-class variabilities are affected. Once the most suitable training set and type of signal are determined, we vary the amount of training data to observe changes in the final performance.

3.1. Results within the Same Visit

For every type of signal and visit, 50% of the data is used for training and testing is carried out with the remaining information in the visit. The random sample selection is the same in ND, FD, and SD. This process gives an idea of how the system configuration works within different user visits. In Table 7 we can see that results are 0 or close enough to 0. As expected, the highest values are calculated under Ex, due to its higher variability. FD is the one with the lowest EER values. SD provides good results under S but performs poorly in the remaining visits.
Figure 5 represents DET graphs for ND, FD, and SD when using Ex in training as an example. As seen in the previous results, FD has the best performance, followed by ND and SD. In the three types of signal, the FNMR reaches higher percentages than the FMR, meaning that false negatives are more common than false positives. In a biometric environment, this could lead to the need of repeating the recognition process even if the user is whom they say to be.

3.2. Results within Different Visits

In the previous experiment, both enrolment and recognition data were acquired under the same scenario. Nonetheless, to represent the potential signal variation in a biometric environment, we need to study the results when using different visits for enrolment and recognition. The scheme in Figure 6 represents the steps followed to obtain the different results when using R1 as a train set or enrolment and the remaining visits as a test set or recognition data. It considers a variable training size that refers to the data proportion that is going to be used for training. In the case it is lower than 1, that is, not all the samples are used, we use Stratified Shuffle Split for the division. The variable xD represents the different types of signals, where x can be changed to refer to ND, FD, or SD. This scheme is followed also by using the remaining visits S, R2, and Ex. Results for all 4 different train sets are represented and discussed in the following paragraphs. For comparison purposes among different graphs, the y-axis is fixed to be common in all of them.
Figure 7 represents all the DET performance curves when R1 is used as the enrolment or train set. The best performance for S is achieved under ND. However, it also provides the worst results of Ex. We can also observe that the results of the three visits are more spread in ND, obtaining very different results depending on the visit. In the cases of FD and SD, the results are more similar among visits, being better in the case of FD as it provides lower FNMR and FMR throughout the graph than SD.
Similarly, Figure 8 represents DETs for S as training. We observe the same pattern as before, where ND provides more differentiated results among visits, whereas FD and SD have similar trends. The best result is always achieved under R1, which tells that the train and test visits are not highly differentiated. It could be due to the influence of being taken on the same day, or a not relevant variation in the data between sitting and standing. The results for Ex and R2 are not significantly different either because they keep a low EER.
In the case of R2 in training, ND keeps performing worse than FD and SD, as seen in Figure 9. However, generally, the three signals have similar performances but reach the best ones under FD, although the improvement is not very noticeable. In these graphs we observe Ex reaching values like those in R1 and S as training, which would not happen in the previous cases. This shows that even R2 and Ex are were taken under different heart rates in the users, the fact that they were close in time results in good performances, obtaining EER around 5%.
Finally, Ex is used as training, collecting its results in Figure 10. As expected, due to the higher heart rate in this visit, results do not improve the ones seen so far. However, it is consistent with FD being the best signal in performance. In addition, results for R2 are the lowest ones, reinforcing the theory that the signal taken in the same day behaves well, despite the different heart rate. Again, R1 and S have similar trends.
When using the same data set in training and testing, we observed that FD achieved better results, and it has been consistent when testing with different visits. Even though best results are obtained with R1 as train set, keeping the EER around 5% or lower when using FD, we have observed that similar results have also been obtained using FD when R2 is the train set. This provides two important pieces of information: resting is the most suitable condition for the user to be enrolled with, and the system still provides good results when identifying with information taken on different days.

Final Configuration and Its Results

Considering all the previous processes and results, the final MLP configuration is summarized in Table 8, considering the previously fixed hyperparameters in Table 3. The enrolment data must be acquired in a relaxed state while relaxing, providing 125 QRS complexes. To reduce the computational cost, the only type of signal fed into the classifier is FD, as it has shown to result in a good performance on its own. Under this configuration, the EER is calculated when using 125 complexes of FD data in R1 to enrol. It results in 3.64%, 3%, and 5.54% for S, R2, and Ex visits, respectively.

3.3. Results for Different Enrolment Sizes

The determination of the most optimal MLP configuration was done using only one of the visits. It was required to split the data between training and validation, using only half of the available data for modeling. Using the highest amount of possible data could lead to better results. However, providing too much information could cause overfitting, resulting in a model that is not capable of generalizing properly with new samples. The previous section already set the most suitable configuration when using 50% of R1 data with FD. The goal of this experiment is to vary the training proportion to observe how the initial results get affected. The training is done using 25%, 40%, 50%, 75%, and 100% of the available R1 FD data, while the testing is the same as the one in the results between different visits. This data is plot in Figure 11. The different performances under S are plot in Figure 11a. The lowest EER is obtained when using 100% of the available data in training, increasing as the proportion decreases. However, we can see that there are not huge differences between using 60% or 75% of the data, as it also happens with 40% and 50%. However, when data is reduced to only 25%, the EER almost doubles the previous result. Similarly, Figure 11b represents these results in the case of R2. In this visit, using 100% provides closer results to those in 80 and 75%. As seen in the case of S, 40% and 50% do not provide significant differences, and decreasing the enrol data to 25% provides the same impact as before. In the case of Ex, represented in Figure 11c, results do not change significantly between sizes. Even though we observe 100% as the higher in EER, given that all results are close to each other, this subtle difference could have nothing to do with the training size.
All the EER results are collected in Table 9. As seen graphically, using only 40% of the available data for enrolment provides similar results to those in 50%, being even slightly improved. This change could mean going from 125 necessary cycles to 100, decreasing the enrolment data acquisition in a few seconds. Increasing to 60% also makes a slight improvement, although not in the case of Ex. Using 75% of the data results in the lowest EER for Ex and one of the lowest for S and R2, being under those obtained in 100%. However, the different EER results when increasing the proportion from 75% to 100% is noticeable, as S and R2 perform better by causing a poorer performance in Ex. This is one of the reasons to avoid using such a long enrolment if there are chances of having very different heart rates in recognition.
After analysing these results, we can say that our system can reach results between 2.79% to 4.95% EER, using 100 QRS complexes. If the purpose of the system requires higher performances with the expense of longer enrolments, increasing the size of the data set is a good approach. Providing 187 QRS complexes for enrolment provides an EER between 2.69% and 4.71%. Finally, results in terms of accuracy are represented in Table 10 applying a threshold of 0.5. The achieved accuracy under every condition is higher than 97% considering single heartbeats. We can also observe that the accuracy increases slightly up to 1% when increasing the training data.

3.4. Results Comparison

There are heterogeneous techniques throughout literature along with all different steps in the process. Not only results but also tools need to be taken into consideration. The selected works to compare our results to, are chosen due to similarities with the present work, as well as being relatively recent, going from 2017 to 2019. The techniques and some of their results are summarized in Table 11.
The criteria to select these works were based on several facts: getting databases whose characteristics were the closest to the proposed one, while having similar segmentation processes and/or same classification algorithm. However, fulfilling all conditions at the same time of providing similar procedures in testing and metric calculations is very complicated. We have found that similar works usually implement the classification accuracy as performance measurement, instead of calculating DET graphs. In addition, due to different procedures and tools, having higher accuracies or lower EERs does not imply better performances, as conditions vary among all of them. There are even discrepancies when working with the proposed database, as the verification performance is obtained considering its extended version.
These drawbacks are also the reason to not test the proposed algorithm in public databases. This work deals with the problem of verification using ECG with significant heart rate variations, but there are no public databases with these characteristics. As most databases are focused on pathology detection, they include non-healthy subjects, colliding with the assumption of healthy sinus rhythm for the QRS segmentation. Moreover, the databases that collect information on healthy users do not provide heart rate changes to test our approach.
Nonetheless, our work provides good results in comparison to the work with the extended version of the database [10]. We need to emphasize that in the extended version, 49 out of 104 subjects provide the Day 1 scenario in both days, making the EERs not comparable. Obtained accuracies for the best solutions (40% and 75% of training data) are also good reaching up to 98.35%. The proposed classifier structure is a very simple ANN that only requires using differentiation in the segmented QRS, avoiding more complex transformations, and decreasing the computational cost, which is important to export algorithms to mobile devices. It also deals with a good number of users and considers different more physiological conditions that are not considered in the ECG-ID database.

4. Discussion

This work proposes a machine learning algorithm for ECG verification by avoiding extra calculations in the feature extraction process. This is achieved by re-using already calculated data in the segmentation process to train the model. Skipping these extra calculations reduces the complexity and time consumed by the system while using simple differentiated QRS complexes. The selected classification algorithm is the MLP neural network, which gets its optimal hyperparameter configuration after a tuning process. To reach this goal, we applied a database which considers specific physical states and different acquisition days to obtain more information about the recognition. The tuning process and the following performance evaluation have provided a deep study in the behaviour of the different types of scenarios.
In addition, the optimal configuration provided an interesting result—the final activation function is identity, which corresponds to a linear function. This feature leads the system to be simplified to a linear system, which implies the possibility of being solved as linear regression. Even though more activations such as ReLU also achieved good results in tuning, finding out the results of this approach could lead to relevant information. Doing further research in this matter can provide information about how these data behave and if the selected features are representative enough, simplifying the classification.
Regarding the final MLP model, we have observed that the first differentiation in the QRS complex provides the best results. From these results, we can infer that how fast the waveform varies is a way to enhance the differences among users. In addition, we have seen how using different sizes of enrolment data affects the performance. The selection of this feature depends on the purposes of the biometric system, as some environments require shorter enrolment processes but do not mind lower performances, while others could choose otherwise. In conclusion, we have shown that using around 100 QRS complexes is good enough for the system, and it is not recommended to exceed 190 in enrolment unless it is known that the recognition data is controlled and heart rates are low enough. These configurations reach the best EER of 2.69% to 4.71% when applying the longest type of enrolment. Nonetheless, looking for more combinations of the signal differentiation could be a new way to improve performances.
The present work also focuses on the differences between three types of physiological situations for the user, due to a lack of public databases with these characteristics. The chosen database collects different visits in different scenarios, proving that MLP is capable of working properly under these conditions. The system also reduces the computational complexity and time with its simplicity. This matter is key for potential algorithm adaptation in mobile devices, setting a good precedent for further research on this issue.
Moreover, we have proven that as it can be expected, different heart rates between the enrolment and recognition data provide lower performance. In addition, even if the scenarios are different, the performance does not decrease as much if the data is taken on the same day. Having huge time differences between data acquisition in enrolment and recognition is unavoidable, so it is very important to have a controlled environment when acquiring enrolment data, being preferable to get it when the user is sitting down and relaxed.
Achieving these results allows us to do further research related to the different stages of a biometric system. Changing the sensor into a more usable one, such as a wearable, would likely provide a signal with lower quality, adding challenges in software. Doing an extension for the original database could also be an interesting approach: adding more types of scenarios such as stress; different physical conditions such as those given by age or exercise frequency or including acquisitions in hot or humid environments. Making ECG biometrics more inclusive with users with pathologies is also a pending task. This could be potentially achieved by avoiding the segmentation process with deep learning techniques.

Author Contributions

Investigation, P.T.-M. (corresponding author); Methodology, P.T.-M.; Supervision, R.S.-R.; Writing – original draft, P.T.-M.; Writing—review and editing, J.L.-J. and J.S.-C. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Ribeiro, A.H.; Ribeiro, M.H.; Paixão, G.M.; Oliveira, D.M.; Gomes, P.R.; Canazart, J.A.; Ferreira, M.P.; Andersson, C.R.; Macfarlane, P.W.; Wagner, M.; et al. Automatic diagnosis of the 12-lead ECG using a deep neural network. Nat. Commun. 2020, 11, 1–9. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  2. Goldberger, A.L.; Goldberger, Z.D.; Shvilkin, A. Clinical Electrocardiography: A Simplified Approach E-Book, 9th ed.; Elsevier: Amsterdam, The Netherlands, 2017. [Google Scholar]
  3. Rajoub, B. Machine Learning in Biomedical Signal Processing with ECG Applications; Elsevier Inc.: Amsterdam, The Netherlands, 2020; pp. 91–112. [Google Scholar] [CrossRef]
  4. Atkielski, A. Schematic Diagram of Normal Sinus Rhythm for a Human Heart as Seen on ECG. Public Domain. 2007. Available online: https://en.wikipedia.org/wiki/File:SinusRhythmLabels.png (accessed on 20 June 2020).
  5. Jain, A.K.; Ross, A.; Prabhakar, S. An Introduction to Biometric Recognition. IEEE Trans. Circuits Syst. Video Technol. (TCSVT) 2004, 14, 1–29. [Google Scholar] [CrossRef] [Green Version]
  6. Simon, B.P.; Eswaran, C. An ECG classifier designed using modified decision based neural networks. Comput. Biomed. Res. 1997, 30, 257–272. [Google Scholar] [CrossRef] [PubMed]
  7. Wübbeler, G.; Stavridis, M.; Kreiseler, D.; Bousseljot, R.D.; Elster, C. Verification of humans using the electrocardiogram. Pattern Recognit. Lett. 2007, 28, 1172–1175. [Google Scholar] [CrossRef]
  8. Biel, L.; Pettersson, O.; Philipson, L.; Wide, P. ECG analysis: A new approach in human identification. IEEE Trans. Instrum. Meas. 2001, 50, 808–812. [Google Scholar] [CrossRef] [Green Version]
  9. Kyoso, M.; Uchiyama, A. Development of an ECG identification system. Annu. Rep. Res. React. Inst. Kyoto Univ. 2001, 4, 3721–3723. [Google Scholar] [CrossRef]
  10. Kim, J.; Sung, D.; Koh, M.J.; Kim, J.; Park, K.S. Electrocardiogram authentication method robust to dynamic morphological conditions. IET Biom. 2019, 8, 401–410. [Google Scholar] [CrossRef]
  11. AliveCor. AliveCor’s Kardiamobile. Available online: https://www.alivecor.es/kardiamobile (accessed on 20 June 2020).
  12. Nymi. Nymi’s Homepage. Available online: https://www.nymi.com (accessed on 20 June 2020).
  13. Palaniappan, R.; Krishnan, S.M. Identifying individuals using ECG beats. In Proceedings of the 2004 International Conference on Signal Processing and Communications, Bangalore, India, 11–14 December 2004; pp. 569–572. [Google Scholar] [CrossRef]
  14. Iqbal, F.T.Z.; Sidek, K.A.; Noah, N.A.; Gunawan, T.S. A comparative analysis of QRS and cardioid graph based ECG biometric recognition in different physiological conditions. In Proceedings of the 2014 IEEE International Conference on Smart Instrumentation, Measurement and Applications (ICSIMA), Kuala Lumpur, Malaysia, 25 November 2014; pp. 25–27. [Google Scholar] [CrossRef]
  15. Wieclaw, L.; Khoma, Y.; Fałat, P.; Sabodashko, D.; Herasymenko, V. Biometrie identification from raw ECG signal using deep learning techniques. In Proceedings of the 9th IEEE International Conference on Intelligent Data Acquisition and Advanced Computing Systems: Technology and Applications, Bucharest, Romania, 21–23 September 2017; Volume 1, pp. 129–133. [Google Scholar]
  16. Satija, U.; Ramkumar, B.; Sabarimalai Manikandan, M. A Review of Signal Processing Techniques for Electrocardiogram Signal Quality Assessment. IEEE Rev. Biomed. Eng. 2018, 11, 36–52. [Google Scholar] [CrossRef]
  17. Srivastava, A.; Yadav, K.T.; Tiwari, R.; Venkateswaran, K. A Brief Study on Noise Reduction Approaches Used in Electrocardiogram. In Proceedings of the Second International Conference on Inventive Research in Computing Applications, Coimbatore, India, 15–17 July 2020; pp. 1023–1027. [Google Scholar]
  18. Choudhary, T.; Manikandan, M.S. A novel unified framework for noise-robust ECG-based biometric authentication. In Proceedings of the 2nd International Conference on Signal Processing and Integrated Networks, Noida, India, 19–20 February 2015; pp. 186–191. [Google Scholar] [CrossRef]
  19. Ribeiro Pinto, J.; Cardoso, J.S.; Lourenco, A. Evolution, current challenges, and future possibilities in ECG Biometrics. IEEE Access 2018, 6, 34746–34776. [Google Scholar] [CrossRef]
  20. Arsene, C.T.; Hankins, R.; Yin, H. Deep learning models for denoising ECG signals. In Proceedings of the European Signal Processing Conference, EURASIP, La Coruña, Spain, 6 September 2019. [Google Scholar] [CrossRef] [Green Version]
  21. Kher, R. Signal Processing Techniques for Removing Noise from ECG Signals. JBER 2019, 3, 1–9. [Google Scholar] [CrossRef] [Green Version]
  22. Pan, J.; Tompkins, W.J. A Real-Time QRS Detection Algorithm. IEEE Trans. Biomed. Eng. (TBME) 1985, 32, 230–236. [Google Scholar] [CrossRef] [PubMed]
  23. Odinaka, I.; Lai, P.H.; Kaplan, A.D.; O’Sullivan, J.A.; Sirevaag, E.J.; Kristjansson, S.D.; Sheffield, A.K.; Rohrbaugh, J.W. ECG biometrics: A robust short-time frequency analysis. In Proceedings of the 2010 IEEE International Workshop on Information Forensics and Security, Seattle, WA, USA, 12–15 December 2010; pp. 1–6. [Google Scholar] [CrossRef]
  24. Singh, Y.N.; Singh, S.K. Evaluation of Electrocardiogram for Biometric Authentication. J. Inf. Secur. (JIS) 2012, 3, 39–48. [Google Scholar] [CrossRef] [Green Version]
  25. Belgacem, N. ECG Based Human Authentication using Wavelets and Random Forests. Int. J. Cryptogr. Inf. Secur. (IJCIS) 2012, 2, 1–11. [Google Scholar] [CrossRef]
  26. Ye, C.; Coimbra, M.T.; Kumar, B.V. Investigation of human identification using two-lead Electrocardiogram (ECG) signals. In Proceedings of the 2010 Fourth IEEE International Conference on Biometrics: Theory, Applications and Systems (BTAS), Washington, DC, USA, 27–29 September 2010. [Google Scholar] [CrossRef]
  27. Hassan, Z.; Gilani, S.O.; Jamil, M. Review of fiducial and non-fiducial techniques of feature extraction in ECG based biometric systems. Indian J. Sci. Technol. (IJST) 2016, 9, 850–855. [Google Scholar] [CrossRef]
  28. Teodoro, F.G.S.; Peres, S.M.; Lima, C.A. Feature selection for biometric recognition based on electrocardiogram signals. In Proceedings of the 2017 International Joint Conference on Neural Networks (IJCNN), Anchorage, AK, USA, 14–19 May 2017; pp. 2911–2920. [Google Scholar] [CrossRef]
  29. Sidek, K.A.; Khalil, I.; Smolen, M. ECG biometric recognition in different physiological conditions using robust normalized QRS complexes. Comput. Cardiol. (CinC) 2012, 39, 97–100. [Google Scholar]
  30. Mai, V.; Khalil, I.; Meli, C. ECG biometric using multilayer perceptron and radial basis function neural networks. In Proceedings of the 2011 Annual International Conference of the IEEE Engineering in Medicine and Biology Society, Boston, MA, USA, 30 August–3 September 2011; pp. 2745–2748. [Google Scholar] [CrossRef]
  31. Pelc, M.; Khoma, Y.; Khoma, V. ECG signal as robust and reliable biometric marker: Datasets and algorithms comparison. Sensors 2019, 19, 2350. [Google Scholar] [CrossRef] [Green Version]
  32. Scikit-learn developers. Neural Network Models (supervised). Available online: https://scikit-learn.org/stable/modules/neural_networks_supervised.html (accessed on 11 May 2020).
  33. Goldberger, A.L.; Amaral, L.A.N.; Glass, L.; Hausdorff, J.M.; Ivanov, P.C.; Mark, R.G.; Mietus, J.E.; Moody, G.B.; Peng, C.K.; Stanley, H.E. PhysioBank, PhysioToolkit, and PhysioNet: Components of a New Research Resource for Complex Physiologic Signals. Circulation 2000, 101, e215–e220. [Google Scholar] [CrossRef] [Green Version]
  34. Bousseljot, R.; Kreiseler, D.; Schnabel, A. Nutzung der EKG-Signaldatenbank CARDIODAT der PTB über das Internet. Biomed. Eng. Biomed. Tech. (BMT) 1995, 40, 317–318. [Google Scholar] [CrossRef]
  35. Nemirko, A.P.; Lugovaya, T.S. Biometric human identification based on electrocardiogram. In The 12th Russian Conference on Mathematical Methods of Pattern Recognition; MAKS Press: Moscow, Russia, 2005; pp. 387–390. [Google Scholar]
  36. ISO/IEC JTC 1/SC 37. Information Technology—Biometric Performance Testing and Reporting—Part 1: Principles and Framework (Standard No. 19795); International Organization for Standardization (ISO): Geneve, Switzerland, 2006; p. 56. [Google Scholar]
  37. Varoquaux, G.; Buitinck, L.; Louppe, G.; Grisel, O.; Pedregosa, F.; Mueller, A. Scikit-learn: Machine Learning in Python. J. Mach. Learn. Res. (JMLR) 2011, 12, 2825–2830. [Google Scholar] [CrossRef]
  38. Panchal, G.; Ganatra, A.; Kosta, Y.P.; Panchal, D. Behaviour Analysis of Multilayer Perceptrons with Multiple Hidden Neurons and Hidden Layers. Int. J. Comput. Theory Eng. (IJCTE) 2011, 3, 332–337. [Google Scholar] [CrossRef] [Green Version]
  39. Kingma, D.P.; Ba, J.L. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
  40. Bergstra, J.; Bengio, Y. Random search for hyper-parameter optimization. J. Mach. Learn. Res. 2012, 13, 281–305. [Google Scholar]
Figure 1. Normal Sinus Rhythm waveform and some of its relevant points [4].
Figure 1. Normal Sinus Rhythm waveform and some of its relevant points [4].
Applsci 10 06896 g001
Figure 2. Scheme for signal pre-processing and data segmentation. Variable r n g 1 represents the data points on the left of the R peak and r n g 2 the data points after, including the R peak. N p e a k s refers to the number of selected R peaks in the matrix.
Figure 2. Scheme for signal pre-processing and data segmentation. Variable r n g 1 represents the data points on the left of the R peak and r n g 2 the data points after, including the R peak. N p e a k s refers to the number of selected R peaks in the matrix.
Applsci 10 06896 g002
Figure 3. Multilayer Perceptron (MLP) with one hidden layer.
Figure 3. Multilayer Perceptron (MLP) with one hidden layer.
Applsci 10 06896 g003
Figure 4. Hyperparameter tuning scheme.
Figure 4. Hyperparameter tuning scheme.
Applsci 10 06896 g004
Figure 5. Detection Error Trade-off (DET) for Ex as train and test set.
Figure 5. Detection Error Trade-off (DET) for Ex as train and test set.
Applsci 10 06896 g005
Figure 6. Performance evaluation scheme for different test sets when using R1 with xD signals in training.
Figure 6. Performance evaluation scheme for different test sets when using R1 with xD signals in training.
Applsci 10 06896 g006
Figure 7. Results for every test set using (a) ND, (b) FD, and (c) SD signals with 50% of R1 as training.
Figure 7. Results for every test set using (a) ND, (b) FD, and (c) SD signals with 50% of R1 as training.
Applsci 10 06896 g007
Figure 8. Results for every test set using (a) ND, (b) FD, and (c) SD signals with 50% of S as training.
Figure 8. Results for every test set using (a) ND, (b) FD, and (c) SD signals with 50% of S as training.
Applsci 10 06896 g008
Figure 9. Results for every test set using (a) ND, (b) FD, and (c) SD signals with 50% of R2 as training.
Figure 9. Results for every test set using (a) ND, (b) FD, and (c) SD signals with 50% of R2 as training.
Applsci 10 06896 g009
Figure 10. Results for every test set using (a) ND, (b) FD, and (c) SD signals with 50% of Ex as training.
Figure 10. Results for every test set using (a) ND, (b) FD, and (c) SD signals with 50% of Ex as training.
Applsci 10 06896 g010
Figure 11. Performance results with different sizes of enrolment for (a) S, (b) R2, and (c) Ex data as test set with the final MLP configuration.
Figure 11. Performance results with different sizes of enrolment for (a) S, (b) R2, and (c) Ex data as test set with the final MLP configuration.
Applsci 10 06896 g011
Table 1. Most common activation functions.
Table 1. Most common activation functions.
NameIdentityLogisticTanhReLU
Formula g ( x ) = x g ( x ) = 1 1 + e x g ( x ) = t a n h x g ( x ) = m a x ( 0 , x )
Table 2. Selected database visit distribution. The four visits are re-named as shown, based on the type of scenario: R1 and R2 for resting while sitting down, S for resting while standing up, and Ex for exercise.
Table 2. Selected database visit distribution. The four visits are re-named as shown, based on the type of scenario: R1 and R2 for resting while sitting down, S for resting while standing up, and Ex for exercise.
Day 1Day 2
Visit 1 (R1)Visit 2 (S)Visit 1 (R2)Visit 2 (Ex)
Resting, sitting downResting, standingResting, sitting downAfter exercise (average 130 bpm)
Table 3. Summarization of the fixed MLP hyperparameters.
Table 3. Summarization of the fixed MLP hyperparameters.
HyperparameterDescriptionValue
Hidden layersNumber of hidden layers.1
SolverFunction used for weight updating.adam
Learning rateFunction used to update the learning rate
that takes part in the solver.
Not applicable
with adam solver.
Learning rate stepStep size for the learning rate updates.0.0001
No change iterationsNumber of iterations with no relevant
change to consider convergence.
10
Table 4. Possible values for the remaining hyperparameters.
Table 4. Possible values for the remaining hyperparameters.
HyperparameterPossible Values
Hidden layer size1, 25, 50, 75, 100, 125, 150, 175, 200, 225, 250
ActivationIdentity, logistic, tanh, ReLU
Alpha0.0001, 0.0005, 0.001, 0.005, 0.01, 0.05, 0.1, 0.5
Tolerance0.01, 0.05, 0.1, 0.5
Table 5. Best hyperparameter values in Random Search and their Equal Error Rates (EERs) for Non-Differentiation (ND), First Differentiation (FD), and Second Differentiation (SD).
Table 5. Best hyperparameter values in Random Search and their Equal Error Rates (EERs) for Non-Differentiation (ND), First Differentiation (FD), and Second Differentiation (SD).
SignalHidden Layer SizeActivationAlphaToleranceMean EER (%)
ND200logistic0.010.050.13
100logistic0.010.050.23
125logistic0.10.010.50
FD200identity0.0010.050.16
200relu0.00050.051.27
75logistic0.00010.011.54
SD150tanh0.00010.010.95
125identity0.00010.011.28
175tanh0.0050.014.59
Table 6. Best hyperparameter values in Grid Search and their EERs for ND, FD and SD.
Table 6. Best hyperparameter values in Grid Search and their EERs for ND, FD and SD.
SignalHidden Layer SizeActivationAlphaToleranceMean EER (%)
ND200logistic0.010.050.13
FD200identity0.0010.050.16
SD150tanh0.00010.010.95
Table 7. EER (%) results for every type of signal under the same visit.
Table 7. EER (%) results for every type of signal under the same visit.
R1SR2Ex
ND0.100.1300.17
FD0000.07
SD0.2000.200.77
Table 8. Final MLP configuration.
Table 8. Final MLP configuration.
Number of hidden layers1
Hidden layer size200
ActivationIdentity
Alpha0.001
Tolerance0.05
SolverAdam
Learning rate step0.0001
No change iterations10
Table 9. EER (%) for different enrolment proportions when using R1 FD for enrolment.
Table 9. EER (%) for different enrolment proportions when using R1 FD for enrolment.
VisitSR2Ex
Enrol Proportion
25%5.423.645.05
40%3.642.794.95
50%3.643.005.45
60%2.732.465.42
75%2.692.094.71
100%2.021.885.52
Table 10. Accuracy (%) for different enrolment proportions when using R1 FD for enrolment.
Table 10. Accuracy (%) for different enrolment proportions when using R1 FD for enrolment.
VisitSR2Ex
Enrol Proportion
25%97.4897.5597.48
40%97.8297.8997.78
50%97.9898.0597.78
60%98.1198.1798.04
75%98.2698.3598.15
100%98.4098.4998.26
Table 11. Summary of results and tools from state-of-the-art with similar characteristics.
Table 11. Summary of results and tools from state-of-the-art with similar characteristics.
ReferenceDatabaseUsersFeaturesClassifierResult
[15]Private18HeartbeatsMLPAccuracy: 89%
[31]ECG-ID90HeartbeatsMLP
LDA
Accuracy: 84–89.3%
Accuracy: 93.28%
[10]Proposed database
(extended)
104SWT
for heartbeats
LDAEER: 1.74–5.47%
Proposed workProposed database
(only users with exercise)
55QRS complexMLPShort enrol:
EER: 2.79–4.95%
Accuracy: 97.78–97.89%
Long enrol:
EER: 2.69–4.71%
Accuracy: 98.15–98.35%

Share and Cite

MDPI and ACS Style

Tirado-Martin, P.; Liu-Jimenez, J.; Sanchez-Casanova, J.; Sanchez-Reillo, R. QRS Differentiation to Improve ECG Biometrics under Different Physical Scenarios Using Multilayer Perceptron. Appl. Sci. 2020, 10, 6896. https://doi.org/10.3390/app10196896

AMA Style

Tirado-Martin P, Liu-Jimenez J, Sanchez-Casanova J, Sanchez-Reillo R. QRS Differentiation to Improve ECG Biometrics under Different Physical Scenarios Using Multilayer Perceptron. Applied Sciences. 2020; 10(19):6896. https://doi.org/10.3390/app10196896

Chicago/Turabian Style

Tirado-Martin, Paloma, Judith Liu-Jimenez, Jorge Sanchez-Casanova, and Raul Sanchez-Reillo. 2020. "QRS Differentiation to Improve ECG Biometrics under Different Physical Scenarios Using Multilayer Perceptron" Applied Sciences 10, no. 19: 6896. https://doi.org/10.3390/app10196896

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop