1. Introduction
Personal devices—such as smartphones, smartwatches, wearables, or smart rings—have become invaluable tools for collecting physiological data [
1], including heart rate [
2] and blood pressure [
3,
4] among others. These devices not only enhance health monitoring but also open new possibilities for authentication systems. Traditional methods, such as passwords and PINs, rely on knowledge-based credentials (“what you know”), making them vulnerable to issues like phishing and theft. In contrast, biometrics—the research field that studies how individuals can be uniquely recognized from their physical, chemical, or behavioral attributes [
5]—rely on traits inherently possessed by individuals (“what you are”). Biometrics operates on the foundational assumption that unique patterns or characteristics can reliably distinguish one individual from another. These traits encompass facial features [
6], fingerprints [
7], and anatomical structures [
8,
9], as well as eyebrow and the nose-tip features for facial recognition [
10], and the structure of the blood vessels underneath the palm skin [
9]. Among these, a widely used physiological signal captured by most wearable devices—including the Apple Watch [
11], Oura Ring [
12], and Samsung Watch [
13]—has demonstrated the ability to encode unique characteristics of heart mechanics and pulse [
14]. This signal also conveys information about capillary blood vessels [
9], skin tone, and body mass index [
15]. These features collectively enable the unique characterization of individuals based on data from wrist-worn wearables [
16].
While prior research has demonstrated that PPG can be effectively used for biometrics [
17,
18,
19,
20], their evaluation methods and the datasets used introduce limitations that can overestimate performance. Specifically, we identify four key factors that impact the validity and generalizability of PPG-based biometric systems: (1) the prevalence of controlled experimental settings, which often fail to capture the variability in physiological data seen in real-world environments; (2) the number of users in the datasets, which may simplify recognition tasks and raise questions about their scalability; (3) the need to address scenarios where new users attempt to authenticate; and (4) the lack of evaluation methods that separate training and test data across distinct time spans, preventing an assessment of model performance on new data collected after an initial training phase. Although these issues are recognized in biometrics research broadly, systematic empirical assessments within the context of PPG-based methods are still lacking. Most studies mention these factors qualitatively [
21], but few explicitly measure their individual effect on model reliability.
Three of these four factors stem from dataset limitations, a common issue in the wearable domain due to the challenges of collecting real-world longitudinal data [
22,
23,
24]. To systematically assess these constraints, we leverage the LAUREATE dataset [
25], one of the few longitudinal datasets that provide raw PPG data from 44 participants. This dataset enables a more rigorous evaluation of model performance over extended periods, addressing the need for real-world variability and long-term assessment. We replicate two prior studies [
26,
27] and introduce a feature-based approach that incorporates both time- and frequency-domain characteristics as a baseline. We then assess the impact of dataset size, real-world variability, unseen users, and temporal separation on biometric performance.
Our findings indicate that PPG-based user recognition is unreliable in real-world conditions. Models that achieve a Matthews Correlation Coefficient (MCC) of 0.94 in controlled settings suffer performance drops of up to 0.61 points when tested on real-world data. Specifically, on the LAUREATE dataset, the MCC declines from 0.70 to 0.09 for one method and from 0.94 to 0.44 for another, underscoring the impact of real-world variability on biometric performance. We further show that small dataset sizes artificially inflate results, as models trained on fewer users face a simpler classification task. Moreover, we highlight that the often-overlooked open-set scenario, where new users not present during training attempt authentication, should be considered. Finally, we observe that random train–test splits substantially inflate performance. When the training and test sets are instead separated in time—preventing the model from exploiting short-term temporal similarities—the MCC drops below 0.17. This indicates that the models do not learn stable biometric characteristics but rather rely on transient patterns present within closely sampled data. To support more rigorous evaluation in future work, we also outline a set of guidelines for PPG-based biometrics, emphasizing the importance of adequate dataset size, real-world variability, unseen users, and proper temporal separation.
The main contributions of this paper are as follows:
A systematic categorization of biometric techniques presented in the literature;
Replication of prior PPG-based biometric methods and comparative evaluation of their performance under real-world constraints;
Identification and assessment of factors that may inflate reported biometric performances in the literature;
Recommendations for an evaluation setup that robustly assesses the performance of biometric models.
The code used to run the analyses presented in this study is available on GitHub:
https://github.com/LidiaAlecci/PotentialPitfallsBiometricPPG (accessed on 9 December 2025). The LAUREATE dataset is available upon signing a data-sharing agreement, whereas the BIDMC PPG and Respiratory dataset and the Real-World PPG dataset are publicly accessible.
2. Background and Related Work
2.1. User Recognition Taxonomy
The reviewed literature uses different terms for user prediction within a dataset:
verification,
identification,
authentication,
re-identification, and
recognition. For example, Bianco and Napoletano [
28] use
recognition and
identification interchangeably; they refer with both terms to a classification task in which a model, given an input, determines which user from the training set it belongs to. However, Biswas et al. [
19] implement
identification using a one-vs.-rest approach, i.e., a separate model is trained for each user to distinguish that specific user from all others. Similarly, Cornelius et al. [
29] consider
identification as the one-vs-rest approach, using
verification for the classification approach. Lastly, both Almuashi et al. [
30] and Peralta et al. [
31] in their literature surveys define
identification as the identification of
n users, and
verification as the binary classification task (one-vs-rest).
We categorize biometric research into three tasks:
verification,
identification, and
re-identification, with
recognition used as the overarching term. This categorization is based on how each system is designed and its intended purpose; a summary of our terminology is provided in
Table 1.
In
identification, the system aims to distinguish an individual from a population, involving a classification task with multiple classes corresponding to
n possible users, similarly to Piciucco et al. [
18]. In
verification, the system seeks to validate a person’s identity, resulting in a binary classification task where the model discriminates between the user it is trained on and all the others as in Biswas et al. [
19].
re-identification, first defined in the domain of multi-camera surveillance systems [
32], focuses on identifying the same person appearing in non-overlapping or disjoint camera views, accounting for variations in lighting conditions, poses, and camera perspectives. In the field of biometrics, Seok et al. [
26] address re-identification using PPG, while Wu et al. [
33] address re-identification within video surveillance systems.
The distinction between re-identification and identification lies in the system’s design. In re-identification, all samples are cross-referenced to create pairs from the original pool of samples, and the system determines whether two samples correspond to the same user or not. Since re-identification focuses solely on pairwise similarity rather than explicit identity assignment, it can generalize to users never encountered during training. In contrast, identification involves a system that, given a sample as input, assigns it to one of the users seen during training. This approach inherently assumes that all possible users are present in the training set. As a result, an identification model must be retrained to accommodate new users, whereas re-identification remains applicable even when encountering users not seen during training.
2.2. Biometrics from PPG
In
Table 2, we summarize current state-of-the-art approaches for identification, verification, and re-identification using the PPG signal. While identifying users through PPG is not a new approach, there remains a gap in evaluating solutions within real-world scenarios, where external factors such as noise and emotional variations can affect the signal, and on longitudinal datasets spanning months of recordings.
Piciucco et al. [
18] explore biometrics using electrodermal activity (EDA) and PPG signals. The dataset used comprises 17 participants engaged in lectures, with data collected during two separate sessions spaced seven days apart. The study employs convolutional neural networks (CNNs) to extract features from the spectrograms of the EDA and PPG signals. The highest accuracy is achieved using a multimodal approach combining both signals, yielding an average correct identification rate of 98.58%. Similarly, Hinatsu et al. [
27] focus on biometric identification using PPG signals but with an emphasis on both heartbeat and respiration-derived features, such as the maximum, minimum, and standard deviation of the pulsatile, non-pulsatile, and respiratory components. Hinatsu et al. [
27] extract Mel-Frequency Cepstrum Coefficients from the PPG signals to capture additional physiological characteristics. Their method, tested on a dataset of 46 users, achieves an accuracy of 92.9% using a random forest classifier. Jindal et al. [
16] instead opt for a deep learning approach, applying Restricted Boltzmann Machines and Deep Belief Networks for classification. They use the TROIKA dataset containing PPG signals from 12 male users, who were recorded during walking and running activities at various speeds. Preprocessing steps include filtering to remove motion artifacts, segmentation, and feature extraction. The method achieves an accuracy of 96.1% using 10-fold cross-validation. Seok et al. [
26] investigate PPG-based biometric authentication using a Siamese 1D convolutional neural network within a re-identification framework, reporting 97.23% accuracy on 35 subjects.
From the results in the mentioned literature, recognizing users across datasets seems feasible. However, we show empirically that prior research tends to overestimate the reliability of their findings for several reasons. For instance, Hinatsu et al. [
27] and Jindal et al. [
16] evaluate their methods exclusively on datasets collected in controlled laboratory settings, leaving their applicability to real-world environments untested. Although Piciucco et al. [
18] utilize a real-world dataset, the study is limited to only 17 users, which restricts the generalizability of the results.
In summary, while the existing literature demonstrates promising results in biometric, several factors suggest that these findings might be overly optimistic. Many studies rely on small sample sizes or controlled environments and lack proper evaluation, raising concerns about their generalizability and scalability in real-world settings. Therefore, in this paper, we critically assess how current PPG-based biometric approaches fail to sustain their effectiveness outside controlled settings. Additionally, we propose guidelines for an evaluation framework that enables future research to design testing setups that more accurately reflect real-world scenarios. While the need to account for environmental variability and unseen users has been discussed in other biometric modalities such as face or fingerprint recognition, their quantitative impact on physiological biometrics remains underexplored.
3. Datasets, Algorithms and Evaluation Metrics
Reproducing previous studies is often challenging due to the unavailability of code or datasets. In our case, as shown in
Table 2, no implementation code was provided for the majority of relevant prior works, and only a few datasets were publicly available. To determine which studies to reproduce, we established the following criteria: the study must use PPG data and rely on a publicly available dataset. The latter criterion is essential, as access to the dataset allows us to verify the implementation of the methods described in their papers and compare our results with their reported performance.
Therefore, we selected the following two studies for comparison: Hinatsu et al. [
27] and Seok et al. [
26].
3.1. Datasets
3.1.1. BIDMC PPG and Respiration Dataset
The dataset presented by Pimentel et al. [
40], hereinafter referred to as the BIDMCPPG dataset, contains PPG signals from 53 adults (median age: 64.81, age range: 19–90+, 32 females), sampled at a frequency of 125 Hz, with each session lasting 480 s. This dataset is used by Hinatsu et al. [
27], although their implementation relies on a subset of 46 subjects rather than the full cohort.
3.1.2. Real-World PPG Dataset
The dataset presented by Siam [
41], hereinafter referred to as the RWPPG dataset, contains PPG signals from 35 users. Most recordings are collected during a single, long session on the same day for each participant and then divided into 6 s segments [
42] sampled at 50 Hz. The dataset is already partitioned into training and test sets: the training set contains approximately 40 samples per participant (1374 samples in total), while the test set includes roughly 20 samples per participant (700 samples in total). Train and test splits, obtained by random selection, are provided by the authors [
43]. This dataset is used by Seok et al. [
26].
3.1.3. LAUREATE
The LAUREATE dataset [
25] was collected during two university courses and includes data from 44 participants: 42 students and two professors.
Each course consisted of biweekly lectures over a 13-week period, totaling 52 recorded sessions. A typical lecture lasted 90 min and followed a frontal teaching format, though some sessions included laboratory activities, quizzes, or exams. In total, the dataset comprises over 1400 h of physiological recordings. For our study, we focus on the 42 students, excluding the professors to prevent artificially inflated results. Professors generally exhibit greater movement during lectures, introducing motion artifacts in the PPG data. This increased variability makes them easier to distinguish from students, thereby facilitating their recognition. To mitigate
device bias—a key challenge in
user recognition [
44,
45]—participants were assigned different devices for each lecture.
We restrict our analysis to data collected during frontal lectures, excluding quizzes, exams, and laboratory sessions. Including these additional sessions substantially increases vigorous movement and device manipulation, introducing severe motion artifacts. Although such artifacts naturally occur in real-world settings, their magnitude strongly depends on device placement—here, the wrist—where even moderate movement can markedly degrade the PPG signal. Because our goal is to evaluate PPG-based authentication under realistic but not extreme conditions, we retain only frontal lecture sessions, where movement is limited yet still naturalistic.
As a result, the final dataset includes 42 participants and comprises around 700 h of recorded data. We selected the LAUREATE dataset for our analysis because it is a longitudinal dataset spanning several months of recordings. By capturing physiological changes influenced by real-world conditions, it provides a unique opportunity to assess how user recognition models handle variability over time.
3.2. Algorithms
Hinatsu et al. [
27] employ a shallow machine learning technique (i.e., Random Forest), while Seok et al. [
26] utilize a deep learning model (i.e., Siamese Neural Network). This selection allows our findings to be generalizable across both paradigms—i.e., both classical machine learning and deep learning approaches. As the original code for both studies is unavailable, we made specific methodological choices in cases of ambiguity or insufficient procedural details. The
Supplementary Materials provides further details to support the reproducibility of our analysis in future research. Our goal is not to outperform these models but to use them as representative anchors for evaluating robustness under different data regimes. In addition to replicating their methods, we developed a baseline approach, hereinafter referred to as feature-based method.
3.2.1. Hinatsu et al. [27]’s Method
They propose a PPG-based subject identification method, leveraging features from heartbeat and respiration for continuous authentication. The signal is segmented into 40 s windows and decomposed into three components: AC (high-frequency), RS (mid-frequency for respiration), and DC (low-frequency for blood volume). Feature extraction includes 20 spectral and statistical features from AC and RS, 5 additional features from DC, and MFCC coefficients. A Random Forest classifier is trained using a leave-one-out evaluation, i.e., one data window is held out at a time while the remaining samples are used for training. More implementation details are provided in the
Supplementary Materials and in the authors’ original paper paper [
27].
3.2.2. Seok et al. [26]’s Method
The authors developed a PPG-based re-identification system using a one-dimensional Siamese neural network. The signal is first segmented into 6 s windows, followed by linear detrending to enhance peak detection. Feature extraction involves peak detection, quadratic spline interpolation for cycle standardization, and multicycle averaging to improve signal consistency. Evaluation is conducted using the predefined train–test split from the RWPPG dataset. More implementation details are provided in the
Supplementary Materials and in their paper [
26].
3.2.3. Our Feature-Based Method
We extract multiple signal components from the raw PPG signal: Fast Fourier Transform, Welch’s method, Wavelet Daubechies 1, Wavelet Daubechies 4, and Wavelet Symlets 4. A total of 160 features are extracted, comprising 144 time-domain and 16 frequency-domain features, computed across all components. The model we employ is XGBoost. Full implementation details are available in the
Supplementary Materials. The evaluation setup and window size are tailored to each task and adjusted based on whether the experiment includes a comparison between feature-based method and the other two algorithms. In comparative cases, the evaluation approach and window size are aligned with those of the other algorithms to ensure a fair and consistent comparison. The specific evaluation procedure for each experiment is detailed before presenting the results. In a re-identification scenario, our feature-based method concatenates the two inputs, which are then fed into the XGBoost model.
3.3. Metrics
We report results using both accuracy and balanced accuracy as commonly recommended in the literature [
46]. Since accuracy alone is not suitable for unbalanced datasets [
47]—as is the case for most of our of our experiments—we also include the Matthews Correlation Coefficient (MCC) [
48], a metric that is robust to class imbalance. Although the Matthews Correlation Coefficient (MCC) was originally introduced for binary classification problems [
48], it admits a straightforward extension to the multi-class case based on the full confusion matrix, where it still measures the correlation between true and predicted labels. In line with recommendations for imbalanced classification tasks [
47], we therefore report the multi-class MCC in our identification experiments, computing it over all users (42 or 53 classes depending on the dataset). This yields a robust performance metric that remains informative even when class distributions are skewed and accuracy alone may be misleading.
Across all experiments, each model is initialized using ten different random seeds. In the tables, we report the standard error as a subscript for each metric, while the plots display the standard error as error bars. When statistical tests are performed, we apply a one-sided Wilcoxon signed-rank test [
49], with Bonferroni correction for multiple comparisons [
50].
4. Results
In this section, we present the results of our study. First, we benchmark the different methods using the datasets used in the literature. We then describe different experiments assessing the factors that, to our understanding, contribute to the overestimation of biometric results in the literature. Rather than pursuing state-of-the-art accuracy, this analysis emphasizes how methodological design choices influence the reported performance of biometric systems. By systematically quantifying the extent of overestimation arising from common evaluation practices, our work complements algorithmic research with an evidence-based perspective on experimental rigor.
4.1. Benchmarking
As the first step of our analysis, we try to reproduce two studies [
26,
27] from the literature whose datasets are publicly available, to create a benchmark of the different methods and datasets. We follow the same evaluation procedure and use the same window size as the original studies. For the BIDMCPPG dataset, we apply a leave-one-out approach—i.e., one test sample is held out at each iteration, and the remaining samples are used for training—and use a 40 s window [
27], while for the RWPPG dataset we use the train–test split provided by the dataset authors with a 6 s window [
26]. We report the results of the benchmark in
Table 3.
Notably, our implementation of Hinatsu et al. [
27] achieves higher accuracy compared to the results published in their paper, with an improvement of approximately 0.02 percentage points. Likewise, our re-implementation of Seok et al. [
26] attains an accuracy that is 0.01 percentage points higher. These differences may stem from random variation in the results or from minor discrepancies in our re-implementation (see the
Supplementary Materials for more information).
Nevertheless, we consider this difference negligible and unlikely to significantly impact our overall conclusions. Our implementation of the method described by Seok et al. [
26] achieves the same accuracy reported in their paper. For this method, we also rely on some assumptions to re-implement the original method, described in detail in the
Supplementary Materials.
Our feature-based method and our implementations of previous works achieve over 0.89 accuracy, underscoring the relative ease of these tasks under controlled conditions. However, we believe these results may be overestimated for several reasons. First, both datasets are collected in laboratory settings where physiological data, though still variable, may not reflect the natural fluctuations encountered in real-world environments [
51]. Second, these works do not account for the challenge of unseen users attempting to authenticate into the system, which is crucial for real-world applications. Finally, these methods have not been evaluated on temporally disjoint data, leaving their long-term effectiveness uncertain.
To address these limitations, in
Section 4.2, we further test both our feature-based method and our implementation of Hinatsu et al. [
27] and Seok et al. [
26] methods on a real-world, longitudinal dataset (LAUREATE) to better evaluate their applicability in practical, everyday scenarios.
4.2. Laboratory vs. Real-World Scenarios
A significant number of experiments are conducted in controlled laboratory environments [
16,
27,
38], where researchers attempt to isolate specific variables. However, these controlled settings do not necessarily translate to real-world conditions, and often fail to capture the inherent variability in human physiology [
52]. Physiological responses are not only highly variable between individuals but can also fluctuate within the same individual depending on the context and even time of day [
53,
54]. This situation not only complicates the consistency and reproducibility of lab experiments but also challenges the assumption that the methods developed and tested in controlled environments are applicable to real-world settings.
To evaluate how these methods perform in a real-world scenario, where users are exposed to external stimuli such as emotions and movement, we test our feature-based method and our implementation of the two previously used methods on the LAUREATE dataset [
25]. To isolate and assess only the impact of the real-world environment, we keep all other relevant factors as close as possible to the previous setting.
First, for each user in the LAUREATE dataset, we only use data from one day—namely, the user’s first day in the dataset. Second, we also segment the data as in the original studies, i.e., 40 s windows for identification (our implementation of Hinatsu et al. [
27]) and 6 s windows for re-identification (i.e., Seok et al. [
26]). Third, we also match the number of samples, using 12 samples per user for identification and 60 for re-identification. Furthermore, to ensure good quality readings (i.e., comparable to laboratory studies) we also discard the first and last 5 min of each session, ensuring samples are taken while students are engaged in class, avoiding noise from putting on or removing the device. We refer to this dataset as “LAUREATE one-day”. Finally, we use the same evaluation procedure as in the previous experiment: leave-one-out approach for identification, and a random 66%-34% train–test split for re-identification.
As shown in
Table 4, the accuracy obtained on the LAUREATE dataset—collected in the real-world—is statistically lower than the results achieved in laboratory settings (reported in
Table 3). For the identification task, the MCC comparison between the laboratory dataset and the LAUREATE dataset reveals a disparity of up to 0.50 MCC points between the two settings— Hinatsu et al. [
27] on their laboratory dataset (0.94) versus the same method on the real-world LAUREATE dataset (0.44). This decline is even more pronounced in the re-identification task, where the MCC difference between the laboratory setting and the LAUREATE dataset reaches up to 0.61 MCC points—0.70 using the Seok et al. [
26] method on the laboratory dataset versus 0.09 on the real-world LAUREATE dataset. This underscores how real-world conditions, where various factors can degrade physiological data quality, add a level of complexity that methods tested only on laboratory data may struggle to manage effectively.
We emphasize that relying solely on accuracy for the re-identification task is misleading, as the task is inherently imbalanced, with the positive class significantly underrepresented in the test set. As Chicco and Jurman [
47] highlight, traditional metrics often fail to provide a meaningful assessment in such cases. Specifically, although our feature-based method for the re-identification task on the LAUREATE one-day dataset achieves an accuracy of 0.82 and a balanced accuracy of 0.76, these values are misleadingly high, as the MCC is only 0.20, indicating poor overall performance.
In the next two experiments, we focus on the best-performing model and the task that yields the highest accuracy. Therefore, we select the identification task and use only our feature-based method.
4.3. Impact of the Number of Users
The datasets used by Hinatsu et al. [
27] and Seok et al. [
26] include 53 and 35 participants, respectively. In contrast, most studies summarized in
Table 2 involve approximately 20 users. While training and testing methods on a limited number of participants already poses generalization issues in domains like human activity recognition [
55], we argue that this challenge is even more critical in the biometric domain. In this context, the number of users determines the number of outputs of the model. Consequently, having fewer users increases the probability of achieving a correct prediction by chance.
To evaluate the impact of the number of participants on prediction performance, we conduct the following experiment. Using only the first day of data for each participant from the LAUREATE dataset, we select a 30 min segment from the middle of the lecture to ensure data quality and maintain balance among all users. This approach aligns with the methodology used by Piciucco et al. [
18] in a similar scenario. Then, we segment the data into 40 s windows, and randomly select 30% as the test set. For this experiment, we use our feature-based method, which demonstrates the highest performance in the closely related previous task (reported in
Table 4).
As shown in
Figure 1, the performance of the model is inversely proportional to the number of users in the dataset. This finding mirrors the results observed by Sugrim et al. [
56] in different domains, such as localization predictions and sports performance. Specifically, in our case, MCC is nearly perfect with only two users but declines rapidly until reaching eighteen users. By plotting these results, we can also assess whether the model remains stable as the number of users increases. Notably, with our model applied to the LAUREATE one-day dataset, performance tends to stabilize around 35 users. For comparison, we also report a random baseline alongside our feature-based method.
4.4. Unseen Users at Testing Time
A key limitation in many existing studies ([
18,
19,
34]) is their focus on verifying users who have already been registered and included in the system’s training data (here referred to as
known or
previously seen users). However, these studies often overlook the system’s vulnerability when faced with users who attempt authentication but were not part of the initial training data. In some cases, training and testing may indeed involve the same set of users, for example, in scenarios where users undergo an initial authentication step and the system’s only role is to assign permissions or manage access within this known group. However, systems designed for open environments must account for the possibility of encountering unknown users. Since it is infeasible to train an authentication system on every potential user, these systems must develop robust strategies to effectively reject unseen users, risking otherwise to incorrectly authenticate unauthorized users. This issue is particularly problematic in machine learning-based identification methods, where models can only predict from a fixed set of classes corresponding to the users seen during training. When an unseen user attempts to authenticate, the model is forced to assign them to one of the known classes, potentially granting access to unauthorized individuals.
To illustrate how the number of unseen users affects a system’s performance, we use the same setup as described in
Section 4.3. We select data from the first day for each participant in the LAUREATE dataset, extracting a 30 min segment from the middle of the lecture on that day and we segment the data into 40 s windows.
To analyze how performance changes as the number of unseen users increases, we start with 15 users in the training set, leaving up to 27 users as unseen cases for evaluation. As shown in
Figure 2, the performance of the model decreases as new unseen users are added to the test set. This decline highlights the inability of the model to effectively handle unseen users, causing it to incorrectly assign them to one of the existing classes. Consequently, the number of false positives increases in direct proportion to the number of unseen user samples and, therefore, the number of unseen users. For comparison, we also report a random baseline alongside our feature-based method.
Notably, there are several approaches for developing a model that can handle unseen users. Techniques like open-set recognition [
57]—where the model is designed to determine whether a given input belongs to a known class or an unknown one—and anomaly detection algorithms can be used to identify and reject unfamiliar users. Additionally, instead of focusing on identification, alternative methods such as verification, where a unique model is created for each user to distinguish that user from others, or re-identification, where the input is compared with all samples encountered during training and either matched with the correct identity or rejected, can also be considered.
4.5. Temporal Separation of Training and Test Data
A common limitation in many machine learning studies involving time-series data, such as in [
26,
58], is the lack of evaluation methods that separate training and test data across distinct time spans. Without this separation, performance metrics are usually inflated [
59], as models tend to leverage short-term sequential patterns rather than genuinely learn relevant predictive features. In contrast, evaluating the model on data from a later, separate period provides a measure of its ability to generalize and predict beyond immediate, localized patterns.
To assess how performance is affected when modifying the evaluation procedure (i.e., from a random split of training and test sets to a more robust scenario), we split the LAUREATE dataset along the temporal dimension. We use the first half of the dataset for training, while we reserve the latter half for testing. We then apply the same models used in
Section 4.2, including our method and our implementation of those from Hinatsu et al. [
27] and Seok et al. [
26]. To compare the results of this new experiment with those from
Section 4.2, which uses only one day of data from the LAUREATE dataset, we select a subset that allows for a similar training set size. Specifically, for the identification task, we extract 24 samples per user. This allows for an even split of 12 samples for training and 12 for testing, resulting in a training set size that matches that of the previous experiment. Similarly, for the re-identification task, we select 120 samples per user. To maintain balance among users, we evenly distribute the samples across the days when each participant is present. For example, if a participant attended 12 days and we need to extract 24 samples per user, we select 2 samples from each day. The sampled windows are randomly chosen from all 30 min lecture segments, centered within the lecture as described in
Section 4.3.
As shown in
Table 5, performance drops significantly when training and test sets are separated along the temporal axis. In this scenario, the MCC is 0.07 and 0.08 for the identification task and 0.02 and 0.06 for the re-identification task, indicating that the model fails to reliably distinguish users. This performance decline can be attributed to two main factors. First, the model struggles to identify users when their individual patterns differ due to varying daily conditions, such as emotions or stress on a given day [
60]. Second, while we reduce the number of samples to ensure a fair comparison between results, this reduction may impair the learning capabilities of the model. To isolate the effect of the first factor, we rerun the experiment without restricting the dataset size. The results are reported in
Table 6.
Using the entire LAUREATE dataset does not present scalability issues for identification tasks; however, this is not the case for re-identification scenarios as detailed in the
Supplementary Materials. Specifically, in the setup used by Seok et al. [
26], where all possible combinations of samples are generated as input for the Siamese Neural Network, increasing the number of samples in the dataset leads to an exponential increase in the number of pairs. While it might be feasible to manage this growth by exploring new solutions, such as subsampling pairs to reduce redundancy (e.g., by excluding positive pairs from the same user collected on the same day), doing so would significantly alter the model initially presented by Seok et al. [
26]. Therefore, we opt not to implement such changes and instead present only the results of the identification task in
Table 6. We believe this decision does not compromise the overall findings, as the re-identification task demonstrates limited effectiveness in handling a real-world dataset like LAUREATE as shown in
Table 4 and
Table 5.
Comparing
Table 5 and
Table 6 reveals a performance improvement in the models, highlighting the importance of having more samples, especially when users must be recognized under varying conditions. While the results improve slightly, with our feature-based method achieving an MCC of 0.16, the models still struggle with user recognition. We show empirically that a longitudinal approach is essential for testing biometric models. In real-world scenarios, training typically occurs during an initial time period, followed by testing on future data with only occasional updates. Therefore, randomly separating training and test sets does not accurately reflect real-world conditions. Given that many current methods struggle to identify users effectively in these situations, we advocate for the development of new strategies and models that can better address these challenges.
5. Guidelines for Evaluating Methods for Biometric Systems
To ensure that biometric systems are effective, scalable, and capable of generalizing to real-world conditions, it is essential to establish rigorous evaluation setups. In this section, we outline key recommendations for evaluating biometric methods. Although our findings focus specifically on methods using PPG data, we believe they generalize to other signals that exhibit similar characteristics, e.g., high variability due to factors such as emotional state or physical activity. Some of these guidelines are derived from the findings of this study, while others represent general best practices [
61].
The guidelines below focus on dataset diversity, robust evaluation setups, appropriate metrics, and reproducibility. Implementing these recommendations enables researchers and practitioners to produce more reliable and replicable results that generalize to real-world conditions.
By following these recommendations, future studies can establish more robust, realistic, and generalizable evaluation setups for biometric methods. To demonstrate the current relevance of these guidelines,
Table 7 presents whether each paper mentioned in the related work section adheres to the proposed guidelines. For the assessment of qualitative guidelines, we define adherence as follows: for the guideline “number of users”, adherence is met if the study includes more than 35 participants; for “demographic diversity”, adherence is met if the study includes participants from at least two distinct demographic groups (in this case, we assessed age and sex, defining age diversity as the presence of participants whose ages differ by at least 50 years; for “health condition diversity”, adherence is met if the study includes both healthy individuals and those with a diagnosed clinical condition; and for “longitudinal data”, adherence is met if data are collected on multiple days for all participants. Although “detailed documentation” is difficult to fully assess, we consider a paper compliant if its code is available and well-commented. If the code is unavailable, the paper should instead provide sufficient detail to enable the reconstruction of preprocessing steps and the implementation of the model. Finally, adherence to “balanced evaluation” is achieved if the test set is balanced, or if, in cases of imbalanced data, the MCC is reported. If no information is provided, adherence for that parameter is not met.
6. Limitations and Future Works
Our study aims to highlight how some of the previous works tend to report results that are not sustained when applied to more real scenarios. This inadvertent overestimation is caused by limitations in the used datasets and chosen evaluation procedures. In particular, we identify four main factors contributing to these overestimated results, and study their impact using both a feature-based method developed by us and two other methods from the literature. While our study does not center on methodological advancement, it nonetheless contributes empirically to ongoing discussions about evaluation standards in physiological biometrics. The insights provided here can guide the design of future, more advanced methods. Nevertheless, our work has several limitations, which are discussed here.
Firstly, the longitudinal dataset we use was collected using the Empatica E4, a medical-grade device that is CE-certified in Europe and has been successfully used in a wide range of studies and tasks [
63,
64]. However, most previous works, including the studies selected for our analysis, employed different sensors, making it unclear whether the observed performance degradation is solely due to the transition from controlled laboratory conditions to more variable real-world settings, or whether differences in the sensing devices also play a role. Future work could focus specifically on assessing the differences in signal quality and model performance when collecting data from the same users in both laboratory and real-world conditions.
Secondly, although our dataset includes multiple users with recordings over several days, all users come from the same university and belong to a similar age range. Future research could explore how increasing diversity in the dataset—across demographics and environments—might further impact biometric performances.
Additionally, among the four identified limitations, only one stems from algorithm design, while the remaining three arise from dataset constraints. Thus, although we evaluate these limitations in the context of biometrics, they may also affect other systems, such as human activity recognition.
7. Conclusions
In this study, we critically evaluated the reliability of biometric methods based on photoplethysmography (PPG) and proposed guidelines to enhance the robustness and generalizability of future research in this domain. By replicating prior work and developing a feature-based method as baseline, our findings reveal that, while existing methods show promise in controlled laboratory settings, they often face significant challenges in real-world applications. Different factors such as small sample sizes, the lack of longitudinal data, or testing protocols that do not account for unseen users, lead to an overestimation of model performance which may hinder the applicability of these systems in practical environments.
We propose a set of guidelines to address these challenges, emphasizing the need for larger and more diverse datasets that reflect the variability encountered in real-world use. This includes leveraging longitudinal data collected over multiple days and conditions to assess model stability over time. We also highlight the importance of demonstrating that the approaches remain valid and stable as the number of users increases, ensuring scalability. Additionally, we recommend implementing temporal splits for training and testing to prevent models from relying on short-term patterns that may not generalize well. Furthermore, testing model performance on unseen users is crucial, as it addresses potential security vulnerabilities, enhancing robustness against unauthorized access. We also recommend transparent reporting practices—such as detailed documentation of preprocessing steps, model parameters, and open access to code and datasets—essential for ensuring reproducibility and advancing the field. Finally, in scenarios with class imbalance, we advocate for metrics like the Matthews Correlation Coefficient, which provides a more reliable evaluation in imbalanced conditions compared to traditional metrics.
By demonstrating and quantifying well-known pitfalls within a reproducible framework, our proposed guidelines aim to create more robust evaluation setups, helping future research to develop biometric systems that are not only effective in controlled settings but also resilient and applicable in the variability of daily life. As the field progresses, we hope that these recommendations will contribute to the creation of biometric systems that are both reliable and adaptable, enabling their use across a wide range of conditions.
Supplementary Materials
The following supporting information can be downloaded at:
https://www.mdpi.com/article/10.3390/s25247586/s1, The Supplementary Materials provide additional implementation details of the reimplemented methods from the literature and of our feature-based approach. Figures S1 and S2 show the processing pipelines of the methods by Hinatsu et al. [
27] and Seok et al. [
26], respectively. Figures S3 and S4 compare intermediate processing steps between our reimplementation of Seok et al. [
26] and their original study. Figures S5 and S6 illustrate our feature-based identification and re-identification pipelines. References [
26,
27,
41,
65,
66,
67,
68,
69,
70,
71,
72,
73,
74,
75,
76] are cited in Supplementary Materials.
Author Contributions
Conceptualization, L.A. (Lidia Alecci), M.L. and S.S.; methodology, L.A. (Lidia Alecci), M.L., L.A. (Leonardo Alchieri), N.A. and S.S.; software, L.A. (Lidia Alecci); Writing—original draft, L.A. (Lidia Alecci), M.L.; Writing—review and editing, L.A. (Lidia Alecci), M.L., L.A. (Leonardo Alchieri), N.A. and S.S.; Supervision, S.S.; project administration, S.S.; funding acquisition, S.S. All authors have read and agreed to the published version of the manuscript.
Funding
This work is partially funded by the Swiss National Science Foundation (SNSF) through the grant 205121_197242 for the project “PROSELF: Semi-automated Self-Tracking Systems to Improve Personal Productivity”.
Institutional Review Board Statement
Not applicable.
Informed Consent Statement
Not applicable.
Data Availability Statement
All three datasets used in this study are publicly accessible to the research community. The BIDMC PPG and Respiration Dataset is available for download from PhysioNet (
https://physionet.org/content/bidmc/1.0.0/ (accessed on 9 December 2025)). Real-World PPG Dataset is available for download from Mendeley Data (
https://data.mendeley.com/datasets/yynb8t9x3d/1 (accessed on 9 December 2025)). Access to the LAUREATE dataset requires signing a data-sharing agreement. Interested researchers should contact the corresponding author of Laporte et al. [
25].
Acknowledgments
The authors used ChatGPT v5 to shorten selected paragraphs within this work. The shortened text was originally written without the help of any generative AI tools. Furthermore, the authors have reviewed and edited the output and take full responsibility for the content of this publication.
Conflicts of Interest
The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript; or in the decision to publish the results.
References
- Shilton, K. Four Billion Little Brothers? Privacy, Mobile Phones, and Ubiquitous Data Collection. Commun. ACM 2009, 52, 48–53. [Google Scholar] [CrossRef]
- Reiss, A.; Indlekofer, I.; Schmidt, P.; Van Laerhoven, K. Deep PPG: Large-Scale Heart Rate Estimation with Convolutional Neural Networks. Sensors 2019, 19, 3079. [Google Scholar] [CrossRef]
- Song, S.H.; Cho, J.S.; Oh, H.S.; Lee, J.S.; Kim, I.Y. Estimation of Blood Pressure Using Photoplethysmography on the Wrist. In Proceedings of the 2009 36th Annual Computers in Cardiology Conference (CinC), Park City, UT, USA, 13–16 September 2009; pp. 741–744. [Google Scholar]
- Paliakaitė, B.; Charlton, P.H.; Rapalis, A.; Pluščiauskaitė, V.; Piartli, P.; Kaniusas, E.; Marozas, V. Blood Pressure Estimation Based on Photoplethysmography: Finger Versus Wrist. In Proceedings of the 2021 Computing in Cardiology (CinC), Brno, Czech Republic, 13–15 September 2021; Volume 48, pp. 1–4. [Google Scholar] [CrossRef]
- Jain, A.K.; Flynn, P.; Ross, A.A. (Eds.) Handbook of Biometrics; Springer: Boston, MA, USA, 2008. [Google Scholar] [CrossRef]
- Cavazos, J.G.; Phillips, P.J.; Castillo, C.D.; O’Toole, A.J. Accuracy Comparison Across Face Recognition Algorithms: Where Are We on Measuring Race Bias? IEEE Trans. Biom. Behav. Identity Sci. 2021, 3, 101–111. [Google Scholar] [CrossRef]
- Ferrara, M.; Cappelli, R.; Maltoni, D. Detecting Double-Identity Fingerprint Attacks. IEEE Trans. Biom. Behav. Identity Sci. 2023, 5, 476–485. [Google Scholar] [CrossRef]
- Packhäuser, K.; Gündel, S.; Münster, N.; Syben, C.; Christlein, V.; Maier, A. Deep Learning-Based Patient Re-Identification Is Able to Exploit the Biometric Nature of Medical Chest X-ray Data. Sci. Rep. 2022, 12, 14851. [Google Scholar] [CrossRef] [PubMed]
- Wu, W.; Elliott, S.J.; Lin, S.; Sun, S.; Tang, Y. Review of Palm Vein Recognition. IET Biom. 2020, 9, 1–10. [Google Scholar] [CrossRef]
- Arca, S.; Campadelli, P.; Lanzarotti, R. A Face Recognition System Based on Automatically Determined Facial Fiducial Points. Pattern Recognit. 2006, 39, 432–443. [Google Scholar] [CrossRef]
- Apple Inc. SensorKit. 2024. Available online: https://developer.apple.com/documentation/sensorkit (accessed on 9 December 2025).
- Oura Team. Technology in the Oura Ring. 2024. Available online: https://ouraring.com/blog/ring-technology/ (accessed on 9 December 2025).
- Samsung Developer. Samsung Health Sensor SDK. 2024. Available online: https://developer.samsung.com/health/sensor/overview.html (accessed on 9 December 2025).
- Akl, T.J.; Wilson, M.A.; Ericson, M.N.; Coté, G.L. Quantifying Tissue Mechanical Properties Using Photoplethysmography. Biomed. Opt. Express 2014, 5, 2362. [Google Scholar] [CrossRef] [PubMed]
- Ajmal; Boonya-Ananta, T.; Rodriguez, A.J.; Du Le, V.N.; Ramella-Roman, J.C. Monte Carlo Analysis of Optical Heart Rate Sensors in Commercial Wearables: The Effect of Skin Tone and Obesity on the Photoplethysmography (PPG) Signal. Biomed. Opt. Express 2021, 12, 7445. [Google Scholar] [CrossRef]
- Jindal, V.; Birjandtalab, J.; Pouyan, M.B.; Nourani, M. An Adaptive Deep Learning Approach for PPG-based Identification. In Proceedings of the 2016 38th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), Orlando, FL, USA, 16–20 August 2016; pp. 6401–6404. [Google Scholar] [CrossRef]
- Everson, L.; Biswas, D.; Panwar, M.; Rodopoulos, D.; Acharyya, A.; Kim, C.H.; Van Hoof, C.; Konijnenburg, M.; Van Helleputte, N. BiometricNet: Deep Learning Based Biometric Identification Using Wrist-Worn PPG. In Proceedings of the 2018 IEEE International Symposium on Circuits and Systems (ISCAS), Florence, Italy, 27–30 May 2018; pp. 1–5. [Google Scholar] [CrossRef]
- Piciucco, E.; Di Lascio, E.; Maiorana, E.; Santini, S.; Campisi, P. Biometric Recognition Using Wearable Devices in Real-Life Settings. Pattern Recognit. Lett. 2021, 146, 260–266. [Google Scholar] [CrossRef]
- Biswas, D.; Everson, L.; Liu, M.; Panwar, M.; Verhoef, B.E.; Patki, S.; Kim, C.H.; Acharyya, A.; Van Hoof, C.; Konijnenburg, M.; et al. CorNET: Deep Learning Framework for PPG-Based Heart Rate Estimation and Biometric Identification in Ambulant Environment. IEEE Trans. Biomed. Circuits Syst. 2019, 13, 282–291. [Google Scholar] [CrossRef]
- Yadav, U.; Abbas, S.N.; Hatzinakos, D. Evaluation of PPG Biometrics for Authentication in Different States. In Proceedings of the 2018 International Conference on Biometrics (ICB), Gold Coast, Australia, 20–23 February 2018; pp. 277–282. [Google Scholar] [CrossRef]
- Donida Labati, R.; Piuri, V.; Rundo, F.; Scotti, F. Photoplethysmographic Biometrics: A Comprehensive Survey. Pattern Recognit. Lett. 2022, 156, 119–125. [Google Scholar] [CrossRef]
- Canali, S.; Schiaffonati, V.; Aliverti, A. Challenges and Recommendations for Wearable Devices in Digital Health: Data Quality, Interoperability, Health Equity, Fairness. PLoS Digit. Health 2022, 1, e0000104. [Google Scholar] [CrossRef]
- Zhang, Y.; Pratap, A.; Folarin, A.A.; Sun, S.; Cummins, N.; Matcham, F.; Vairavan, S.; Dineley, J.; RanJanuary, Y.; Rashid, Z.; et al. Long-Term Participant Retention and Engagement Patterns in an App and Wearable-Based Multinational Remote Digital Depression Study. npj Digit. Med. 2023, 6, 25. [Google Scholar] [CrossRef]
- Van Der Donckt, J.; Vandenbussche, N.; Van Der Donckt, J.; Chen, S.; Stojchevska, M.; De Brouwer, M.; Steenwinckel, B.; Paemeleire, K.; Ongenae, F.; Van Hoecke, S. Mitigating Data Quality Challenges in Ambulatory Wrist-Worn Wearable Monitoring through Analytical and Practical Approaches. Sci. Rep. 2024, 14, 17545. [Google Scholar] [CrossRef]
- Laporte, M.; Gjoreski, M.; Langheinrich, M. LAUREATE: A Dataset for Supporting Research in Affective Computing and Human Memory Augmentation. Proc. ACM Interact. Mob. Wearable Ubiquitous Technol. 2023, 7, 106. [Google Scholar] [CrossRef]
- Seok, C.L.; Song, Y.D.; An, B.S.; Lee, E.C. Photoplethysmogram Biometric Authentication Using a 1D Siamese Network. Sensors 2023, 23, 4634. [Google Scholar] [CrossRef] [PubMed]
- Hinatsu, S.; Suzuki, D.; Ishizuka, H.; Ikeda, S.; Oshiro, O. Photoplethysmographic Subject Identification by Considering Feature Values Derived from Heartbeat and Respiration. In Proceedings of the 2020 42nd Annual International Conference of the IEEE Engineering in Medicine & Biology Society (EMBC), Montreal, QC, Canada, 20–24 July 2020; pp. 902–905. [Google Scholar] [CrossRef]
- Bianco, S.; Napoletano, P. Biometric Recognition Using Multimodal Physiological Signals. IEEE Access 2019, 7, 83581–83588. [Google Scholar] [CrossRef]
- Cornelius, C.; Peterson, R.; Skinner, J.; Halter, R.; Kotz, D. A Wearable System That Knows Who Wears It. In Proceedings of the 12th Annual International Conference on Mobile Systems, Applications, and Services, Bretton Woods, NH, USA, 16–19 June 2014; pp. 55–67. [Google Scholar] [CrossRef]
- Almuashi, M.; Mohd Hashim, S.Z.; Mohamad, D.; Alkawaz, M.H.; Ali, A. Automated Kinship Verification and Identification through Human Facial Images: A Survey. Multimed. Tools Appl. 2017, 76, 265–307. [Google Scholar] [CrossRef]
- Peralta, D.; Galar, M.; Triguero, I.; Paternain, D.; García, S.; Barrenechea, E.; Benítez, J.M.; Bustince, H.; Herrera, F. A Survey on Fingerprint Minutiae-Based Local Matching for Verification and Identification: Taxonomy and Experimental Evaluation. Inf. Sci. 2015, 315, 67–87. [Google Scholar] [CrossRef]
- Huang, T.; Russell, S. Object Identification in a Bayesian Context. In Proceedings of the Fifteenth International Joint Conference on Artifical Intelligence, Nagoya, Japan, 23–29 August 1997; Morgan Kaufmann Publishers Inc.: San Francisco, CA, USA, 1997; Volume 2, pp. 1276–1282. [Google Scholar]
- Wu, Y.; He, G.D.; Wen, L.H.; Qin, X.; Yuan, C.A.; Gribova, V.; Filaretov, V.F.; Huang, D.S. Discriminative Local Representation Learning for Cross-Modality Visible-Thermal Person Re-Identification. IEEE Trans. Biom. Behav. Identity Sci. 2023, 5, 1–14. [Google Scholar] [CrossRef]
- Yaacoubi, C.; Besrour, R.; Lachiri, Z. A Multimodal Biometric Identification System Based on ECG and PPG Signals. In Proceedings of the 2nd International Conference on Digital Tools & Uses Congress, Virtual Event, Tunisia, 15–17 October 2020; pp. 1–6. [Google Scholar] [CrossRef]
- Liu, X.; Yuan, Z.; Ma, D. Deep Learning Framework for Biometric Identification from Wrist-Worn PPG with Acceleration Signals. In Proceedings of the 2021 IEEE 6th International Conference on Signal and Image Processing (ICSIP), Nanjing, China, 22–24 October 2021; pp. 1–5. [Google Scholar] [CrossRef]
- Reşit Kavsaoğlu, A.; Polat, K.; Recep Bozkurt, M. A Novel Feature Ranking Algorithm for Biometric Recognition with PPG Signals. Comput. Biol. Med. 2014, 49, 1–14. [Google Scholar] [CrossRef] [PubMed]
- Blasco, J.; Peris-Lopez, P. On the Feasibility of Low-Cost Wearable Sensors for Multi-Modal Biometric Verification. Sensors 2018, 18, 2782. [Google Scholar] [CrossRef] [PubMed]
- Luque, J.; Cortes, G.; Segura, C.; Maravilla, A.; Esteban, J.; Fabregat, J. END-to-END Photoplethysmography (PPG) Based Biometric Authentication by Using Convolutional Neural Networks. In Proceedings of the 2018 26th European Signal Processing Conference (EUSIPCO), Rome, Italy, 3–7 September 2018; pp. 538–542. [Google Scholar] [CrossRef]
- Alam, M.A.U. Person Re-identification Attack on Wearable Sensing. arXiv 2021, arXiv:2106.11900. [Google Scholar] [CrossRef]
- Pimentel, M.A.F.; Johnson, A.E.W.; Charlton, P.H.; Birrenkott, D.; Watkinson, P.J.; Tarassenko, L.; Clifton, D.A. Toward a Robust Estimation of Respiratory Rate From Pulse Oximeters. IEEE Trans. Biomed. Eng. 2017, 64, 1914–1923. [Google Scholar] [CrossRef]
- Siam, A. Real-World PPG Dataset, Version 1; Mendeley Data: Amsterdam, The Netherlands, 2019. [Google Scholar] [CrossRef]
- Siam, A.I. (Department of Embedded Network Systems Technology, Faculty of Artificial Intelligence, Kafrelsheikh University, Kafr El Sheikh, Egypt). Private communication via emails, 2023.
- Siam, A.I. (Department of Embedded Network Systems Technology, Faculty of Artificial Intelligence, Kafrelsheikh University, Kafr El Sheikh, Egypt). Private communication via emails, 2024.
- Das, A.; Borisov, N.; Caesar, M. Exploring Ways To Mitigate Sensor-Based Smartphone Fingerprinting. arXiv 2015, arXiv:1503.01874. [Google Scholar] [CrossRef]
- Stragapede, G.; Vera-Rodriguez, R.; Tolosana, R.; Morales, A. BehavePassDB: Public Database for Mobile Behavioral Biometrics and Benchmark Evaluation. Pattern Recognit. 2023, 134, 109089. [Google Scholar] [CrossRef]
- Dalianis, H. Evaluation Metrics and Evaluation. In Clinical Text Mining: Secondary Use of Electronic Patient Records; Dalianis, H., Ed.; Springer International Publishing: Cham, Switzerland, 2018; pp. 45–53. [Google Scholar] [CrossRef]
- Chicco, D.; Jurman, G. The Advantages of the Matthews Correlation Coefficient (MCC) over F1 Score and Accuracy in Binary Classification Evaluation. BMC Genom. 2020, 21, 6. [Google Scholar] [CrossRef]
- Matthews, B.W. Comparison of the Predicted and Observed Secondary Structure of T4 Phage Lysozyme. Biochim. Biophys. Acta (Bba)-Protein Struct. 1975, 405, 442–451. [Google Scholar] [CrossRef]
- Wilcoxon, F. Individual Comparisons by Ranking Methods. Biom. Bull. 1945, 1, 80–83. [Google Scholar] [CrossRef]
- Dunn, O.J. Multiple Comparisons Among Means. J. Am. Stat. Assoc. 1961, 56, 52–64. [Google Scholar] [CrossRef]
- Xu, Y.; Hübener, I.; Seipp, A.K.; Ohly, S.; David, K. From the Lab to the Real-World: An Investigation on the Influence of Human Movement on Emotion Recognition Using Physiological Signals. In Proceedings of the 2017 IEEE International Conference on Pervasive Computing and Communications Workshops (PerCom Workshops), Kona, HI, USA, 13–17 March 2017; pp. 345–350. [Google Scholar] [CrossRef]
- Johnston, D.W.; Anastasiades, P.; Wood, C. The Relationship Between Cardiovascular Responses in the Laboratory and in the Field. Psychophysiology 1990, 27, 34–44. [Google Scholar] [CrossRef] [PubMed]
- van Geert, P.; van Dijk, M. Focus on Variability: New Tools to Study Intra-Individual Variability in Developmental Data. Infant Behav. Dev. 2002, 25, 340–374. [Google Scholar] [CrossRef]
- Guidetti, L.; Rivellini, G.; Figura, F. EMG Patterns during Running: Intra- and Inter-Individual Variability. J. Electromyogr. Kinesiol. 1996, 6, 37–48. [Google Scholar] [CrossRef]
- Lara, O.D.; Labrador, M.A. A Survey on Human Activity Recognition Using Wearable Sensors. IEEE Commun. Surv. Tutor. 2013, 15, 1192–1209. [Google Scholar] [CrossRef]
- Sugrim, S.; Liu, C.; Lindqvist, J. Recruit Until It Fails: Exploring Performance Limits for Identification Systems. Proc. ACM Interact. Mob. Wearable Ubiquitous Technol. 2019, 3, 104:1–104:26. [Google Scholar] [CrossRef]
- Geng, C.; Huang, S.J.; Chen, S. Recent Advances in Open Set Recognition: A Survey. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 43, 3614–3631. [Google Scholar] [CrossRef]
- Gu, W.; Yang, Z.; Shangguan, L.; Sun, W.; Jin, K.; Liu, Y. Intelligent Sleep Stage Mining Service with Smartphones. In Proceedings of the 2014 ACM International Joint Conference on Pervasive and Ubiquitous Computing, Seattle, WA, USA, 13–17 September 2014; Association for Computing Machinery: New York, NY, USA, 2014; pp. 649–660. [Google Scholar] [CrossRef]
- Hammerla, N.Y.; Plötz, T. Let’s (Not) Stick Together: Pairwise Similarity Biases Cross-Validation in Activity Recognition. In Proceedings of the 2015 ACM International Joint Conference on Pervasive and Ubiquitous Computing, Osaka, Japan, 7–11 September 2015; Association for Computing Machinery: New York, NY, USA, 2015; pp. 1041–1051. [Google Scholar] [CrossRef]
- Alecci, L.; Zanzi, A.; Laporte, M.; Alchieri, L.; Abdalazim, N.; Felice, G.D.; Santini, S. Robust PPG Authentication: The Role of Temporal, Individual, and State Variability. In Proceedings of the Companion of the 2025 ACM International Joint Conference on Pervasive and Ubiquitous Computing (UbiComp Companion ’25), Espoo, Finland, 12–16 October 2025; ACM: New York, NY, USA, 2025; p. 8p. [Google Scholar] [CrossRef]
- Ioannidis, J.P.A.; Fanelli, D.; Dunne, D.D.; Goodman, S.N. Meta-Research: Evaluation and Improvement of Research Methods and Practices. PLoS Biol. 2015, 13, e1002264. [Google Scholar] [CrossRef]
- Gill, S.K.; Barsky, A.; Guan, X.; Bunting, K.V.; Karwath, A.; Tica, O.; Stanbury, M.; Haynes, S.; Folarin, A.; Dobson, R.; et al. Consumer Wearable Devices for Evaluation of Heart Rate Control Using Digoxin versus Beta-Blockers: The RATE-AF Randomized Trial. Nat. Med. 2024, 30, 2030–2036. [Google Scholar] [CrossRef] [PubMed]
- Hosseini, S.; Gottumukkala, R.; Katragadda, S.; Bhupatiraju, R.T.; Ashkar, Z.; Borst, C.W.; Cochran, K. A Multimodal Sensor Dataset for Continuous Stress Detection of Nurses in a Hospital. Sci. Data 2022, 9, 255. [Google Scholar] [CrossRef]
- Ragot, M.; Martin, N.; Em, S.; Pallamin, N.; Diverrez, J.M. Emotion Recognition Using Physiological Signals: Laboratory vs. Wearable Sensors. In Advances in Human Factors in Wearable Technologies and Game Design, Proceedings of the AHFE 2017 International Conference on Advances in Human Factors and Wearable Technologies, Los Angeles, CA, USA, 17–21 July 2017; Ahram, T., Falcão, C., Eds.; Springer: Cham, Switzerland, 2018; pp. 15–22. [Google Scholar] [CrossRef]
- Temko, A. Accurate Heart Rate Monitoring During Physical Exercises Using PPG. IEEE Trans. Biomed. Eng. 2017, 64, 2016–2024. [Google Scholar] [CrossRef]
- Lapitan, D.G.; Rogatkin, D.A.; Molchanova, E.A.; Tarasov, A.P. Estimation of Phase Distortions of the Photoplethysmographic Signal in Digital IIR Filtering. Sci. Rep. 2024, 14, 6546. [Google Scholar] [CrossRef]
- Virtanen, P.; Gommers, R.; Oliphant, T.E.; Haberland, M.; Reddy, T.; Cournapeau, D.; Burovski, E.; Peterson, P.; Weckesser, W.; Bright, J.; et al. SciPy 1.0: Fundamental Algorithms for Scientific Computing in Python. Nat. Methods 2020, 17, 261–272. [Google Scholar] [CrossRef]
- Shannon, C.E. A Mathematical Theory of Communication. Bell Syst. Tech. J. 1948, 27, 379–423. [Google Scholar] [CrossRef]
- Brigham, E.O. The Fast Fourier Transform and Its Applications; Prentice-Hall Signal Processing Series; Prentice Hall: Englewood Cliffs, NJ, USA, 1988. [Google Scholar]
- Welch, P.D. The Use of Fast Fourier Transform for the Estimation of Power Spectra: A Method Based on Time Averaging over Short, Modified Periodograms. IEEE Trans. Audio Electroacoust. 1967, 15, 70–73. [Google Scholar] [CrossRef]
- Daubechies, I. Ten Lectures on Wavelets; Number 61 in CBMS-NSF Regional Conference Series in Applied Mathematics; Society for Industrial and Applied Mathematics: Philadelphia, PA, USA, 1992. [Google Scholar] [CrossRef]
- Daubechies, I. Orthonormal Bases of Compactly Supported Wavelets. Commun. Pure Appl. Math. 1988, 41, 909–996. [Google Scholar] [CrossRef]
- Chen, T.; Guestrin, C. XGBoost: A Scalable Tree Boosting System. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, New York, NY, USA, 13–17 August 2016; pp. 785–794. [Google Scholar] [CrossRef]
- Ramdani, F.; Furqon, M.T. The Simplicity of XGBoost Algorithm versus the Complexity of Random Forest, Support Vector Machine, and Neural Networks Algorithms in Urban Forest Classification. F1000Research 2022, 11, 1069. [Google Scholar] [CrossRef]
- Didavi, A.B.K.; Agbokpanzo, R.G.; Agbomahena, M. Comparative Study of Decision Tree, Random Forest and XGBoost Performance in Forecasting the Power Output of a Photovoltaic System. In Proceedings of the 2021 4th International Conference on Bio-Engineering for Smart Technologies (BioSMART), Paris, France, 8–10 December 2021; pp. 1–5. [Google Scholar] [CrossRef]
- Li, Y.; Chen, C.L.P.; Zhang, T. A Survey on Siamese Network: Methodologies, Applications, and Opportunities. IEEE Trans. Artif. Intell. 2022, 3, 994–1014. [Google Scholar] [CrossRef]
| Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).