Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

Article Types

Countries / Regions

Search Results (54)

Search Parameters:
Keywords = remote heart rate estimation

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
15 pages, 3599 KB  
Article
High-Fidelity rPPG Waveform Reconstruction from Palm Videos Using GANs
by Tao Li and Yuliang Liu
Sensors 2026, 26(2), 563; https://doi.org/10.3390/s26020563 - 14 Jan 2026
Viewed by 276
Abstract
Remote photoplethysmography (rPPG) enables non-contact acquisition of human physiological parameters using ordinary cameras, and has been widely applied in medical monitoring, human–computer interaction, and health management. However, most existing studies focus on estimating specific physiological metrics, such as heart rate and heart rate [...] Read more.
Remote photoplethysmography (rPPG) enables non-contact acquisition of human physiological parameters using ordinary cameras, and has been widely applied in medical monitoring, human–computer interaction, and health management. However, most existing studies focus on estimating specific physiological metrics, such as heart rate and heart rate variability, while paying insufficient attention to reconstructing the underlying rPPG waveform. In addition, publicly available datasets typically record facial videos accompanied by fingertip PPG signals as reference labels. Since fingertip PPG waveforms differ substantially from the true photoplethysmography (PPG) signals obtained from the face, deep learning models trained on such datasets often struggle to recover high-quality rPPG waveforms. To address this issue, we collected a new dataset consisting of palm-region videos paired with wrist-based PPG signals as reference labels, and experimentally validated its effectiveness for training neural network models aimed at rPPG waveform reconstruction. Furthermore, we propose a generative adversarial network (GAN)-based pulse-wave synthesis framework that produces high-quality rPPG waveforms by denoising the mean green-channel signal. By incorporating time-domain peak-aware loss, frequency-domain loss, and adversarial loss, our method achieves promising performance, with an RMSE (Root Mean Square Error) of 0.102, an MAPE (Mean Absolute Percentage Error) of 0.028, a Pearson correlation of 0.987, and a cosine similarity of 0.989. These results demonstrate the capability of the proposed approach to reconstruct high-fidelity rPPG waveforms with improved morphological accuracy compared to noisy raw rPPG signals, rather than directly validating health monitoring performance. This study presents a high-quality rPPG waveform reconstruction approach from both data and model perspectives, providing a reliable foundation for subsequent physiological signal analysis, waveform-based studies, and potential health-related applications. Full article
(This article belongs to the Special Issue Systems for Contactless Monitoring of Vital Signs)
Show Figures

Figure 1

17 pages, 291 KB  
Article
A Unified Benchmarking Framework for Classical Machine Learning Based Heart Rate Estimation from RGB and NIR rPPG
by Sahar Qaadan, Ghassan Al Jayyousi and Adam Alkhalaileh
Electronics 2026, 15(1), 218; https://doi.org/10.3390/electronics15010218 - 2 Jan 2026
Viewed by 323
Abstract
This work presents a unified benchmarking framework for evaluating classical machine-learning–based heart-rate estimation from remote photoplethysmography (rPPG) across both RGB and near-infrared (NIR) modalities. Despite extensive research on algorithmic rPPG methods, their relative robustness across datasets, illumination conditions, and sensor types remains inconsistently [...] Read more.
This work presents a unified benchmarking framework for evaluating classical machine-learning–based heart-rate estimation from remote photoplethysmography (rPPG) across both RGB and near-infrared (NIR) modalities. Despite extensive research on algorithmic rPPG methods, their relative robustness across datasets, illumination conditions, and sensor types remains inconsistently reported. To address this gap, we standardize ROI extraction, signal preprocessing, rPPG computation, handcrafted feature generation, and label formation across four publicly available datasets: UBFC-rPPG Part 1, UBFC-rPPG Part 2, VicarPPG-2, and IMVIA-NIR. We benchmark five rPPG extraction methods (Green, POS, CHROM, PBV, PCA/ICA) combined with four classical regressors using MAE, RMSE, and R2, complemented by permutation feature importance for interpretability. Results show that CHROM is consistently the most reliable algorithm across all RGB datasets, providing the lowest error and highest stability, particularly when paired with tree-based models. For NIR recordings, PCA with spatial patch decomposition substantially outperforms ICA, highlighting the importance of spatial redundancy when color cues are absent. While handcrafted features and classical regressors offer interpretable baselines, their generalization is limited by small-sample datasets and the absence of temporal modeling. The proposed pipeline establishes robust cross-dataset baselines and offers a standardized foundation for future deep-learning architectures, hybrid algorithmic–learned models, and multimodal sensor-fusion approaches in remote physiological monitoring. Full article
(This article belongs to the Special Issue Image Processing and Analysis)
22 pages, 2586 KB  
Article
Noncontact Visualization of Respiration and Vital Sign Monitoring Using a Single Mid-Wave Infrared Thermal Camera: Preliminary Proof-of-Concept
by Takashi Suzuki
Sensors 2026, 26(1), 98; https://doi.org/10.3390/s26010098 - 23 Dec 2025
Viewed by 492
Abstract
Infrared thermal cameras can noninvasively measure the surface temperatures of objects and are widely used as fever-screening systems for infectious diseases. However, body temperature measurements alone are often insufficient for identifying people with infections. To address the inherent limitations of fever-based screening, this [...] Read more.
Infrared thermal cameras can noninvasively measure the surface temperatures of objects and are widely used as fever-screening systems for infectious diseases. However, body temperature measurements alone are often insufficient for identifying people with infections. To address the inherent limitations of fever-based screening, this study aimed to develop analytical methods that enable multi-vital sensing alongside body temperature measurement using a single mid-wave infrared (MWIR) camera. Respiratory parameters were assessed by visualizing exhaled airflow based on MWIR absorption by carbon dioxide, whereas the heart rate was estimated from subtle temperature fluctuations captured using high thermal resolution. The experimental results validated the proposed method, showing that the developed system achieved good agreement with reference measurements; the respiratory rate, heart rate, and body temperature showed strong correlations (r = 0.864–0.987) and acceptable limits of agreement in Bland–Altman analyses. The exhalation volume was quantified from the visualized airflow and was found to align with the expected physiological ranges. These results demonstrate that noncontact multi-vital sensing can be achieved using a single MWIR camera, without the need for complex instrumentation. The proposed method holds promise for high-precision infection screening, remote health monitoring, and in-home physiological assessment. Full article
(This article belongs to the Collection Biomedical Imaging and Sensing)
Show Figures

Figure 1

22 pages, 2396 KB  
Article
CHROM-Y: Illumination-Adaptive Robust Remote Photoplethysmography Through 2D Chrominance–Luminance Fusion and Convolutional Neural Networks
by Mohammed Javidh, Ruchi Shah, Mohan Uma, Sethuramalingam Prabhu and Rajendran Beaulah Jeyavathana
Signals 2025, 6(4), 72; https://doi.org/10.3390/signals6040072 - 9 Dec 2025
Viewed by 759
Abstract
Remote photoplethysmography (rPPG) enables non-contact heart rate estimation but remains highly sensitive to illumination variation and dataset-dependent factors. This study proposes CHROM-Y, a robust 2D feature representation that combines chrominance (Ω, Φ) with luminance (Y) to improve physiological signal extraction under varying lighting [...] Read more.
Remote photoplethysmography (rPPG) enables non-contact heart rate estimation but remains highly sensitive to illumination variation and dataset-dependent factors. This study proposes CHROM-Y, a robust 2D feature representation that combines chrominance (Ω, Φ) with luminance (Y) to improve physiological signal extraction under varying lighting conditions. The proposed features were evaluated using U-Net, ResNet-18, and VGG16 for heart rate estimation and waveform reconstruction on the UBFC-rPPG and BhRPPG datasets. On UBFC-rPPG, U-Net with CHROM-Y achieved the best performance with a Peak MAE of 3.62 bpm and RMSE of 6.67 bpm, while ablation experiments confirmed the importance of the Y-channel, showing degradation of up to 41.14% in MAE when removed. Although waveform reconstruction demonstrated low Pearson correlation, dominant frequency preservation enabled reliable frequency-based HR estimation. Cross-dataset evaluation revealed reduced generalization (MAE up to 13.33 bpm and RMSE up to 22.80 bpm), highlighting sensitivity to domain shifts. However, fine-tuning U-Net on BhRPPG produced consistent improvements across low, medium, and high illumination levels, with performance gains of 11.18–29.47% in MAE and 12.48–27.94% in RMSE, indicating improved adaptability to illumination variations. Full article
Show Figures

Figure 1

18 pages, 4262 KB  
Article
A Dual-Branch Spatio-Temporal Feature Differencing Method for Robust rPPG Estimation
by Gyumin Cho, Man-Je Kim and Chang Wook Ahn
Mathematics 2025, 13(23), 3830; https://doi.org/10.3390/math13233830 - 29 Nov 2025
Viewed by 372
Abstract
Remote photoplethysmography (rPPG) is a non-contact technology that estimates physiological signals, such as Heart Rate (HR), by capturing subtle skin color changes caused by periodic blood volume variations using only a standard RGB camera. While cost-effective and convenient, it suffers from a fundamental [...] Read more.
Remote photoplethysmography (rPPG) is a non-contact technology that estimates physiological signals, such as Heart Rate (HR), by capturing subtle skin color changes caused by periodic blood volume variations using only a standard RGB camera. While cost-effective and convenient, it suffers from a fundamental limitation: performance degrades severely in dynamic environments due to susceptibility to noise, such as abrupt illumination changes or motion blur. This study presents a deep learning framework that combines two structural modifications to ensure robustness in dynamic environments, specifically modeling movement noise and illumination change noise. The proposed framework structurally cancels global disturbances, such as illumination changes or global motion, through a dual-branch pipeline that encodes the face and background in parallel after Video Color Magnification (VCM) and then performs differencing. Subsequently, it utilizes a structure that injects a Temporal Shift Module (TSM) into the Spatio-Temporal Feature Extraction (SSFE) block to preserve long- and short-term temporal correlations and smooth noise, even amidst short and irregular movements. We measured MAE, RMSE, and correlation on the standard dataset UBFC-rPPG under four noise conditions: clean, illumination change noise, Movement Noise, Both Noise and the real-world in-vehicle dataset MR-NIRP (Stationary and Driving). Experimental results showed that the proposed method achieved consistent error reduction and correlation improvement compared to the VS-Net baseline in the illumination change noise-only and combined noise environments (UBFC-rPPG) and in the high-noise driving scenario (MR-NIRP). It maintained competitive performance in motion-only noise. Conversely, a modest performance disadvantage was observed under clean conditions (UBFC) and quasi-clean stationary conditions (MR-NIRP), interpreted as a design trade-off focused on global noise cancellation and temporal smoothing. Ablation studies demonstrated that the dual-branch pipeline is the primary contributor under illumination change noise, while TSM is the key contributor under movement noise, and that the combination of both elements achieves optimal robustness in the most complex scenarios. Full article
Show Figures

Figure 1

31 pages, 1305 KB  
Review
Artificial Intelligence in Cardiac Electrophysiology: A Clinically Oriented Review with Engineering Primers
by Giovanni Canino, Assunta Di Costanzo, Nadia Salerno, Isabella Leo, Mario Cannataro, Pietro Hiram Guzzi, Pierangelo Veltri, Sabato Sorrentino, Salvatore De Rosa and Daniele Torella
Bioengineering 2025, 12(10), 1102; https://doi.org/10.3390/bioengineering12101102 - 13 Oct 2025
Cited by 2 | Viewed by 5082
Abstract
Artificial intelligence (AI) is transforming cardiac electrophysiology across the entire care pathway, from arrhythmia detection on 12-lead electrocardiograms (ECGs) and wearables to the guidance of catheter ablation procedures, through to outcome prediction and therapeutic personalization. End-to-end deep learning (DL) models have achieved cardiologist-level [...] Read more.
Artificial intelligence (AI) is transforming cardiac electrophysiology across the entire care pathway, from arrhythmia detection on 12-lead electrocardiograms (ECGs) and wearables to the guidance of catheter ablation procedures, through to outcome prediction and therapeutic personalization. End-to-end deep learning (DL) models have achieved cardiologist-level performance in rhythm classification and prognostic estimation on standard ECGs, with a reported arrhythmia classification accuracy of ≥95% and an atrial fibrillation detection sensitivity/specificity of ≥96%. The application of AI to wearable devices enables population-scale screening and digital triage pathways. In the electrophysiology (EP) laboratory, AI standardizes the interpretation of intracardiac electrograms (EGMs) and supports target selection, and machine learning (ML)-guided strategies have improved ablation outcomes. In patients with cardiac implantable electronic devices (CIEDs), remote monitoring feeds multiparametric models capable of anticipating heart-failure decompensation and arrhythmic risk. This review outlines the principal modeling paradigms of supervised learning (regression models, support vector machines, neural networks, and random forests) and unsupervised learning (clustering, dimensionality reduction, association rule learning) and examines emerging technologies in electrophysiology (digital twins, physics-informed neural networks, DL for imaging, graph neural networks, and on-device AI). However, major challenges remain for clinical translation, including an external validation rate below 30% and workflow integration below 20%, which represent core obstacles to real-world adoption. A joint clinical engineering roadmap is essential to translate prototypes into reliable, bedside tools. Full article
(This article belongs to the Special Issue Mathematical Models for Medical Diagnosis and Testing)
Show Figures

Figure 1

18 pages, 8336 KB  
Article
Contactless Estimation of Heart Rate and Arm Tremor from Real Competition Footage of Elite Archers
by Byeong Seon An, Song Hee Park, Ji Yeon Moon and Eui Chul Lee
Electronics 2025, 14(18), 3650; https://doi.org/10.3390/electronics14183650 - 15 Sep 2025
Viewed by 1211
Abstract
This study investigates the effects of heart rate and arm tremor on performance in elite archery, using non-contact physiological monitoring from real Olympic competition footage. A total of 50 video segments were extracted from publicly available international broadcasts, comprising athletes of various backgrounds. [...] Read more.
This study investigates the effects of heart rate and arm tremor on performance in elite archery, using non-contact physiological monitoring from real Olympic competition footage. A total of 50 video segments were extracted from publicly available international broadcasts, comprising athletes of various backgrounds. From these, heart rate signals were estimated via remote photoplethysmography (rPPG) from facial regions, and micro-movements were quantified from right and left arm regions using feature point tracking. Ordinal logistic regression was employed to evaluate the relationship between biometric variables and archery scores (10, 9, ≤8 points). Results showed that elevated heart rate (β = −0.1166; p< 0.001) and greater right-arm movement (β = −6.1747; p = 0.008) were significantly associated with lower scores. Athletes scoring 10 points exhibited significantly lower heart rate (p< 0.001) and reduced right-arm tremor (p = 0.010) compared to others. These findings support the hypothesis that physiological arousal and biomechanical instability impair performance, and they further demonstrate the feasibility of contactless monitoring in real competition environments. The proposed method enables objective, in-game performance evaluation and supports the development of personalized training systems for precision sports. Full article
(This article belongs to the Special Issue Artificial Intelligence, Computer Vision and 3D Display)
Show Figures

Figure 1

24 pages, 624 KB  
Review
Integrating Artificial Intelligence into Perinatal Care Pathways: A Scoping Review of Reviews of Applications, Outcomes, and Equity
by Rabie Adel El Arab, Omayma Abdulaziz Al Moosa, Zahraa Albahrani, Israa Alkhalil, Joel Somerville and Fuad Abuadas
Nurs. Rep. 2025, 15(8), 281; https://doi.org/10.3390/nursrep15080281 - 31 Jul 2025
Cited by 2 | Viewed by 4409
Abstract
Background: Artificial intelligence (AI) and machine learning (ML) have been reshaping maternal, fetal, neonatal, and reproductive healthcare by enhancing risk prediction, diagnostic accuracy, and operational efficiency across the perinatal continuum. However, no comprehensive synthesis has yet been published. Objective: To conduct a scoping [...] Read more.
Background: Artificial intelligence (AI) and machine learning (ML) have been reshaping maternal, fetal, neonatal, and reproductive healthcare by enhancing risk prediction, diagnostic accuracy, and operational efficiency across the perinatal continuum. However, no comprehensive synthesis has yet been published. Objective: To conduct a scoping review of reviews of AI/ML applications spanning reproductive, prenatal, postpartum, neonatal, and early child-development care. Methods: We searched PubMed, Embase, the Cochrane Library, Web of Science, and Scopus through April 2025. Two reviewers independently screened records, extracted data, and assessed methodological quality using AMSTAR 2 for systematic reviews, ROBIS for bias assessment, SANRA for narrative reviews, and JBI guidance for scoping reviews. Results: Thirty-nine reviews met our inclusion criteria. In preconception and fertility treatment, convolutional neural network-based platforms can identify viable embryos and key sperm parameters with over 90 percent accuracy, and machine-learning models can personalize follicle-stimulating hormone regimens to boost mature oocyte yield while reducing overall medication use. Digital sexual-health chatbots have enhanced patient education, pre-exposure prophylaxis adherence, and safer sexual behaviors, although data-privacy safeguards and bias mitigation remain priorities. During pregnancy, advanced deep-learning models can segment fetal anatomy on ultrasound images with more than 90 percent overlap compared to expert annotations and can detect anomalies with sensitivity exceeding 93 percent. Predictive biometric tools can estimate gestational age within one week with accuracy and fetal weight within approximately 190 g. In the postpartum period, AI-driven decision-support systems and conversational agents can facilitate early screening for depression and can guide follow-up care. Wearable sensors enable remote monitoring of maternal blood pressure and heart rate to support timely clinical intervention. Within neonatal care, the Heart Rate Observation (HeRO) system has reduced mortality among very low-birth-weight infants by roughly 20 percent, and additional AI models can predict neonatal sepsis, retinopathy of prematurity, and necrotizing enterocolitis with area-under-the-curve values above 0.80. From an operational standpoint, automated ultrasound workflows deliver biometric measurements at about 14 milliseconds per frame, and dynamic scheduling in IVF laboratories lowers staff workload and per-cycle costs. Home-monitoring platforms for pregnant women are associated with 7–11 percent reductions in maternal mortality and preeclampsia incidence. Despite these advances, most evidence derives from retrospective, single-center studies with limited external validation. Low-resource settings, especially in Sub-Saharan Africa, remain under-represented, and few AI solutions are fully embedded in electronic health records. Conclusions: AI holds transformative promise for perinatal care but will require prospective multicenter validation, equity-centered design, robust governance, transparent fairness audits, and seamless electronic health record integration to translate these innovations into routine practice and improve maternal and neonatal outcomes. Full article
Show Figures

Figure 1

22 pages, 3866 KB  
Article
Evaluating the Accuracy of Low-Cost Wearable Sensors for Healthcare Monitoring
by Tatiana Pereira Filgueiras, Pedro Bertemes-Filho and Fabrício Noveletto
Micromachines 2025, 16(7), 791; https://doi.org/10.3390/mi16070791 - 2 Jul 2025
Cited by 1 | Viewed by 3561
Abstract
This study evaluates the accuracy of a low-cost wearable system for the continuous monitoring of vital signs, including heart rate, blood oxygen saturation (SpO2), blood pressure trend (BPT), and body temperature. The prototype was built using the nRF52840 microcontroller, which [...] Read more.
This study evaluates the accuracy of a low-cost wearable system for the continuous monitoring of vital signs, including heart rate, blood oxygen saturation (SpO2), blood pressure trend (BPT), and body temperature. The prototype was built using the nRF52840 microcontroller, which integrates photoplethysmography and infrared sensors. The heart rate and SpO2 data were collected under three body positions (Rest, Sitting, and Standing), while all measurements were performed using both anatomical configurations: BPT-Finger and BPT-Earlobe. Results were compared against validated commercial devices: UT-100 for heart rate and SpO2, G-TECH LA800 for blood pressure, and G-TECH THGTSC3 for body temperature. Ten participants were monitored over a ten-day period. Bland–Altman analysis revealed clinically acceptable agreement thresholds of ±5 mmHg for blood pressure, ±5–10 bpm for heart rate, ±4% for SpO2, and ±0.5 °C for temperature. Both wearable configurations demonstrated clinically acceptable agreement across all vital signs. The BPT-Earlobe configuration exhibited superior stability and lower variability in the Rest and Sitting positions, likely due to reduced motion artifacts. Conversely, the BPT-Finger configuration showed higher SpO2 accuracy in the Standing position, with narrower limits of agreement. These findings highlight the importance of sensor placement in maintaining measurement consistency across physiological conditions. With an estimated cost of only ~USD 130—compared to ~USD 590 for the commercial alternatives—the proposed system presents a cost-effective, scalable, and accessible solution for decentralized health monitoring, particularly in underserved or remote environments. Full article
(This article belongs to the Special Issue Advanced Flexible Electronic Devices for Biomedical Application)
Show Figures

Figure 1

22 pages, 6248 KB  
Article
Situational Awareness Prediction for Remote Tower Controllers Based on Eye-Tracking and Heart Rate Variability Data
by Weijun Pan, Ruihan Liang, Yuhao Wang, Dajiang Song and Zirui Yin
Sensors 2025, 25(7), 2052; https://doi.org/10.3390/s25072052 - 25 Mar 2025
Cited by 2 | Viewed by 1402
Abstract
Remote tower technology is an important development direction for air traffic control to reduce the construction and operation costs of small or remote airports. However, its digital and virtualized working environment poses new challenges to controllers’ situational awareness (SA). In this study, a [...] Read more.
Remote tower technology is an important development direction for air traffic control to reduce the construction and operation costs of small or remote airports. However, its digital and virtualized working environment poses new challenges to controllers’ situational awareness (SA). In this study, a dataset is constructed by collecting eye-tracking (ET) and heart rate variability (HRV) data from participants in a remote tower simulation control experiment. At the same time, probe questions are designed that correspond to the SA hierarchy in conjunction with the remote tower control task flow, and the dataset is annotated using the scenario presentation assessment method (SPAM). The annotated dataset containing 25 ET and HRV features is trained using the LightGBM model optimized by a Tree-structured Parzen Estimator, and feature selection and model interpretation are performed using the SHapley Additive exPlanations (SHAP) analysis. The results show that the TPE-LightGBM model exhibits excellent prediction capability, obtaining an RMSE, MAE and adjusted R2 of 0.0909, 0.0730 and 0.7845, respectively. This study presents an effective method for assessing and predicting controllers’ SA in remote tower environments. It further provides a theoretical basis for understanding the effect of the physiological state of remote tower controllers on their SA. Full article
(This article belongs to the Section Biosensors)
Show Figures

Figure 1

15 pages, 2854 KB  
Article
Designing a Remote Photoplethysmography-Based Heart Rate Estimation Algorithm During a Treadmill Exercise
by Yusang Nam, Junghwan Lee, Jihong Lee, Hyuntae Lee, Dongwook Kwon, Minsoo Yeo, Sayup Kim, Ryanghee Sohn and Cheolsoo Park
Electronics 2025, 14(5), 890; https://doi.org/10.3390/electronics14050890 - 24 Feb 2025
Cited by 1 | Viewed by 3462
Abstract
Remote photoplethysmography is a technology that estimates heart rate by detecting changes in blood volume induced by heartbeats and the resulting changes in skin color through imaging. This technique is fundamental for the noncontact acquisition of physiological signals from the human body. Despite [...] Read more.
Remote photoplethysmography is a technology that estimates heart rate by detecting changes in blood volume induced by heartbeats and the resulting changes in skin color through imaging. This technique is fundamental for the noncontact acquisition of physiological signals from the human body. Despite the notable progress in remote-photoplethysmography algorithms for estimating heart rate from facial videos, challenges remain in accurately assessing heart rate during cardiovascular exercises such as treadmill or elliptical workouts. To address these issues, research has been conducted in various fields. For example, an understanding of optics can help solve these issues. Careful design of video production is also crucial. Approaches in computer vision and deep learning with neural networks can also be applied. We focused on developing a practical approach to improve heart rate estimation algorithms under constrained conditions. To address the limitations of motion blur during high-motion activities, we introduced a novel motion-based algorithm. While existing methods like CHROM, LGI, OMIT, and POS incorporate correction processes, they have shown limited success in environments with significant motion. By analyzing treadmill data, we identified a relationship between motion changes and heart rate. With an initial heart rate provided, our algorithm achieved over a 15 bpm improvement in mean absolute error and root mean squared error compared to existing methods, along with more than double the Pearson correlation. We hope this research contributes to advancements in healthcare and monitoring. Full article
Show Figures

Figure 1

17 pages, 3294 KB  
Article
Hybrid Neural Network Models to Estimate Vital Signs from Facial Videos
by Yufeng Zheng
BioMedInformatics 2025, 5(1), 6; https://doi.org/10.3390/biomedinformatics5010006 - 22 Jan 2025
Cited by 4 | Viewed by 2949
Abstract
Introduction: Remote health monitoring plays a crucial role in telehealth services and the effective management of patients, which can be enhanced by vital sign prediction from facial videos. Facial videos are easily captured through various imaging devices like phone cameras, webcams, or [...] Read more.
Introduction: Remote health monitoring plays a crucial role in telehealth services and the effective management of patients, which can be enhanced by vital sign prediction from facial videos. Facial videos are easily captured through various imaging devices like phone cameras, webcams, or surveillance systems. Methods: This study introduces a hybrid deep learning model aimed at estimating heart rate (HR), blood oxygen saturation level (SpO2), and blood pressure (BP) from facial videos. The hybrid model integrates convolutional neural network (CNN), convolutional long short-term memory (convLSTM), and video vision transformer (ViViT) architectures to ensure comprehensive analysis. Given the temporal variability of HR and BP, emphasis is placed on temporal resolution during feature extraction. The CNN processes video frames one by one while convLSTM and ViViT handle sequences of frames. These high-resolution temporal features are fused to predict HR, BP, and SpO2, capturing their dynamic variations effectively. Results: The dataset encompasses 891 subjects of diverse races and ages, and preprocessing includes facial detection and data normalization. Experimental results demonstrate high accuracies in predicting HR, SpO2, and BP using the proposed hybrid models. Discussion: Facial images can be easily captured using smartphones, which offers an economical and convenient solution for vital sign monitoring, particularly beneficial for elderly individuals or during outbreaks of contagious diseases like COVID-19. The proposed models were only validated on one dataset. However, the dataset (size, representation, diversity, balance, and processing) plays an important role in any data-driven models including ours. Conclusions: Through experiments, we observed the hybrid model’s efficacy in predicting vital signs such as HR, SpO2, SBP, and DBP, along with demographic variables like sex and age. There is potential for extending the hybrid model to estimate additional vital signs such as body temperature and respiration rate. Full article
(This article belongs to the Section Applied Biomedical Data Science)
Show Figures

Figure 1

14 pages, 474 KB  
Communication
Low-Complexity Timing Correction Methods for Heart Rate Estimation Using Remote Photoplethysmography
by Chun-Chi Chen, Song-Xian Lin and Hyundoo Jeong
Sensors 2025, 25(2), 588; https://doi.org/10.3390/s25020588 - 20 Jan 2025
Cited by 3 | Viewed by 4048
Abstract
With the rise of modern healthcare monitoring, heart rate (HR) estimation using remote photoplethysmography (rPPG) has gained attention for its non-contact, continuous tracking capabilities. However, most HR estimation methods rely on stable, fixed sampling intervals, while practical image capture often involves irregular frame [...] Read more.
With the rise of modern healthcare monitoring, heart rate (HR) estimation using remote photoplethysmography (rPPG) has gained attention for its non-contact, continuous tracking capabilities. However, most HR estimation methods rely on stable, fixed sampling intervals, while practical image capture often involves irregular frame rates and missing data, leading to inaccuracies in HR measurements. This study addresses these issues by introducing low-complexity timing correction methods, including linear, cubic, and filter interpolation, to improve HR estimation from rPPG signals under conditions of irregular sampling and data loss. Through a comparative analysis, this study offers insights into efficient timing correction techniques for enhancing HR estimation from rPPG, particularly suitable for edge-computing applications where low computational complexity is essential. Cubic interpolation can provide robust performance in reconstructing signals but requires higher computational resources, while linear and filter interpolation offer more efficient solutions. The proposed low-complexity timing correction methods improve the reliability of rPPG-based HR estimation, making it a more robust solution for real-world healthcare applications. Full article
Show Figures

Figure 1

17 pages, 7070 KB  
Article
Research on Heart Rate Detection from Facial Videos Based on an Attention Mechanism 3D Convolutional Neural Network
by Xiujuan Sun, Ying Su, Xiankai Hou, Xiaolan Yuan, Hongxue Li and Chuanjiang Wang
Electronics 2025, 14(2), 269; https://doi.org/10.3390/electronics14020269 - 10 Jan 2025
Cited by 1 | Viewed by 3114
Abstract
Remote photoplethysmography (rPPG) has attracted growing attention due to its non-contact nature. However, existing non-contact heart rate detection methods are often affected by noise from motion artifacts and changes in lighting, which can lead to a decrease in detection accuracy. To solve this [...] Read more.
Remote photoplethysmography (rPPG) has attracted growing attention due to its non-contact nature. However, existing non-contact heart rate detection methods are often affected by noise from motion artifacts and changes in lighting, which can lead to a decrease in detection accuracy. To solve this problem, this paper initially employs manual extraction to precisely define the facial Region of Interest (ROI), expanding the facial area while avoiding rigid regions such as the eyes and mouth to minimize the impact of motion artifacts. Additionally, during the training phase, illumination normalization is employed on video frames with uneven lighting to mitigate noise caused by lighting fluctuations. Finally, this paper introduces a 3D convolutional neural network (CNN) method incorporating an attention mechanism for heart rate detection from facial videos. We optimize the traditional 3D-CNN to capture global features in spatiotemporal data more effectively. The SimAM attention mechanism is introduced to enable the model to precisely focus on and enhance facial ROI feature representations. Following the extraction of rPPG signals, a heart rate estimation network using a bidirectional long short-term memory (BiLSTM) model is employed to derive the heart rate from the signals. The method introduced here is experimentally validated on two publicly available datasets, UBFC-rPPG and PURE. The mean absolute errors were 0.24 bpm and 0.65 bpm, the root mean square errors were 0.63 bpm and 1.30 bpm, and the Pearson correlation coefficients reached 0.99, confirming the method’s reliability. Comparisons of predicted signals with ground truth signals further validated its accuracy. Full article
(This article belongs to the Section Artificial Intelligence)
Show Figures

Figure 1

14 pages, 793 KB  
Article
MFF-Net: A Lightweight Multi-Frequency Network for Measuring Heart Rhythm from Facial Videos
by Wenqin Yan, Jialiang Zhuang, Yuheng Chen, Yun Zhang and Xiujuan Zheng
Sensors 2024, 24(24), 7937; https://doi.org/10.3390/s24247937 - 12 Dec 2024
Cited by 2 | Viewed by 1431
Abstract
Remote photo-plethysmography (rPPG) is a useful camera-based health motioning method that can measure the heart rhythm from facial videos. Many well-established deep learning models can provide highly accurate and robust results in measuring heart rate (HR) and heart rate variability (HRV). However, these [...] Read more.
Remote photo-plethysmography (rPPG) is a useful camera-based health motioning method that can measure the heart rhythm from facial videos. Many well-established deep learning models can provide highly accurate and robust results in measuring heart rate (HR) and heart rate variability (HRV). However, these methods are unable to effectively eliminate illumination variation and motion artifact disturbances, and their substantial computational resource requirements significantly limit their applicability in real-world scenarios. Hence, we propose a lightweight multi-frequency network named MFF-Net to measure heart rhythm via facial videos in a short time. Firstly, we propose a multi-frequency mode signal fusion (MFF) mechanism, which can separate the characteristics of different modes of the original rPPG signals and send them to a processor with independent parameters, helping the network recover blood volume pulse (BVP) signals accurately under a complex noise environment. In addition, in order to help the network extract the characteristics of different modal signals effectively, we designed a temporal multiscale convolution module (TMSC-module) and spectrum self-attention module (SSA-module). The TMSC-module can expand the receptive field of the signal-refining network, obtain more abundant multiscale information, and transmit it to the signal reconstruction network. The SSA-module can help a signal reconstruction network locate the obvious inferior parts in the reconstruction process so as to make better decisions when merging multi-dimensional signals. Finally, in order to solve the over-fitting phenomenon that easily occurs in the network, we propose an over-fitting sampling training scheme to further improve the fitting ability of the network. Comprehensive experiments were conducted on three benchmark datasets, and we estimated HR and HRV based on the BVP signals derived by MFF-Net. Compared with state-of-the-art methods, our approach achieves better performance both on HR and HRV estimation with lower computational burden. We can conclude that the proposed MFF-Net has the opportunity to be applied in many real-world scenarios. Full article
(This article belongs to the Section Sensor Networks)
Show Figures

Figure 1

Back to TopTop