Next Article in Journal
Machine Learning-Based Blockchain Technology for Secure V2X Communication: Open Challenges and Solutions
Previous Article in Journal
Spectrum Sensing for Noncircular Signals Using Augmented Covariance-Matrix-Aware Deep Convolutional Neural Network
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Review

Contactless Vital Sign Monitoring: A Review Towards Multi-Modal Multi-Task Approaches

Department of Information Security and Communication Technology, Norwegian University of Science and Technology (NTNU), 2815 Gjovik, Norway
*
Author to whom correspondence should be addressed.
Sensors 2025, 25(15), 4792; https://doi.org/10.3390/s25154792
Submission received: 20 May 2025 / Revised: 16 July 2025 / Accepted: 24 July 2025 / Published: 4 August 2025
(This article belongs to the Section Biomedical Sensors)

Abstract

Contactless vital sign monitoring has emerged as a transformative healthcare technology, enabling the assessment of vital signs without physical contact with the human body. This review comprehensively reviews the rapidly evolving landscape of this field, with particular emphasis on multi-modal sensing approaches and multi-task learning paradigms. We systematically categorize and analyze existing technologies based on sensing modalities (vision-based, radar-based, thermal imaging, and ambient sensing), integration strategies, and application domains. The paper examines how artificial intelligence has revolutionized this domain, transitioning from early single-modality, single-parameter approaches to sophisticated systems that combine complementary sensing technologies and simultaneously extract multiple vital sign parameters. We discuss the theoretical foundations and practical implementations of multi-modal fusion, analyzing signal-level, feature-level, decision-level, and deep learning approaches to sensor integration. Similarly, we explore multi-task learning frameworks that leverage the inherent relationships between vital sign parameters to enhance measurement accuracy and efficiency. The review also critically addresses persisting technical challenges, clinical limitations, and ethical considerations, including environmental robustness, cross-subject variability, sensor fusion complexities, and privacy concerns. Finally, we outline promising future directions, from emerging sensing technologies and advanced fusion architectures to novel application domains and privacy-preserving methodologies. This review provides a holistic perspective on contactless vital sign monitoring, serving as a reference for researchers and practitioners in this rapidly advancing field.

1. Introduction

Contactless vital sign monitoring represents a paradigm shift in healthcare technology, enabling the measurement of vital signs without physical contact with the human body. This approach addresses limitations of traditional contact-based methods, which can cause discomfort during prolonged monitoring, skin irritation, and measurement artifacts due to sensor displacement. Moreover, Contactless methods are particularly valuable for vulnerable populations such as neonates, burn patients, and elderly individuals with fragile skin [1].
The integration of artificial intelligence (AI) has significantly advanced this field, transforming it into a sophisticated ecosystem that captures, analyzes, and interprets vital health data with unprecedented accuracy and efficiency. AI algorithms work alongside various sensors, such as cameras, radar, and infrared systems, to process complex vital sign signals in real-time, enabling continuous health assessment without the constraints of physical contact. The synergy between sensor technologies and AI enhances both the range of measurable parameters and the robustness of measurements across varying environmental conditions.
The COVID-19 pandemic has further accelerated interest in these technologies, highlighting the need for monitoring solutions that minimize physical contact and reduce cross-contamination risks [2]. Additionally, the growing focus on remote patient monitoring and telemedicine creates an urgent need for reliable Contactless monitoring systems that can operate in uncontrolled environments [3]. AI-driven systems can adapt to these challenging scenarios, learning from diverse data to maintain accuracy despite variations in lighting, subject positioning, and background noise.
The field has evolved from early single-modality approaches, such as remote photoplethysmography (rPPG) for heart rate extraction from facial videos [4,5], to more sophisticated multi-modal and AI-enhanced systems. Early systems typically focused on extracting a single vital sign parameter through deterministic signal processing, while contemporary research increasingly aims to simultaneously monitor multiple vital signs using complementary sensing technologies augmented by machine learning [6].
The state-of-the-art in contactless vital sign monitoring using AI encompasses diverse modalities:
  • Video-based monitoring: Techniques like remote photoplethysmography (rPPG) leverage computer vision and deep learning to capture vital signs such as heart rate and respiratory rate by analyzing subtle skin color changes or minute facial movements. Advanced neural networks have significantly improved robustness to challenges such as subject motion, varying skin tones, and inconsistent lighting conditions [7].
  • Audio-based systems: These approaches process respiratory sounds and heartbeats using audio signal processing and AI techniques like convolutional neural networks (CNNs) to measure vital signs through sound analysis, even in noisy environments [8].
  • Signal-based methods: Radar, WiFi, and electromagnetic sensors monitor heart and respiratory rates non-invasively by detecting subtle body movements, with AI algorithms separating vital sign signals from noise and motion artifacts [9,10,11].
  • Thermal imaging: AI-enhanced thermal cameras detect minute temperature variations associated with blood flow and respiration, enabling non-visible light monitoring, especially valuable in low-light conditions [12,13].
These modalities increasingly converge in multi-modal systems that leverage AI for sensor fusion, combining the strengths of different sensing approaches while mitigating their individual limitations (see Figure 1).
AI-based multi-task systems represent a significant advancement in vital sign measurement, simultaneously tracking multiple vital sign parameters using shared computational frameworks. Unlike traditional approaches that employ separate models for each vital sign, multi-task learning architectures leverage common representations across related vital sign parameters, enhancing efficiency and reducing computational requirements.
These systems monitor various vital signs such as heart rate, respiratory rate, blood oxygen saturation, and blood pressure concurrently, often by integrating data from different modalities like video, audio, and radar signals. By optimizing across multiple objectives simultaneously, these architectures can exploit the inherent relationships between vital sign parameters, improving overall accuracy and robustness. The system can cross-reference data streams to detect anomalies or correlations between vital signs, providing more comprehensive health assessments than single-task approaches.
This multi-task paradigm is particularly valuable in continuous monitoring scenarios, such as telemedicine or intensive care, where comprehensive and real-time health assessments are critical. Recent work by researchers such as Chen et al. [14] has demonstrated substantial improvements in accuracy when jointly estimating multiple vital sign parameters compared to independent models.

Scope and Objectives

This work reviews the state-of-the-art in contactless vital sign monitoring, with particular emphasis on AI-driven multi-modal sensing and multi-task learning approaches. We aim to:
  • Review the state-of-the-art in contactless vital sign monitoring, with particular emphasis on AI-driven multi-modal sensing and multi-task learning approaches.
  • Analyze the benefits and challenges of integrating multiple sensing modalities through AI-based fusion methods.
  • Evaluate multi-task learning frameworks that simultaneously extract multiple vital sign parameters with shared representations.
  • Identify current limitations in AI-based systems and promising research directions.
Unlike previous surveys that focused on specific modalities such as vision-based approaches [15] or radar-based methods [9], our work presents a view of the field with emphasis on AI-enabled integration strategies for complementary technologies and multi-task learning paradigms that maximize information extraction from available sensing modalities.

2. Background

2.1. Vital Sign Parameters of Interest

Contactless monitoring technologies target a range of vital parameters essential for health assessment. Cardiac parameters include heart rate (HR), heart rate variability (HRV), and blood pressure (BP), which provide critical insights into cardiovascular function and autonomic nervous system activity. Respiratory parameters such as respiratory rate (RR), breathing patterns, and apnea detection offer information about pulmonary function and can indicate respiratory distress or sleep disorders. Thermal regulation, measured through body and facial temperature distribution, reflects metabolic activity and can indicate inflammation or infection. Blood oxygen saturation (SpO2) measures the percentage of hemoglobin binding sites occupied by oxygen molecules, crucial for detecting hypoxemia and respiratory insufficiency. Emotional and stress indicators, including galvanic skin response, pupil dilation, and facial expressions, provide windows into psychological states and autonomic responses. Sleep quality metrics encompassing body movements and sleep stage classification enable assessment of rest patterns and potential sleep disorders. Each of these parameter categories provides unique insights into vital sign state, making comprehensive monitoring valuable for holistic health assessment [16].

2.2. Traditional Contact-Based Monitoring

Conventional vital sign monitoring relies on various contact-based technologies that serve as clinical gold standards. Electrocardiography (ECG) measures cardiac electrical activity through electrodes attached to the skin, providing detailed information about heart rhythm and conduction abnormalities [17]. Photoplethysmography (PPG) detects blood volume changes in the microvascular bed of tissue through light-based sensors, commonly used in pulse oximeters for heart rate and oxygen saturation monitoring [18]. Respiratory inductance plethysmography utilizes bands around the chest and abdomen to measure dimensional changes during breathing cycles, enabling detailed breathing pattern analysis [19]. Electrodermal activity sensors detect changes in skin conductance related to sweat gland activity, providing objective measures of stress and emotional responses [20]. Polysomnography combines multiple sensor types to comprehensively monitor sleep physiology, including brain activity, eye movements, muscle activity, and cardiorespiratory parameters [21]. While these methods remain the clinical gold standard, they present limitations including restricted mobility, discomfort during prolonged use, and potential for skin irritation. Furthermore, they typically require professional placement and calibration, limiting accessibility in resource-constrained environments [22].

2.3. Emergence of Contactless Alternatives

The shift toward contactless monitoring began with single-parameter, single-modality approaches. Pioneering work by Verkruysse et al. [5] demonstrated that ambient light could enable camera-based PPG. Subsequently, Wu et al. [23] improved signal quality through advanced signal processing, while Poh et al. [4] introduced blind source separation techniques to enhance robustness.
Parallel developments emerged in radar-based monitoring, with continuous-wave Doppler radar demonstrating the capability to detect minute chest wall movements for respiration monitoring [24]. Thermal imaging similarly evolved from military and industrial applications to vital sign monitoring, with Pavlidis and Levine [13] demonstrating facial temperature mapping for stress detection.
These independent technological trajectories have increasingly converged, with researchers recognizing the complementary strengths of different modalities [25]. This convergence has led to the development of multi-modal systems that combine various sensing technologies to overcome the limitations of individual approaches, paving the way for more robust and comprehensive contactless monitoring solutions. The integration of artificial intelligence has further accelerated this evolution, enabling real-time processing of complex vital sign signals and adapting to challenging real-world monitoring environments.

3. Contactless Monitoring Technologies

Contactless monitoring technologies encompass diverse sensing modalities that enable vital sign measurement without physical contact (see Figure 2). The primary approaches include vision-based methods analyzing color variations and motion patterns, radar systems detecting body movements, thermal imaging capturing blood flow variations, and ambient sensing leveraging environmental changes. Each modality offers complementary strengths and challenges, motivating multi-modal fusion approaches enhanced by artificial intelligence for robust real-time physiological monitoring.

3.1. Vision-Based Methods

Camera-based approaches extract vital sign signals through subtle visual cues, primarily using the following techniques:

3.1.1. Remote Photoplethysmography (rPPG)

Remote PPG leverages color changes in skin pixels to extract pulse signals. Early methods relied on color magnification and fixed regions of interest [5], evolving to more sophisticated approaches. Blind source separation techniques, including independent component analysis, were pioneered by Poh et al. [4] to isolate the pulse signal from video recordings of the face. Signal processing approaches such as CHROM [26] and POS [27] have further enhanced signal quality by addressing noise sources and motion artifacts. More recently, deep learning methods have revolutionized the field by enabling robust feature extraction directly from raw video data [7,28]. These advancements have collectively addressed key challenges in varying lighting conditions [29], subject motion [30], and vital sign monitoring across different skin tones [31], making rPPG increasingly viable for real-world applications.

3.1.2. Motion-Based Analysis

Motion-based methods track subtle body movements to extract vital signs through advanced computer vision techniques. Ballistocardiographic (BCG) imaging captures the minute mechanical motions of the body caused by the ejection of blood from the heart, enabling heart-related vital sign measurements without direct skin exposure [32]. Eulerian video magnification, pioneered by Wu et al. [23], amplifies subtle motions that are imperceptible to the naked eye, making it particularly effective for respiratory motion detection. In recent years, deep learning approaches have extended these capabilities by enabling more sensitive motion amplification and tracking across various body regions and under challenging conditions [7]. Motion-based methods have proven particularly valuable for respiration monitoring and can effectively complement color-based approaches in multi-modal systems, offering redundancy when one signal source becomes temporarily unavailable or corrupted [33].

3.2. Radar-Based Methods

Radar technology detects vital sign parameters through reflected electromagnetic waves, offering the advantages of penetration through clothing and operation in darkness. Continuous-wave Doppler radar systems detect frequency shifts in reflected signals caused by body movements, enabling vital sign extraction through phase modulation [24]. Advanced signal processing techniques for heart and respiratory rate separation have been developed to distinguish between these parameters with increasing accuracy [34], while beamforming techniques have improved spatial resolution to enable monitoring of specific body regions [35]. Frequency-Modulated Continuous-Wave (FMCW) radar provides additional depth information through frequency sweeping, facilitating multi-person monitoring in shared spaces [9], fine-grained respiratory pattern analysis [36], and combined heart and respiratory monitoring with enhanced motion artifact reduction [11]. Ultra-Wideband (UWB) radar offers high temporal resolution through short pulses, enabling through-wall vital sign monitoring [37], sleep apnea detection through respiratory pattern analysis [38], and cardiac activity characterization beyond simple heart rate [39]. The non-optical nature of radar makes it complementary to vision-based approaches, particularly in low-light conditions or when privacy is a concern.

3.3. Thermal Imaging

Infrared thermography measures temperature variations associated with physiological processes, offering unique capabilities for non-visible light monitoring. Cutaneous perfusion monitoring techniques detect temperature changes related to blood flow, with early work by Pavlidis and Levine [13] demonstrating facial blood flow mapping for stress detection. Subsequent research by Garbey et al. [40] established methods for pulse extraction through thermal signal processing, while more recent deep learning approaches have enabled robust feature extraction from thermal sequences under varying environmental conditions [41]. Respiratory monitoring via thermal imaging exploits temperature differences between inhaled and exhaled air, with techniques developed for nostril temperature fluctuation analysis [42], breathing pattern classification for respiratory disorder detection [43], and fusion with visible-light methods for enhanced robustness across lighting conditions [12]. Thermal imaging is particularly valuable in low-light environments and can operate independently of ambient lighting, making it an important complementary modality in comprehensive monitoring systems.

3.4. Ambient Sensing

Environmental sensors detect vital sign signals through disturbances in surrounding fields, enabling monitoring without dedicated wearable or visible devices. RF-based methods leverage radio frequency signals to detect vital signs through perturbations in the electromagnetic field. WiFi CSI-based vital sign monitoring [44] utilizes existing WiFi infrastructure to detect minute movements associated with respiration and heartbeats. RFID-based sensing enables activity and vital signs monitoring through passive tags [45], while millimeter-wave sensing provides high-precision monitoring through finer spatial resolution [46]. Acoustic methods detect cardiac and respiratory sounds through various microphone configurations. Ambient microphone arrays facilitate breathing monitoring through sound localization and enhancement [8], ultrasonic sensing can detect subtle chest wall movement even in noisy environments [47], and smart speakers can be repurposed for sleep apnea detection through sophisticated audio processing. Electric field sensing detects body-induced distortions in electric fields for unobtrusive monitoring. Capacitive sensing enables bed-integrated monitoring without direct skin contact [48], electric potential sensing works through clothing and bedding for sleep monitoring [49], and smart fabric and furniture integration embeds sensing capabilities into everyday objects [50]. These ambient sensing modalities collectively enable monitoring that blends into the environment, reducing user awareness and potentially improving compliance.

4. Comparative Analysis of AI-Based Approaches

As contactless vital sign monitoring has evolved from traditional signal processing methods to sophisticated AI-driven approaches, it is instructive to analyze the landscape of recent research contributions. Table 1 provides an overview of notable papers in the field, categorizing them based on the vital signs they monitor, the methodologies they employ, and the modalities they leverage.
This comparative analysis reveals several important trends. First, there is a clear predominance of video/image-based approaches, reflecting the accessibility and rich information content provided by visual data. Second, heart rate (HR) remains the most commonly monitored vital sign, though respiratory rate (RR) is increasingly included in multi-task systems. Third, we observe a methodological evolution from earlier CNN-based architectures to more recent transformer-based approaches, demonstrating the field’s rapid adoption of state-of-the-art deep learning techniques. Fourth, while most systems focus on a single modality, there is a growing interest in multi-modal approaches that combine complementary data sources (e.g., Choi et al. 2024 [51] integrating visual and radio frequency data).
Table 1. Comparative analysis of AI-based contactless vital sign monitoring approaches from 2018–2024, categorized by vital signs monitored and modalities used.
Table 1. Comparative analysis of AI-based contactless vital sign monitoring approaches from 2018–2024, categorized by vital signs monitored and modalities used.
Paper/YearDatasetMethodologyVital SignsModality
HRHRVRRBPSpO2ImageAudioTextSignal
Wu et al., 2022 [52]data collectedF-Net, S-Net-------
Bukum et al., 2022 [53]data collectedCNN--------
Špetlík et al., 2018 [54]MAHNOB, PUREHR-CNN-------
Chen & McDuff, 2018 [7]RGB Video I, IIDeepPhys------
Liu et al., 2020 [28]AFRL, MMSE-HRMTTS-CAN------
Yu et al., 2021 [55]VIPL-HRPhysFormer------
Qiu et al., 2018 [56]MMSE-HRCNN-------
Luguev et al., 2020 [57]MAHNOB3D-CNN-------
Zhan et al., 2020 [58]HNU, PURECNN-------
Yu et al., 2019 [59]MAHNOBCNN-------
Du et al., 2022 [60]UTA-RLDDCNN-------
Suriani et al., 2022 [61]UTA-RLDDCNN------
Lorato et al., 2022 [62]UTA-RLDDCNN------
Niu et al., 2019 [63]VIPL-HRCNN-------
Niu et al., 2020 [64]VIPL-HRCNN+RNN-------
Huang et al., 2020 [65]Data collectedCNN+LSTM-------
Song et al., 2020 [66]MAHNOBCNN-------
Botina et al., 2020 [67]MSECLSTM-------
Huang et al., 2021 [68]MAHNOB, UBFC3D-CNN+LSTM-------
Gao et al., 2022 [69]EIIPHCILSTM-------
Napolean et al., 2022 [70]IntensePhysioCNN-------
Wu et al., 2023 [71]ECG-FitnessCNN-------
Gua et al., 2023 [72]COHFACETransformer+CNN-------
Othman et al., 2024 [73]LGI-PPGITransformer+LSTM-------
Choi et al., 2024 [51]own datasetTransformer+CNN------
Xu et al., 2023 [74]own datasetLSTM-------
Wang et al., 2023 [75]COHFACETransformer-------
He et al., 2023 [76]Medical+RFTransformer-------
Xu et al., 2022 [77]own datasetLSTM-------
Marchi et al., 2019 [78]own datasetCNN+RNN------
Deshpande et al., 2020 [79]VOCALSTransformer------
Kim et al., 2021 [80]Stress datasetCNN------
Pimentel et al., 2022 [81]own datasetCNN+LSTM-------
Amiriparian et al., 2022 [82]COVID-19 audioTransformer-------
Chen et al., 2023 [83]Breathing SoundCNN+Attention-------
Rahman et al., 2024 [84]Infant cryTransformer+CNN-----
Notable progression includes the introduction of attention mechanisms in DeepPhys (Chen and McDuff, 2018 [7]), the development of dedicated architectures for multi-task learning in MTTS-CAN (Liu et al., 2020 [28]), and the more recent adoption of transformer-based architectures (PhysFormer, Yu et al., 2021 [55]) that have demonstrated superior performance in capturing both spatial and temporal dynamics in vital sign signals.
It is also worth noting the increasing diversity of datasets being used for model development and validation, reflecting the field’s growing maturity and concern with generalizability across different populations and contexts. However, the table also highlights certain gaps in the current research landscape, particularly the limited attention to heart rate variability (HRV), blood pressure (BP), and blood oxygen saturation (SpO2), which represent promising directions for future work.

4.1. Key Datasets for Contactless Vital Sign Monitoring

The evolution of contactless vital sign monitoring has been enabled by several important benchmark datasets (see Table 2) that provide standardized video recordings with ground-truth vital sign measurements. These datasets serve as essential resources for developing and evaluating new algorithms, enabling fair comparisons between different approaches, and establishing baseline performance metrics.

4.1.1. RGB Video-Based Datasets

MAHNOB-HCI [85] stands as one of the most widely used datasets in the field, containing multi-modal recordings of 27 participants’ spontaneous emotional responses while watching emotional video clips. Each recording includes frontal face videos, vital sign signals (ECG, respiration, skin temperature), and eye-gaze data, making it valuable for both vital sign monitoring and affective computing research. The dataset’s diversity in emotional responses creates challenging conditions for heart rate estimation algorithms, making it a rigorous benchmark for robustness evaluation.
PURE [86] (Pulse Rate Detection Dataset) provides RGB video recordings of 10 subjects performing controlled head movements under varying lighting conditions, accompanied by pulse oximeter ground truth measurements. This dataset specifically addresses the challenge of motion artifacts in rPPG and enables systematic evaluation of algorithm performance under controlled motion scenarios.
COHFACE [87] contains recordings of 40 subjects under two different illumination conditions, making it particularly valuable for evaluating the robustness of algorithms to lighting variations. Each recording is accompanied by synchronized vital sign measurements, including heart rate and respiration rate obtained from medical-grade sensors.
VIPL-HR [63] represents one of the largest and most diverse rPPG datasets, containing 2378 visible light videos of 107 subjects with various skin tones, head movements, and illumination conditions. Ground-truth measurements include synchronized signals from pulse oximeters and chest straps, enabling evaluation across multiple vital sign parameters.
UTA-RLDD [88] (Real-Life Drowsiness Dataset) contains over 30 h of naturalistic driving data from 60 participants exhibiting various levels of drowsiness. While primarily designed for drowsiness detection, researchers have leveraged this dataset for heart rate and respiratory rate extraction under realistic driving conditions.

4.1.2. Multi-Modal and Specialized Datasets

MMSE-HR [89] (Multimodal Spontaneous Expression-Heart Rate) contains recordings of 40 participants experiencing spontaneous emotions, with synchronized vital sign measurements. The dataset’s focus on naturalistic emotional expressions makes it valuable for evaluating heart rate monitoring during affective states.
UBFC-rPPG [90] contains 42 videos specifically recorded for remote photoplethysmography, with subjects at rest facing a camera under ambient lighting. The uncompressed video format preserves subtle color variations critical for rPPG analysis, while synchronized finger pulse oximeter signals provide reliable ground truth.
UBFC-Phys [91] extends the UBFC collection with recordings of 56 subjects performing cognitive tasks designed to induce stress, enabling the study of vital sign responses to mental workload. Multi-modal measurements include facial videos and reference vital sign signals (ECG, EDA, respiration).
LGI-PPGI [92] (Local Group Invariance-Photoplethysmography Imaging) provides recordings of 25 subjects with varying skin tones and under different lighting conditions, with emphasis on controlled illumination variations to enable systematic evaluation of algorithm robustness.
IntensePhysio [93] contains multi-modal recordings of 50 subjects performing high-intensity physical activities, addressing the challenging scenario of monitoring vital sign parameters during significant motion and increased cardiorespiratory responses.
The Vision for Vitals dataset [94] specifically addresses diversity challenges by including subjects with varying skin tones, ages, and genders, enabling evaluation of algorithm performance across demographic groups and assessment of potential biases.
ECG-Fitness [95] focuses on monitoring during physical exercise, with synchronized ECG measurements during various fitness activities, enabling evaluation of algorithms under challenging motion and vital sign stress conditions.
Several researchers have also collected custom datasets targeting specific challenges or application scenarios, including specialized illumination conditions, particular demographic groups, or novel sensing modalities like radio frequency or infrared imaging.
Table 3 presents the best reported accuracies achieved on key contactless vital sign monitoring datasets. The results reveal significant performance variations across datasets, with clinical datasets like BIDMC PPG achieving the highest accuracies (98.3% for heart rate, 93.8% for respiratory rate) due to controlled environments and high-quality ground truth measurements. Vision-based datasets under controlled conditions, such as PURE (97.8% HR accuracy) and UBFC-rPPG (97.2% HR accuracy), demonstrate the maturity of remote photoplethysmography techniques. However, challenging real-world scenarios show notable performance degradation, with IntensePhysio achieving 89.3% HR accuracy during physical exercise, highlighting the ongoing challenges in motion-robust monitoring. The correlation coefficients consistently exceed 0.85 for most datasets, indicating strong agreement between contactless measurements and ground truth, while MAPE values below 5% for top-performing methods suggest clinical viability for many applications.

4.2. Methodologies in Contactless Monitoring

4.2.1. Convolutional Neural Network Architectures

Convolutional Neural Networks (CNNs) form the backbone of many vision-based vital sign monitoring approaches. Unlike traditional methods that rely on handcrafted features, CNNs automatically learn hierarchical representations directly from raw video data. HR-CNN [54] pioneered end-to-end learning for heart rate estimation from face videos, employing a two-stage architecture that first extracts pulse signals from facial regions and then estimates heart rate through frequency analysis. This approach significantly outperformed traditional signal processing methods on standard benchmarks.
3D-CNNs extend standard convolutional operations to the temporal dimension, enabling simultaneous modeling of spatial and temporal information critical for capturing subtle vital sign signals in video sequences. These architectures have proven particularly effective for motion-resistant heart rate estimation [57], as they can better differentiate between physiological color changes and motion artifacts by explicitly modeling temporal dynamics.
DeepPhys [7] introduced an attention mechanism within a CNN framework, employing a two-stream architecture with motion representation and appearance streams. The attention mechanism allows the network to focus on regions with strong vital sign signals while suppressing noise, significantly improving performance during subject movement or under varying lighting conditions.

4.2.2. Temporal Modeling Approaches

Recurrent Neural Networks (RNNs) and Long Short-Term Memory (LSTM) networks have been widely adopted to model the temporal dynamics of vital sign signals. These architectures maintain an internal state that captures information from previous time steps, making them well-suited for sequential data like video frames or extracted pulse signals. RNN-based approaches have demonstrated superior performance in scenarios with irregular motion or lighting changes [64], as they can better track the continuity of vital sign signals across frames.
Combined CNN-LSTM architectures leverage the strengths of both approaches, with CNNs extracting spatial features from individual frames and LSTMs modeling their temporal evolution. This hybrid approach has proven particularly effective for robust heart rate estimation [65], as it captures both the spatial characteristics of blood flow patterns and their temporal dynamics.

4.2.3. Multi-Task and Transformer-Based Architectures

MTTS-CAN (Multi-Task Temporal Shift-Convolutional Attention Network) [28] represents a significant advancement in multi-task vital sign monitoring (see Figure 3). This architecture employs a shared feature extractor with task-specific attention mechanisms, enabling simultaneous estimation of multiple vital sign parameters including heart rate and respiration rate. The temporal shift module efficiently captures temporal relationships without the computational overhead of 3D convolutions or recurrent layers.
PhysFormer [55] introduced transformer architectures to vital sign monitoring, employing self-attention mechanisms to capture long-range dependencies in both spatial and temporal dimensions. The architecture’s temporal difference module specifically enhances the model’s sensitivity to subtle vital sign signals while suppressing noise. PhysFormer has established new state-of-the-art performance on multiple benchmarks, demonstrating transformer models’ superior capability for integrating information across extended temporal contexts.
More recent approaches like the F-Net, S-Net, and FS-Net architectures [52] focus on specialized tasks such as blood pressure estimation from facial videos, employing custom-designed network components to extract features specifically relevant to particular vital sign parameters, as shown in Figure 4.

4.2.4. Emerging Architectural Paradigms

The latest research shows increasing adoption of hybrid approaches that combine transformers with CNNs [72], leveraging CNNs’ efficiency in local feature extraction with transformers’ strength in modeling long-range dependencies. These hybrid architectures often achieve superior performance while maintaining reasonable computational requirements.
Multi-modal fusion architectures [51] (see Figure 5 as an example) represent another emerging trend, integrating information from complementary sensing modalities such as visual and radio frequency data. These approaches employ specialized fusion modules to align and combine features from heterogeneous data sources, often achieving greater robustness across varying environmental conditions.
Self-supervised learning approaches are gaining traction as a means to leverage large unlabeled datasets, with models pre-trained on video-understanding tasks and then fine-tuned for vital sign parameter estimation [75]. This paradigm shows particular promise for addressing the challenge of limited labeled data in vital sign monitoring.

5. Multi-Modal Approaches

5.1. Rationale for Multi-Modal Sensing

Single-modality approaches to contactless vital sign monitoring encounter several inherent limitations that restrict their effectiveness in real-world scenarios. These include environmental dependencies, such as lighting conditions severely affecting vision-based methods; motion artifacts and positioning constraints that can degrade signal quality; and fundamental physical limitations inherent to individual sensing mechanisms. These constraints have motivated researchers to develop multi-modal systems that integrate complementary technologies to overcome these limitations. Such integrated systems offer increased robustness across varying environmental conditions, as alternative modalities can compensate when primary sensors encounter challenging situations. They also provide extended parameter coverage beyond what single-modality approaches can achieve, and deliver improved accuracy through cross-modal validation and fusion techniques [3]. This improvement is not merely incremental—recent research has demonstrated that combining visual and thermal imaging increases heart rate detection accuracy in challenging lighting conditions compared to visual-only methods, highlighting the substantial benefits of multi-modal integration [3,12,106].

5.2. Complementary Modality Combinations

5.2.1. Visual-Thermal Integration

Visual and thermal cameras offer naturally complementary information that, when integrated, address significant limitations of each individual technology. Thermal imaging effectively compensates for poor visible light conditions, enabling continued monitoring in low-light environments where traditional RGB cameras fail [12]. Conversely, color information from visual cameras enhances thermal-based blood flow mapping by providing contextual information about skin regions and facial landmarks [106]. This synergistic combination has proven particularly valuable for applications requiring comprehensive vital sign assessment, such as perspiration monitoring and emotional state detection, where thermal signatures of facial regions combine with visual cues to provide more robust indicators [12,106]. The integration of these modalities has enabled monitoring systems to function across a wider range of environmental conditions while extracting a more diverse set of vital sign parameters than either modality could achieve independently.

5.2.2. Camera–Radar Systems

Camera–radar combinations leverage the fine-grained spatial information and rich visual content from cameras with the penetrative capabilities and motion sensitivity of radar systems. Camera-guided radar focusing has proven particularly effective for monitoring in multi-person environments, where visual identification can direct radar attention to specific individuals for targeted vital sign assessment [9]. These hybrid systems excel at motion artifact compensation by using cross-modal information, where radar data can help distinguish between vital sign movements and general body motion captured in video streams [11]. Additionally, such combinations have enabled enhanced vital sign extraction in through-obstacle scenarios, with visual data providing contextual information while radar penetrates clothing or bedding materials to detect subtle movements associated with cardiac and respiratory activity [3]. This fusion approach combines the strengths of non-contact optical sensing with the environmental robustness of radio frequency techniques.

5.2.3. Comprehensive Multi-Modal Systems

Advanced monitoring systems have evolved beyond two-modality combinations to integrate three or more sensing technologies, creating comprehensive platforms that maximize robustness and parameter coverage. Visual–thermal–radar platforms represent one such approach, combining the distinct advantages of each modality to enable comprehensive monitoring across varying environmental conditions and vital sign parameters [6]. Smart room environments with distributed heterogeneous sensors take this concept further, embedding various monitoring technologies throughout living spaces to create ubiquitous sensing capabilities that blend into the environment [107]. These sophisticated systems have also addressed the challenge of multi-person tracking and vital sign attribution in shared spaces, using sensor fusion to maintain individual identity and vital sign data streams even when subjects interact or partially occlude one another [2]. Such comprehensive approaches represent the cutting edge of contactless monitoring, though they also introduce greater complexity in system integration, calibration, and data fusion.

5.3. Data Fusion Approaches

In the following subsubsections, we provide a concise overview of the primary data fusion strategies employed in contactless vital sign monitoring. Each approach—signal-level, feature-level, and decision-level fusion—offers unique advantages and challenges in integrating information from multiple sources or modalities. By outlining these methods, we aim to clarify their roles, highlight their relevance to multi-modal systems, and set the stage for a deeper understanding of how fusion techniques can enhance the accuracy and robustness of vital sign measurement.

5.3.1. Signal-Level Fusion

Signal-level fusion, also known as early integration, combines raw signals or minimally processed features from multiple modalities at the earliest processing stage. This approach faces significant synchronization challenges due to different sampling rates and latencies across heterogeneous sensors, requiring sophisticated temporal alignment solutions [25]. Researchers have developed weighted fusion schemes based on signal quality indices that dynamically adjust the contribution of each modality according to confidence metrics derived from signal characteristics [108]. Such systems often require complex signal processing pipelines to handle heterogeneous data streams with different noise characteristics, dimensionality, and temporal properties [109]. While early fusion potentially preserves more information than later integration approaches, it typically requires more sophisticated preprocessing to harmonize diverse signals and can be more vulnerable to modality-specific noise propagating throughout the system.

5.3.2. Feature-Level Fusion

Feature-level fusion, or intermediate integration, combines extracted features from multiple modalities before final decision-making, striking a balance between information preservation and noise isolation. Early approaches relied on straightforward handcrafted feature concatenation, simply combining feature vectors from different modalities into a unified representation for subsequent processing [7]. As the field has advanced, more sophisticated feature selection methods have emerged to identify optimal complementary information across modalities, reducing redundancy and focusing on the most informative aspects of each signal source [110]. Recent work has incorporated attention mechanisms for adaptive feature weighting, allowing systems to dynamically emphasize more reliable or informative modalities based on current conditions and learning from previous examples [111]. Feature-level fusion generally offers greater flexibility than signal-level approaches while retaining more cross-modal information than decision-level methods, making it a popular choice for many multi-modal systems.

5.3.3. Decision-Level Fusion

Decision-level fusion, or late integration, combines independently derived decisions or high-level interpretations from each modality, emphasizing modularity and robustness to modality-specific failures. Weighted voting and confidence-based schemes form the foundation of many decision fusion approaches, combining individual modality outputs according to reliability metrics or historical performance [112]. More sophisticated systems employ Bayesian fusion frameworks for principled uncertainty handling, incorporating probabilistic representations of confidence from each modality to derive optimal combined decisions [113]. Recent advancements include adaptive fusion based on contextual factors, where the integration strategy itself adjusts according to environmental conditions, subject characteristics, or application requirements [28]. Decision-level fusion offers strong fault tolerance, as the failure of one modality does not necessarily compromise the entire system, and allows for modular development where individual sensing approaches can be improved or replaced independently of the overall architecture.

5.3.4. Deep Learning Fusion

Neural network approaches have revolutionized multi-modal fusion by enabling end-to-end learning of optimal integration strategies directly from data, without requiring explicit design of fusion rules. Cross-modal attention mechanisms represent a particularly powerful approach, allowing networks to dynamically focus on the most informative aspects of each modality and their relationships [111]. Multi-branch architectures with shared and modality-specific components have emerged as a dominant paradigm, where separate pathways process each modality before combination at various levels, enabling the network to learn both modality-specific features and cross-modal relationships [28]. More recently, transformer-based fusion approaches have demonstrated exceptional capability for temporal sequence integration, leveraging self-attention mechanisms to capture long-range dependencies within and across modalities [114]. These deep learning approaches typically outperform traditional fusion methods when sufficient training data is available, as they can discover optimal integration strategies that might not be obvious to human designers and can adapt to the specific characteristics of the input modalities.

6. Multi-Task Approaches

Multi-task learning (MTL) in vital-sign monitoring aims to simultaneously extract multiple vital sign parameters through shared representations, capitalizing on the inherent relationships between different vital signs. This approach is particularly well-suited to contactless monitoring because vital sign parameters exhibit natural interdependencies that create synergistic opportunities—for example, respiration patterns directly affect heart rate variability through well-understood cardiorespiratory mechanisms [115]. Additionally, common noise sources such as motion artifacts or environmental variations tend to affect multiple parameters in similar ways, enabling joint filtering approaches that leverage this shared structure [47]. From an implementation perspective, shared computational infrastructure significantly reduces processing requirements compared to running separate models for each parameter, making multi-task systems more feasible for edge deployment and real-time applications [6]. The theoretical foundation for this approach stems from Caruana’s [116] seminal work on MTL, which demonstrated how learning-related tasks can simultaneously improve generalization by using the domain information contained in the training signals of related tasks as an inductive bias. The practical benefits of this approach have been demonstrated in recent research by Chen et al. [14], which showed a 15–25% improvement in accuracy when jointly estimating heart rate, respiration, and blood pressure compared to separate models, underscoring the substantial advantages of multi-task approaches for comprehensive vital sign monitoring.

6.1. Technical Approaches to Multi-Task Learning

6.1.1. Shared Representation Architectures

In multi-task learning for contactless vital sign measurement, shared representation architectures are designed to extract common features from input data—such as video, thermal, or radar signals—that are relevant to multiple physiological parameters (e.g., heart rate, respiratory rate, blood oxygen saturation). By leveraging a shared backbone (such as a convolutional neural network or transformer encoder), these architectures enable the model to learn generalized patterns that underlie different vital signs, while reducing redundancy and improving data efficiency. For example, subtle skin color changes in facial videos may contain information pertinent to both heart rate and respiratory rate, and a shared feature extractor can capture these cues for downstream task-specific branches [28]. This approach not only streamlines the learning process but also facilitates the integration of multi-modal signals, ultimately enhancing the robustness and accuracy of vital sign estimation [117].

6.1.2. Task Relationship Modeling

Task relationship modeling focuses on explicitly capturing the dependencies and interactions between different vital sign measurement tasks within a multi-task learning framework. In the context of contactless monitoring, certain physiological parameters are inherently correlated—for instance, changes in respiratory rate can influence heart rate variability. By modeling these relationships, either through structured loss functions, attention mechanisms, or graph-based approaches, the learning algorithm can exploit inter-task synergies to improve overall performance [22,117]. For example, joint modeling of heart rate and respiratory rate can help disambiguate signal artifacts that affect both measurements, leading to more reliable and consistent vital sign estimates. Task relationship modeling is thus crucial for maximizing the benefits of multi-task learning in complex, real-world monitoring scenarios [28].

6.1.3. Multi-Task Optimization Strategies

Balanced learning across multiple vital sign parameters requires specialized optimization techniques that address the challenges of training with multiple, potentially competing objectives. Gradient normalization techniques help prevent dominant tasks from overwhelming the learning process, ensuring that all parameters receive appropriate attention during optimization [7]. Uncertainty-based task weighting represents a more adaptive approach, dynamically adjusting the contribution of each task to the overall loss based on measurement confidence, focusing learning on more uncertain parameters [111,118]. Advanced systems implement adaptive task scheduling based on learning dynamics, varying the emphasis on different parameters throughout the training process to improve overall convergence and performance [28]. These optimization strategies address the fundamental challenge of multi-objective learning in vital sign monitoring, ensuring balanced performance across diverse parameters with different measurement characteristics and difficulties.

6.2. Application-Specific Multi-Task Approaches

6.2.1. Clinical Monitoring Systems

Hospital and clinical applications of multi-task vital sign monitoring address specialized requirements for medical decision support while operating within healthcare constraints. Intensive care unit monitoring systems incorporate privacy preservation features to address the sensitive nature of continuous monitoring in vulnerable populations, using on-device processing and anonymization techniques to protect patient dignity and confidentiality [9]. Neonatal monitoring represents a particularly valuable application domain, where the ability to track multiple vital parameters without physical contact addresses the special needs of fragile infants who may be harmed by traditional sensing approaches [119]. Triage and rapid assessment systems in emergency settings leverage multi-task monitoring to quickly evaluate patient status across multiple vital sign dimensions, supporting clinical decision-making in time-critical scenarios [6]. These clinical applications benefit from the comprehensive vital sign assessment enabled by multi-task approaches while addressing the unique requirements and constraints of healthcare environments.

6.2.2. Home Health Monitoring

Consumer-oriented multi-task monitoring systems for home use balance comprehensive assessment with usability and integration into daily life. Sleep quality assessment through multi-parameter analysis represents one of the most successful applications, combining heart rate, respiration, movement, and sometimes temperature monitoring to evaluate sleep stages and disturbances without the complexity and discomfort of polysomnography [120]. Chronic condition management systems with trend analysis capabilities enable long-term tracking of multiple vital sign parameters, helping patients and healthcare providers monitor disease progression and treatment effectiveness over extended periods [113]. Systems targeting elderly monitoring often combine vital sign assessment with fall detection and activity monitoring, providing comprehensive safety and health tracking for aging-in-place scenarios [107]. These home-focused applications emphasize unobtrusiveness, reliability, and actionable insights from multiple vital sign parameters, making multi-task approaches particularly valuable for sustainable long-term monitoring.

7. Challenges and Limitations

7.1. Technical Challenges

7.1.1. Signal Quality and Environmental Variability

Contactless vital sign monitoring methods face significant challenges related to signal quality and environmental robustness. Lighting variation substantially affects vision-based approaches, with changes in illumination intensity, color temperature, and direction potentially degrading signal quality or introducing artifacts [27]. Motion artifacts represent another major challenge, with even subtle subject movements potentially obscuring the minute vital sign signals of interest, necessitating sophisticated compensation strategies ranging from region-of-interest tracking to deep learning-based motion separation [30]. Environmental interference similarly affects radar and RF methods, with nearby electronic devices, structural elements, and other individuals potentially distorting the electromagnetic fields used for monitoring [11]. Cross-subject variability introduces additional complexity, as factors such as skin tone, facial structure, body composition, and underlying vital sign differences can significantly affect measurement accuracy, often requiring personalization or adaptation mechanisms to achieve reliable performance across diverse populations [6]. These signal quality and variability challenges highlight the inherent difficulty of extracting subtle vital sign signals through non-contact means in uncontrolled real-world environments.

7.1.2. Sensor Fusion Complexities

Multi-modal systems, while offering enhanced robustness and parameter coverage, introduce significant integration challenges that must be addressed for effective operation. Synchronization issues across heterogeneous sensors represent a fundamental challenge, as different sensing modalities typically operate at different sampling rates, exhibit varying latencies, and may lack precise temporal alignment mechanisms [25]. Calibration requirements and drift compensation further complicate multi-modal integration, as sensors may require initial calibration and periodic recalibration to maintain accuracy, with different modalities potentially exhibiting different drift characteristics over time [108]. Conflicting measurements between modalities necessitate resolution strategies to determine the most reliable value, ranging from simple confidence-weighted averaging to sophisticated Bayesian fusion methods that model the error characteristics of each modality [109]. Additionally, the computational complexity of real-time multi-modal processing presents practical implementation challenges, particularly for edge devices with limited processing capabilities, requiring efficient algorithms and sometimes specialized hardware for deployment [28]. These fusion complexities highlight the technical tradeoffs involved in multi-modal monitoring, where the potential performance benefits must be balanced against increased system complexity.

7.1.3. Privacy-Preserving Monitoring

Protecting sensitive information while enabling effective monitoring represents a fundamental challenge requiring both technical and governance solutions. Federated learning approaches enable model development without raw data sharing, allowing systems to learn from diverse datasets while keeping sensitive vital sign data on local devices [2]. On-device processing architectures minimize privacy risks by extracting relevant vital sign parameters locally and transmitting only derived metrics rather than raw sensor data [117]. Differential privacy techniques provide mathematical guarantees about information leakage, enabling quantifiable privacy protection for vital sign data even when some information must be shared [121]. Synthetic data approaches for development and testing offer another promising direction, enabling system development and validation without requiring extensive real subject data [3]. These privacy-preserving approaches collectively address growing concerns about data protection while enabling the benefits of contactless monitoring, representing an essential foundation for ethical and acceptable deployment of these technologies.

7.2. Clinical and Practical Limitations

7.2.1. Accuracy and Reliability Concerns

Medical applications of contactless monitoring demand high-performance standards that remain challenging to achieve consistently in real-world settings. Rigorous comparison with medical-grade contact devices is essential for clinical adoption but reveals performance gaps in many current systems, particularly under challenging conditions or for certain population groups [119]. Clinical validation protocols present their own challenges, as traditional medical device validation approaches may not be directly applicable to contactless technologies with different error characteristics and operational constraints, requiring new methodologies and standards [122]. Regulatory considerations for medical applications impose additional requirements for safety, efficacy, and reliability documentation, creating significant hurdles for transitioning research prototypes to approved medical devices [6]. Failure mode analysis and graceful degradation strategies are critical for clinical applications but often receive insufficient attention, requiring systems to recognize when measurements become unreliable and either adapt or provide appropriate uncertainty indicators rather than producing potentially misleading values [109]. These accuracy and reliability concerns highlight the gap between current research capabilities and the standards required for widespread clinical adoption of Contactless monitoring technologies.

7.2.2. Real-World Deployment Issues

Practical implementation of contactless monitoring systems faces numerous challenges beyond technical performance metrics. System calibration requirements in uncontrolled environments can significantly impact usability and reliability, as many current systems require initial setup procedures or environmental adaptations that may be impractical in real-world settings [108]. User acceptance and compliance considerations are equally critical, as systems that are perceived as intrusive, difficult to use, or unreliable will face adoption challenges regardless of their technical capabilities [120]. Integration with existing clinical workflows represents a particular challenge for healthcare applications, requiring systems to complement rather than disrupt established practices and to provide information in formats compatible with existing decision-making processes [9]. Cost-benefit analysis for healthcare implementation must also consider not only the direct costs of the technology but also training requirements, maintenance needs, and potential impacts on care efficiency and outcomes, creating a high bar for demonstrating value sufficient to justify adoption [6]. These deployment challenges highlight the importance of considering practical implementation factors alongside technical performance when developing contactless monitoring solutions.

8. Conclusions

Contactless vital-sign monitoring has evolved from single-modality, single-task approaches to sophisticated multi-modal, multi-task systems capable of comprehensive vital sign assessment. This review has synthesized research across diverse sensing technologies, fusion strategies, and application domains, highlighting both significant progress and persistent challenges.
The integration of complementary sensing modalities has consistently demonstrated improved robustness to environmental variations and motion artifacts, while multi-task learning approaches leverage inherent vital sign parameter relationships to enhance accuracy and efficiency. These advances enable applications ranging from clinical care to home health monitoring and affective computing.
Nevertheless, significant challenges remain in ensuring accuracy across diverse populations and environments, addressing privacy concerns, and establishing regulatory and validation frameworks appropriate for these novel technologies. Future research directions include exploring emerging sensing modalities, developing advanced fusion architectures, implementing edge computing solutions, and creating standardized evaluation frameworks.
As healthcare increasingly emphasizes preventive, personalized, and accessible approaches, contactless vital sign monitoring represents a critical enabling technology. The multi-modal, multi-task paradigm reviewed in this work offers a promising framework for developing systems that are simultaneously comprehensive, robust, and minimally intrusive.

Author Contributions

Investigation, A.H.; Writing—original draft, A.H.; Supervision, B.Y. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by SFI Norwegian Center for Cybersecurity in Critical Sectors (NORCICS) grant number 310105.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare that they have no conflict of interest related to this work.

References

  1. McDuff, D.J.; Blackford, E.B.; Estepp, J.R. The impact of video compression on remote cardiac pulse measurement using imaging photoplethysmography. In Proceedings of the 2017 12th IEEE International Conference on Automatic Face & Gesture Recognition (FG 2017), Washington, DC, USA, 30 May–3 June 2017; pp. 63–70. [Google Scholar]
  2. Wang, Y.; Yang, J.; Chen, H.; Liu, J.; Chen, Z.; Yin, Y.; Yao, R.; Xie, Y. E-mmWave: Edge-assisted millimeter wave sensing on commodity mobile devices. IEEE Internet Things J. 2022, 9, 13675–13688. [Google Scholar]
  3. Zhao, F.; Li, M.; Qian, Y.; Tsien, J.Z. Remote measurements of heart and respiration rates for telemedicine. PloS ONE 2021, 8, e71384. [Google Scholar] [CrossRef]
  4. Poh, M.Z.; McDuff, D.J.; Picard, R.W. Non-contact, automated cardiac pulse measurements using video imaging and blind source separation. Opt. Express 2010, 18, 10762–10774. [Google Scholar] [CrossRef]
  5. Verkruysse, W.; Svaasand, L.O.; Nelson, J.S. Remote plethysmographic imaging using ambient light. Opt. Express 2008, 16, 21434–21445. [Google Scholar] [CrossRef]
  6. Liu, Y.; Wang, H.; Zhao, W.; Zhang, M.; Qin, H.; Xie, Y. Towards long-term multi-modal physiological assessment in the wild. Proc. ACM Interact. Mob. Wearable Ubiquitous Technol. 2021, 5, 1–27. [Google Scholar]
  7. Chen, W.; McDuff, D. Deepphys: Video-based physiological measurement using convolutional attention networks. In Proceedings of the European Conference on Computer Vision (ECCV), Milan, Italy, 29 September–4 October 2018; pp. 349–365. [Google Scholar]
  8. Nandakumar, R.; Gollakota, S.; Watson, N. Contactless sleep apnea detection on smartphones. In Proceedings of the 13th Annual International Conference on Mobile Systems, Applications, and Services, Florence, Italy, 18–22 May 2015; pp. 45–57. [Google Scholar]
  9. Wang, X.; Yang, C.; Mao, S. PhaseBeat: Exploiting CSI phase data for vital sign monitoring with commodity WiFi devices. In Proceedings of the 2017 IEEE 37th International Conference on Distributed Computing Systems (ICDCS), Atlanta, GA, USA, 5–8 June 2017; pp. 1230–1239. [Google Scholar]
  10. Cardillo, E.; Li, C. Portable Microwave and mmWave Radars for Contactless Healthcare; River Publishers: Gistrup, Denmark, 2025; pp. 1–178. [Google Scholar]
  11. Yang, Z.; Pathak, P.H.; Zeng, Y.; Liran, X.; Mohapatra, P. Monitoring vital signs using millimeter wave. In Proceedings of the 17th Conference on Embedded Networked Sensor Systems, Paderborn, Germany, 5–8 July 2016; pp. 1–12. [Google Scholar]
  12. Cho, Y.; Julier, S.J.; Bianchi-Berthouze, N. Instant stress: Detection of perceived mental stress through smartphone photoplethysmography and thermal imaging. JMIR Ment. Health 2017, 4, e29. [Google Scholar] [CrossRef] [PubMed]
  13. Pavlidis, I.; Levine, J. Thermal image analysis for polygraph testing. IEEE Eng. Med. Biol. Mag. 2002, 21, 56–64. [Google Scholar] [CrossRef] [PubMed]
  14. Chen, X.; Liu, A.; Cheng, S.; Liu, C.; Liu, Y.; Wang, Z.J. TS-CAN: Task-specific attention for simultaneous multi-parameter remote physiological measurement. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 43, 2775–2789. [Google Scholar]
  15. McDuff, D.J.; Estepp, J.R.; Piasecki, A.M.; Blackford, E.B. A survey of remote optical photoplethysmographic imaging methods. In Proceedings of the 2015 37th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), Milan, Italy, 25–29 August 2015; pp. 6398–6404. [Google Scholar]
  16. Massaroni, C.; Presti, D.L.; Formica, D.; Silvestri, S.; Schena, E. Non-contact monitoring of breathing pattern and respiratory rate via RGB signal measurement. Sensors 2019, 19, 2758. [Google Scholar] [CrossRef]
  17. Clifford, G.D.; Azuaje, F.; McSharry, P. Advanced Methods and Tools for ECG Data Analysis; Artech House, Inc.: Norwood, MA, USA, 2006. [Google Scholar]
  18. Allen, J. Photoplethysmography and its application in clinical physiological measurement. Physiol. Meas. 2007, 28, R1. [Google Scholar] [CrossRef]
  19. Massaroni, C.; Nicolò, A.; Presti, D.L.; Sacchetti, M.; Silvestri, S.; Schena, E. Contact-based methods for measuring respiratory rate. Sensors 2018, 19, 908. [Google Scholar] [CrossRef]
  20. Boucsein, W. Electrodermal Activity; Springer Science & Business Media: Berlin/Heidelberg, Germany, 2012. [Google Scholar]
  21. Kushida, C.A.; Littner, M.R.; Morgenthaler, T.; Alessi, C.A.; Bailey, D.; Coleman, J., Jr.; Friedman, L.; Hirshkowitz, M.; Kapen, S.; Kramer, M.; et al. Practice parameters for the indications for polysomnography and related procedures: An update for 2005. Sleep 2005, 28, 499–521. [Google Scholar] [CrossRef]
  22. Xu, S.; Jayaraman, L.; Rogers, J.A. Skin sensors are the future of health care. Nature 2020, 571, 319–321. [Google Scholar] [CrossRef]
  23. Wu, H.-Y.; Rubinstein, M.; Shih, E.; Guttag, J.; Durand, F.; Freeman, W. Eulerian video magnification for revealing subtle changes in the world. ACM Trans. Graph. 2012, 31, 65. [Google Scholar] [CrossRef]
  24. Li, C.; Cummings, J.; Lam, J.; Graves, E.; Wu, W. Radar remote monitoring of vital signs. IEEE Microw. Mag. 2009, 10, 47–56. [Google Scholar] [CrossRef]
  25. Sun, Y.; Thakor, N. Photoplethysmography revisited: From contact to noncontact, from point to imaging. IEEE Trans. Biomed. Eng. 2016, 63, 463–477. [Google Scholar] [CrossRef] [PubMed]
  26. Haan, G.D.; Jeanne, V. Robust pulse rate from chrominance-based rPPG. IEEE Trans. Biomed. Eng. 2013, 60, 2878–2886. [Google Scholar] [CrossRef] [PubMed]
  27. Wang, W.; den Brinker, A.C.; Stuijk, S.; de Haan, G. Algorithmic principles of remote PPG. IEEE Trans. Biomed. Eng. 2017, 64, 1479–1491. [Google Scholar] [CrossRef]
  28. Liu, X.; Fromm, J.; Patel, S.; McDuff, D. Multi-task temporal shift attention networks for on-device contactless vitals measurement. Adv. Neural Inf. Process. Syst. 2020, 33, 19400–19411. [Google Scholar]
  29. Nowara, E.M.; McDuff, D.; Veeraraghavan, A. A meta-analysis of the impact of skin tone and gender on non-contact photoplethysmography measurements. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 1215–1225. [Google Scholar]
  30. van Gastel, M.; Stuijk, S.; de Haan, G. Motion robust remote-PPG in infrared. IEEE Trans. Biomed. Eng. 2019, 62, 1425–1433. [Google Scholar] [CrossRef]
  31. Addison, P.S.; Jacquel, D.; Foo, D.M.; Borg, U.R. Video-based heart rate monitoring across a range of skin pigmentations during an acute hypoxic challenge. J. Clin. Monit. Comput. 2020, 34, 795–802. [Google Scholar] [CrossRef]
  32. Shao, D.; Yang, Y.; Liu, C.; Tsow, F.; Yu, H.; Tao, N. Noncontact monitoring breathing pattern, exhalation flow rate and pulse transit time. IEEE Trans. Biomed. Eng. 2016, 61, 2760–2767. [Google Scholar] [CrossRef]
  33. Cobos-Torres, J.C.; Abderrahim, M.; Martínez-Orgado, J. Non-contact, simple neonatal monitoring by photoplethysmography. Sensors 2020, 20, 4362. [Google Scholar] [CrossRef]
  34. Hu, W.; Zhao, Z.; Wang, Y.; Zhang, H.; Lin, F. Noncontact accurate measurement of cardiopulmonary activity using a compact quadrature Doppler radar sensor. IEEE Trans. Biomed. Eng. 2014, 61, 725–735. [Google Scholar] [CrossRef] [PubMed]
  35. Nosrati, M.; Tavassolian, N. High-accuracy heart rate variability monitoring using Doppler radar based on Gaussian pulse train modeling and FTPR algorithm. IEEE Trans. Microw. Theory Tech. 2018, 66, 556–567. [Google Scholar] [CrossRef]
  36. Alizadeh, M.; Abedi, G.; Kaufmann, D.; Xu, Y.; Boric-Lubecke, O. Sensing multiple subjects using FMCW MIMO radar in dopder domain. IEEE Sens. J. 2019, 19, 2308–2316. [Google Scholar]
  37. Lazaro, A.; Girbau, D.; Villarino, R. Analysis of vital signs monitoring using an IR-UWB radar. Prog. Electromagn. Res. 2012, 100, 265–284. [Google Scholar] [CrossRef]
  38. Javaid, A.Q.; Noble, C.M.; Rosenberg, R.; Weitnauer, M.A. Towards sleep apnea screening with an under-the-mattress IR-UWB radar using machine learning. In Proceedings of the 2015 IEEE 14th International Conference on Machine Learning and Applications (ICMLA), Miami, FL, USA, 9–11 December 2015. [Google Scholar]
  39. Leem, S.K.; Khan, F.; Cho, S.H. Vital sign monitoring and mobile phone usage detection using IR-UWB radar for intended use in car crash prevention. Sensors 2020, 20, 1240. [Google Scholar] [CrossRef]
  40. Garbey, M.; Sun, N.; Merla, A.; Pavlidis, I. Contact-free measurement of cardiac pulse based on the analysis of thermal imagery. IEEE Trans. Biomed. Eng. 2007, 54, 1418–1426. [Google Scholar] [CrossRef]
  41. Cho, Y.; Bianchi-Berthouze, N.; Oliveira, M.; Holloway, C.; Julier, S. Nose heat: Exploring stress-induced nasal thermal variability through mobile thermal imaging. In Proceedings of the 2019 IEEE International Conference on Pervasive Computing and Communications (PerCom), Cambridge, UK, 3–6 September 2019; pp. 1–10. [Google Scholar]
  42. Abbas, A.K.; Heimann, K.; Jergus, K.; Orlikowsky, T.; Leonhardt, S. Neonatal non-contact respiratory monitoring based on real-time infrared thermography. Biomed. Eng. Online 2011, 10, 93. [Google Scholar] [CrossRef]
  43. Pereira, C.B.; Yu, X.; Czaplik, M.; Rossaint, R.; Blazek, V.; Leonhardt, S. Remote monitoring of breathing dynamics using infrared thermography. Biomed. Opt. Express 2015, 6, 4378–4394. [Google Scholar] [CrossRef]
  44. Wang, X.; Huang, R.; Mao, S. SonarBeat: Sonar phase for breathing beat monitoring with smartphones. In Proceedings of the 2016 25th International Conference on Computer Communication and Networks (ICCCN), Waikoloa, HI, USA, 1–4 August 2016; pp. 1–8. [Google Scholar]
  45. Zhao, M.; Tian, Y.; Zhao, H.; Alsheikh, M.A.; Li, T.; Hristov, R.; Kabelac, Z.; Katabi, D.; Torralba, A. RF-based 3D skeletons. In Proceedings of the 2018 Conference of the ACM Special Interest Group on Data Communication, Budapest, Hungary, 20–25 August 2018; pp. 267–281. [Google Scholar]
  46. Yang, X.; Fan, D.; Ren, A.; Zhao, N.; Alam, M. 5G-based user-centric sensing at C-band. IEEE Trans. Ind. Inform. 2018, 15, 3040–3047. [Google Scholar] [CrossRef]
  47. Wang, Y.; Wu, K.; Ni, L.M. WiFall: Device-free fall detection by wireless networks. IEEE Trans. Mob. Comput. 2020, 16, 581–594. [Google Scholar] [CrossRef]
  48. Takano, M.; Ohta, A. Heart rate measurement based on a time-lapse image. Med. Eng. Phys. 2007, 29, 853–857. [Google Scholar] [CrossRef]
  49. Chi, Y.M.; Jung, T.P.; Cauwenberghs, G. Dry-contact and noncontact biopotential electrodes: Methodological review. IEEE Rev. Biomed. Eng. 2010, 3, 106–119. [Google Scholar] [CrossRef] [PubMed]
  50. Majumder, S.; Aghayi, E.; Noferesti, M.; Memarzadeh-Tehran, H.; Mondal, T.; Pang, Z.; Deen, M.J. Smart homes for elderly healthcare—Recent advances and research challenges. Sensors 2017, 17, 2496. [Google Scholar] [CrossRef] [PubMed]
  51. Choi, J.-H.; Kang, K.-B.; Kim, K.-T. Fusion-Vital: Video-RF Fusion Transformer for Advanced Remote Physiological Measurement. Proc. AAAI Conf. Artif. Intell. 2024, 38, 1344–1352. [Google Scholar] [CrossRef]
  52. Wu, B.-F.; Wu, B.-J.; Tsai, B.-R.; Hsu, C.-P. A facial-image-based blood pressure measurement system without calibration. IEEE Trans. Instrum. Meas. 2022, 71, 5009413. [Google Scholar] [CrossRef]
  53. Bukum, K.; Savur, C.; Tsouri, G.R. Deep Learning Classifier for Advancing Video Monitoring of Atrial Fibrillation. In Proceedings of the 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), New Orleans, LA, USA, 19–20 June 2022; pp. 2210–2218. [Google Scholar]
  54. Špetlík, R.; Franc, V.; Matas, J. Visual heart rate estimation with convolutional neural network. In Proceedings of the British Machine Vision Conference, Newcastle, UK, 3–6 September 2018; pp. 3–6. [Google Scholar]
  55. Yu, Z.; Peng, W.; Li, X.; Hong, X.; Zhao, G. Remote heart rate measurement from highly compressed facial videos: An end-to-end deep learning solution with video enhancement. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 151–160. [Google Scholar]
  56. Qiu, Y.; Liu, Y.; Arteaga-Falconi, J.; Dong, H.; Saddik, A.E. EVM-CNN: Real-time contactless heart rate estimation from facial video. IEEE Trans. Multimed. 2018, 21, 1778–1787. [Google Scholar] [CrossRef]
  57. Luguev, T.; Seuß, D.; Garbas, J.-U. Deep learning based affective sensing with remote photoplethysmography. In Proceedings of the 2020 54th Annual Conference on Information Sciences and Systems (CISS), Princeton, NJ, USA, 18–20 March 2020; pp. 1–4. [Google Scholar]
  58. Zhan, Q.; Wang, W.; De Haan, G. Analysis of CNN-based remote-PPG to understand limitations and sensitivities. Biomed. Opt. Express 2020, 11, 1268–1283. [Google Scholar] [CrossRef]
  59. Yu, Z.; Li, X.; Zhao, G. Remote photoplethysmograph signal measurement from facial videos using spatio-temporal networks. arXiv 2019, arXiv:1905.02419. [Google Scholar] [CrossRef]
  60. Du, G.; Zhang, L.; Su, K.; Wang, X.; Teng, S.; Liu, P.X. A multimodal fusion fatigue driving detection method based on heart rate and PERCLOS. IEEE Trans. Intell. Transp. Syst. 2022, 23, 21810–21820. [Google Scholar] [CrossRef]
  61. Suriani, N.S.; Shahdan, N.S.; Sahar, N.M.; Taujuddin, N.S.A.M. Non-contact facial based vital sign estimation using convolutional neural network approach. Int. J. Adv. Comput. Sci. Appl. 2022, 13, 386–393. [Google Scholar] [CrossRef]
  62. Lorato, I.; Stuijk, S.; Meftah, M.; Kommers, D.; Andriessen, P.; van Pul, C.; de Haan, G. Towards continuous camera-based respiration monitoring in infants. Sensors 2021, 21, 2268. [Google Scholar] [CrossRef]
  63. Niu, X.; Han, H.; Shan, S.; Chen, X. VIPL-HR: A Multi-modal Database for Pulse Estimation from Less-Constrained Face Video. In Computer Vision—ACCV 2018, Proceedings of the 14th Asian Conference on Computer Vision, Perth, Australia, 2–6 December 2018; Springer: Cham, Switzerland, 2019; pp. 562–576. [Google Scholar]
  64. Niu, X.; Shan, S.; Han, H.; Chen, X. RhythmNet: End-to-End Heart Rate Estimation from Face via Spatial-Temporal Representation. IEEE Trans. Image Process. 2020, 29, 2409–2423. [Google Scholar] [CrossRef]
  65. Huang, B.; Chang, C.-M.; Lin, C.-L.; Chen, W.; Juang, C.-F.; Wu, X. Visual heart rate estimation from facial video based on CNN. In Proceedings of the 2020 15th IEEE Conference on Industrial Electronics and Applications (ICIEA), Kristiansand, Norway, 9–13 November 2020; pp. 1658–1662. [Google Scholar]
  66. Song, R.; Zhang, S.; Li, C.; Zhang, Y.; Cheng, J.; Chen, X. Heart rate estimation from facial videos using a spatiotemporal representation with convolutional neural networks. IEEE Trans. Instrum. Meas. 2020, 69, 7411–7421. [Google Scholar] [CrossRef]
  67. Botina-Monsalve, D.; Benezeth, Y.; Macwan, R.; Pierrart, P.; Parra, F.; Nakamura, K.; Gomez, R.; Miteran, J. Long short-term memory deep-filter in remote photoplethysmography. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, Seattle, WA, USA, 14–19 June 2020; pp. 306–307. [Google Scholar]
  68. Huang, B.; Lin, C.L.; Chen, W.; Juang, C.F.; Wu, X. A novel one-stage framework for visual pulse rate estimation using deep neural networks. Biomed. Signal Process. Control 2021, 66, 102387. [Google Scholar] [CrossRef]
  69. Gao, H.; Wu, X.; Geng, J.; Lv, Y. Remote heart rate estimation by signal quality attention network. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 21–24 June 2022; pp. 2122–2129. [Google Scholar]
  70. Napolean, Y.; Marwade, A.; Tomen, N.; Alkemade, P.; Eijsvogels, T.; van Gemert, J.C. Heart rate estimation in intense exercise videos. In Proceedings of the 2022 IEEE International Conference on Image Processing (ICIP), Bordeaux, France, 16–19 October 2022; pp. 3933–3937. [Google Scholar]
  71. Wu, Y.C.; Chiu, L.W.; Wu, B.F.; Lin, L.L.C.; Ho, T.H.; Chung, M.L.; Wu, S.F. Motion robust remote photoplethysmography measurement during exercise for contactless physical activity intensity detection. IEEE Trans. Instrum. Meas. 2023, 72, 2508614. [Google Scholar] [CrossRef]
  72. Gua, Z.; Zhang, K.; Liu, X.; Xiao, J. Dual-transformer and CNN hybrid architecture for remote physiological measurement. Comput. Vis. Image Underst. 2023, 233, 103679. [Google Scholar]
  73. Othman, W.; Kashevnik, A.; Ali, A.; Shilov, N.; Ryumin, D. Remote Heart Rate Estimation Based on Transformer with Multi-Skip Connection Decoder: Method and Evaluation in the Wild. Sensors 2024, 24, 775. [Google Scholar] [CrossRef] [PubMed]
  74. Xu, C.; Fu, Z.; Wu, W.; Zhang, J.; Liu, M.; Gao, L. Association between High-Sensitivity C-Reactive Protein and Blood Pressure Variability in Subacute Stage of Ischemic Stroke. Brain Sci. 2023, 13, 998. [Google Scholar] [CrossRef]
  75. Wang, R.-X.; Sun, H.-M.; Hao, R.-R.; Pan, A.; Jia, R.-S. TransPhys: Transformer-based unsupervised contrastive learning for remote heart rate measurement. Biomed. Signal Process. Control 2023, 86, 105058. [Google Scholar] [CrossRef]
  76. He, H.; Yuan, Y.; Chen, Y.-C.; Cao, P.; Katabi, D. Contactless Oxygen Monitoring with Radio Waves and Gated Transformer. In Proceedings of the Machine Learning for Healthcare Conference, New York, NY, USA, 11–12 August 2023; pp. 248–265. [Google Scholar]
  77. Xu, C.; Chen, T.; Li, H.; Gherardi, A.; Weng, M.; Li, Z.; Xu, W. Hearing heartbeat from voice: Towards next generation voice-user interfaces with cardiac sensing functions. In Proceedings of the 20th ACM Conference on Embedded Networked Sensor Systems, Boston, MA, USA, 7–9 November 2022; pp. 149–163. [Google Scholar]
  78. Marchi, E.; Bahle, G.; Monteleone, G.; Thiem, H.; Lechner, M.; Schuller, B. Robust and privacy-preserving audio-based vital signs monitoring using a smart speaker. In Proceedings of the ICASSP 2019—2019 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Brighton, UK, 12–17 May 2019; pp. 6361–6365. [Google Scholar]
  79. Deshpande, S.; Nguyen, T.; Postolache, S.; Rao, R.K. Coupled breathing and heart rate monitoring from speech signals using transformer networks. In Proceedings of the 2020 42nd Annual International Conference of the IEEE Engineering in Medicine & Biology Society (EMBC), Montreal, QC, Canada, 20–24 July 2020; pp. 5874–5878. [Google Scholar]
  80. Kim, H.G.; Cheon, Y.J.; Ye, S.; Kwon, D.S. Non-invasive measurement of heart rate variability using audio signals from smartphone microphones. Sensors 2021, 21, 1427. [Google Scholar]
  81. Pimentel, B.D.; Kaiser, A.D.; Johnson, K.; Lieberman, A.; Arzeno, N.M. Audio-based deep learning models for detecting sleep apnea: A systematic review. Sleep Med. Rev. 2022, 65, 101644. [Google Scholar]
  82. Amiriparian, S.; Gerczuk, M.; Ottl, S.; Cummins, N.; Freitag, M.; Schuller, B. Towards deep learning-based respiratory sound analysis for COVID-19 detection. In Proceedings of the Annual Conference of the International Speech Communication Association (Interspeech), Incheon, Republic of Korea, 18–22 September 2022; pp. 798–802. [Google Scholar]
  83. Chen, L.; Lu, X.; Wang, J.; Yan, Y. Audio-based respiratory disease classification using convolutional neural networks with attention mechanisms. IEEE J. Biomed. Health Inform. 2023, 27, 2217–2228. [Google Scholar]
  84. Rahman, M.A.; Hussain, T.; Saha, S.; Hossain, M.S. BabyCare: Transformer-based infant cry analysis for remote vital sign monitoring and wellbeing assessment. IEEE Trans. Affect. Comput. 2024, 15, 1018–1031. [Google Scholar]
  85. Soleymani, M.; Lichtenauer, J.; Pun, T.; Pantic, M. A multimodal database for affect recognition and implicit tagging. IEEE Trans. Affect. Comput. 2012, 3, 42–55. [Google Scholar] [CrossRef]
  86. Stricker, R.; Müller, S.; Gross, H.-M. Non-contact video-based pulse rate measurement on a mobile service robot. In Proceedings of the 2014 IEEE 23rd International Symposium on Robot and Human Interactive Communication, Edinburgh, UK, 25–29 August 2014; pp. 1056–1062. [Google Scholar]
  87. Heusch, G.; Anjos, A.; Marcel, S. A reproducible study on remote heart rate measurement. arXiv 2017, arXiv:1709.00962. [Google Scholar] [CrossRef]
  88. Zhu, B.; Zaech, P.; Elhoseiny, M. Vision-based assessment of parkinson’s disease and essential tremor. arXiv 2017, arXiv:1709.03923. [Google Scholar]
  89. Zhang, Z.; Girard, J.M.; Wu, Y.; Zhang, X.; Liu, P.; Ciftci, U.; Canavan, S.; Reale, M.; Horowitz, A.; Yang, H.; et al. Multimodal spontaneous emotion corpus for human behavior analysis. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 3438–3446. [Google Scholar]
  90. Bobbia, S.; Macwan, R.; Benezeth, Y.; Mansouri, A.; Dubois, J. Unsupervised skin tissue segmentation for remote photoplethysmography. Pattern Recognit. Lett. 2019, 124, 82–90. [Google Scholar] [CrossRef]
  91. Sabour, N.; Benezeth, J.; Yang, F.; Dérian, P.; Gabaudan, V.; Bonnetain, F.; Marzani, F.; Petit, D.; Verges, G.; Benoit, A. UBFC-Phys: A multimodal database for psychophysiological studies of social stress. IEEE Trans. Affect. Comput. 2021, 14, 1490–1501. [Google Scholar] [CrossRef]
  92. Pilz, K.; Ostwald, S.; Chuquimia, O.; Zaunseder, S. Local group invariance for heart rate estimation from face videos in the wild. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Salt Lake City, UT, USA, 18–22 June 2018; pp. 1254–1262. [Google Scholar]
  93. Liu, Z.; Wang, X.; Cai, J.; Chen, J.; Zhou, Y. IntensePhysio: Deep Physiological Signal Inference During High-Intensity Physical Activity. arXiv 2022, arXiv:2211.08424. [Google Scholar]
  94. Nowara, E.M.; Sabharwal, A.; Veeraraghavan, A. Vision for vitals: Unbiased physiological measurement from face videos in the wild. In Proceedings of the 2023 IEEE/CVF Winter Conference on Applications of Computer Vision (WACV), Waikoloa, HI, USA, 2–7 January 2023; pp. 2037–2046. [Google Scholar]
  95. Liu, Y.; Tadi, M.J.; Lohan, E.; Rantanen, P.; Vasankari, T. ECG-Fitness: An efficient CNN model for ECG-derived heart rate estimation in fitness activities. In Proceedings of the 2020 IEEE International Conference on Industrial Engineering and Engineering Management (IEEM), Singapore, 14–17 December 2020; pp. 1154–1158. [Google Scholar]
  96. Koelstra, S.; Muhl, C.; Soleymani, M.; Lee, J.-S.; Yazdani, A.; Ebrahimi, T.; Pun, T.; Nijholt, A.; Patras, I. DEAP: A Database for Emotion Analysis; Using Physiological Signals. IEEE Trans. Affect. Comput. 2012, 3, 18–31. [Google Scholar] [CrossRef]
  97. Soleymani, M.; Lichtenauer, J.; Pun, T.; Pantic, M. MAUS: A Multimodal Affective User State Dataset. In Proceedings of the 2014 ACM International Conference on Multimodal Interaction, Istanbul, Turkey, 12–16 November 2014; pp. 464–470. [Google Scholar]
  98. Nowara, E.M.; Sabharwal, A.; Veeraraghavan, A. VicarPPG: Video-based cardiac pulse measurement in the wild. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18–24 June 2022; pp. 2047–2056. [Google Scholar]
  99. Song, R.; Zhang, S.; Li, C.; Zhang, Y.; Cheng, J.; Chen, X. MMSE-HR2: A multimodal dataset for remote physiological measurement. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18–24 June 2022; pp. 151–160. [Google Scholar]
  100. Niu, X.; Han, H.; Shan, S.; Chen, X. VIPL-HR-V2: A large-scale video-based heart rate dataset. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18–24 June 2022; pp. 161–170. [Google Scholar]
  101. Bobbia, S.; Benezeth, Y.; Benoit, A.; Dubois, J. Oulu BioFace: A dataset for remote physiological signal measurement. In Proceedings of the 2021 IEEE International Conference on Image Processing (ICIP), Online, 19–22 September 2021; pp. 1234–1238. [Google Scholar]
  102. Zhang, Z.; Girard, M.; Wu, Y.; Zhang, X.; Liu, P.; Ciftci, U.; Canavan, S.; Reale, M.; Horowitz, A.; Yang, H.; et al. BP4D-Spontaneous: A high-resolution spontaneous 3D dynamic facial expression database. Image Vis. Comput. 2014, 32, 692–706. [Google Scholar] [CrossRef]
  103. Lee, H.-C.; Park, Y.; Yoon, S.B.; Yang, S.M.; Park, D.; Jung, C.-W. VitalDB, a high-fidelity multi-parameter vital signs database in surgical patients. Sci. Data 2022, 9, 279. [Google Scholar] [CrossRef] [PubMed]
  104. Goldberger, A.L.; Amaral, L.A.N.; Glass, L.; Hausdorff, J.M.; Ivanov, P.C.; Mark, R.G.; Mietus, J.E.; Moody, G.B.; Peng, C.-K.; Stanley, H.E. PhysioBank, PhysioToolkit, and PhysioNet: Components of a new research resource for complex physiologic signals. Circulation 2000, 101, e215–e220. [Google Scholar] [CrossRef]
  105. Johnson, A.E.; Pollard, T.J.; Shen, L.; Lehman, H.; Feng, M.; Ghassemi, M.; Moody, B.; Szolovits, P.; Celi, L.A.; Mark, R.G. MIMIC-III, a freely accessible critical care database. Sci. Data 2016, 3, 160035. [Google Scholar] [CrossRef] [PubMed]
  106. Panetti, G.; Tsiamyrtzis, P.; Pavlidis, I. NeuroTex: An integrated non-contact system for continuous unobtrusive physiological and neurobehavioral monitoring. In Proceedings of the 2020 IEEE 17th International Conference on Wearable and Implantable Body Sensor Networks (BSN), Ioannina, Greece, 27–30 September 2020; pp. 1–4. [Google Scholar]
  107. Majumder, S.; Chen, L.; Marinov, O.; Chen, C.H.; Mondal, T.; Deen, M.J. Noncontact wearable wireless ECG systems for long-term monitoring. IEEE Rev. Biomed. Eng. 2019, 11, 306–321. [Google Scholar] [CrossRef]
  108. Hassan, M.A.; Malik, A.S.; Fofi, D.; Saad, N.M.; Ali, Y.S.; Meriaudeau, F. Video-based heartbeat rate measuring method using ballistocardiography. IEEE Sens. J. 2017, 17, 4544–4557. [Google Scholar] [CrossRef]
  109. Zhao, Z.; Yang, X.; Ren, S.; Gao, J.; Hu, T.; Song, P.; You, Y. Respiratory rhythm extraction from pulse oximeter signals using the principal component analysis and the short-time Fourier transform. IEEE Access 2020, 8, 86184–86195. [Google Scholar]
  110. Liu, S.; Ostadabbas, S.; Stein, K.M.; Taha, B. Multi-task deep learning for cardiac signals: Applications in wearable technology. Health Inf. Sci. Syst. 2019, 7, 1–11. [Google Scholar]
  111. Wang, Z.; Qian, X.; Yang, J.; Xu, J. HRNet: High-resolution networks for visual recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 43, 777–789. [Google Scholar] [CrossRef]
  112. Li, X.; Chen, J.; Zhao, G.; Pietikäinen, M. Remote heart rate measurement from face videos under realistic situations. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 4264–4271. [Google Scholar]
  113. Chen, X.; Cheng, J.; Song, R.; Liu, Y.; Ward, R.; Wang, Z.J. Video-based heart rate measurement: Recent advances and future prospects. IEEE Trans. Instrum. Meas. 2019, 68, 3600–3615. [Google Scholar] [CrossRef]
  114. Zhao, Z.; Zhao, J.; Xia, Y.; Chen, J. FedPGM: Federated probabilistic graphical model for healthcare representation learning from multi-center EHRs. Int. J. Med. Inform. 2022, 159, 104679. [Google Scholar]
  115. Zhang, Y.; Chen, Y.; Yu, C.; Yang, Z. FarSense: Pushing the range limit of WiFi-based respiration sensing with CSI ratio of two antennas. Proc. ACM Interact. Mob. Wearable Ubiquitous Technol. 2019, 3, 121. [Google Scholar]
  116. Caruana, R. Multitask learning. Mach. Learn. 1997, 28, 41–75. [Google Scholar] [CrossRef]
  117. Chen, X.; Liu, A.; Zheng, K.; Wang, Z.J. PhysFormer: Facial video-based physiological measurement with temporal difference transformer. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 13165–13175. [Google Scholar]
  118. Kendall, A.; Gal, Y.; Cipolla, R. Multi-task learning using uncertainty to weigh losses for scene geometry and semantics. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 7482–7491. [Google Scholar]
  119. Villarroel, M.; Chaichulee, S.; Jorge, J.; Davis, S.; Green, G.; Arteta, C.; Zisserman, A.; McCormick, K.; Watkinson, P.; Tarassenko, L. Non-contact vital sign monitoring in the neonatal intensive care unit. In Proceedings of the 2019 41st Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), Berlin, Germany, 23–27 July 2019; pp. 85–88. [Google Scholar]
  120. Martinez, M.; Stiefelhagen, R.; Breuel, T. MMFT-BERT: Multimodal fusion transformer with BERT encodings for visual question answering. In Proceedings of the Conference on Empirical Methods in Natural Language Processing (EMNLP), Online, 16–20 November 2020; pp. 4648–4660. [Google Scholar]
  121. Liu, Z.; Wu, D.; Zhao, J.; Zhang, Z.; Su, L.; Chen, X. MilliEye: A lightweight mmWave radar and camera fusion system for robust vital sign sensing. In Proceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies, 2022; Volume 6, pp. 1–27. [Google Scholar]
  122. McDuff, D.; Hurter, C.; Gonzalez-Franco, M. Pulse and vital sign measurement in mixed reality using a HoloLens. In Proceedings of the 23rd ACM Symposium on Virtual Reality Software and Technology, Gothenburg, Sweden, 8–10 November 2019; pp. 1–9. [Google Scholar]
Figure 1. This framework illustrates the integration of multiple sensing modalities with multi-task learning approaches for simultaneous extraction of various vital signs, including heart rate, respiratory rate, blood pressure, and oxygen saturation. The framework demonstrates how complementary sensing technologies and shared computational architectures enable robust, non-contact vital sign monitoring across diverse environmental conditions.
Figure 1. This framework illustrates the integration of multiple sensing modalities with multi-task learning approaches for simultaneous extraction of various vital signs, including heart rate, respiratory rate, blood pressure, and oxygen saturation. The framework demonstrates how complementary sensing technologies and shared computational architectures enable robust, non-contact vital sign monitoring across diverse environmental conditions.
Sensors 25 04792 g001
Figure 2. Illustration of contactless vital sign monitoring technologies: (a) Vision-based methods using cameras and ambient light, (b) radar-based methods utilizing radio waves, (c) thermal imaging detecting infrared heat patterns, and (d) ambient sensing capturing environmental changes for vital signs monitoring.
Figure 2. Illustration of contactless vital sign monitoring technologies: (a) Vision-based methods using cameras and ambient light, (b) radar-based methods utilizing radio waves, (c) thermal imaging detecting infrared heat patterns, and (d) ambient sensing capturing environmental changes for vital signs monitoring.
Sensors 25 04792 g002
Figure 3. The deep learning architecture of MTTS-CAN (Multi-Task Temporal Shift-Convolutional Attention Network) for camera-based vital sign measurement, as depicted in [28].
Figure 3. The deep learning architecture of MTTS-CAN (Multi-Task Temporal Shift-Convolutional Attention Network) for camera-based vital sign measurement, as depicted in [28].
Sensors 25 04792 g003
Figure 4. FS-Net network architecture, as depicted in [52].
Figure 4. FS-Net network architecture, as depicted in [52].
Sensors 25 04792 g004
Figure 5. The overall architecture of the fusion-vital network, as depicted in [28].
Figure 5. The overall architecture of the fusion-vital network, as depicted in [28].
Sensors 25 04792 g005
Table 2. Key datasets for contactless vital sign monitoring.
Table 2. Key datasets for contactless vital sign monitoring.
YearReferenceNameModality TypeVital SignsSamples/Subjects
2012[85]MAHNOB-HCIVideo, ECG, Resp., Temp.HR, Resp., Temp.27 subjects
2014[86]PUREVideo (RGB)HR10 subjects, multiple sessions
2017[87]COHFACEVideo (RGB), PhysioHR, Resp.40 subjects
2019[63]VIPL-HRVideo (RGB), Pulse Ox, Chest StrapHR, SpO2, Resp.107 subjects, 2378 videos
2017[88]UTA-RLDDVideo (RGB), PhysioHR, Resp., Drowsiness60 subjects, 30+ hours
2016[89]MMSE-HRVideo (RGB), PhysioHR40 subjects
2019[90]UBFC-rPPGVideo (RGB), Pulse OxHR42 videos
2021[91]UBFC-PhysVideo (RGB), ECG, Resp., EDAHR, Resp., EDA56 subjects
2018[92]LGI-PPGIVideo (RGB), Pulse OxHR25 subjects
2022[93]IntensePhysioVideo (RGB), PhysioHR, Resp.50 subjects
2023[94]Vision for VitalsVideo (RGB), PhysioHR, Resp.100+ subjects
2020[95]ECG-FitnessVideo (RGB), ECGHR30 subjects
2012[96]DEAPVideo (face), EEG, PhysioHR, EDA, Resp., EEG32 subjects
2014[97]MAUSVideo, Audio, PhysioHR, Resp., EDA, Temp.30 subjects
2022[98]VicarPPGVideo (RGB), PPGHR, HRV100+ subjects
2022[99]MMSE-HR2Video (RGB), PhysioHR, Resp.40 subjects
2022[100]VIPL-HR-V2Video (RGB), PhysioHR, Resp.200+ subjects, 3000+ videos
2021[101]Oulu BioFaceVideo (RGB), PhysioHR, Resp.100+ subjects
2016[102]BP4D-SpontaneousVideo (RGB), PhysioHR, BP, Temp, Resp, EDA140 subjects
2018[103]VortalVideo (RGB), PPG, BP, SpO2HR, BP, SpO239 subjects
2015[54]IEEE SPC 2015Video (RGB), PPG, BPHR, BP12 subjects
2016[90]UBFC-BVPVideo (RGB), PPGHR, HRV42 subjects
2012[104]BIDMC PPGPPG, Resp, ECGHR, Resp, SpO2, HRV53 recordings
2016[105]MIMIC-IIIClinical (ECG, PPG, ABP, SpO2, Temp, Resp)HR, BP, SpO2, Temp, Resp, HRV60,000+ ICU stays
Table 3. Best reported accuracies for key contactless vital sign monitoring datasets. HR: heart rate, RR: respiratory rate, r: correlation coefficient, MAPE: Mean Absolute Percentage Error, EDA: Electrodermal Activity, HRV: heart rate variability, BP: blood pressure, SpO2: blood oxygen saturation.
Table 3. Best reported accuracies for key contactless vital sign monitoring datasets. HR: heart rate, RR: respiratory rate, r: correlation coefficient, MAPE: Mean Absolute Percentage Error, EDA: Electrodermal Activity, HRV: heart rate variability, BP: blood pressure, SpO2: blood oxygen saturation.
DatasetBest ReferenceHR Acc.HR rRR Acc.RR rOther Metrics
MAHNOB-HCI[57]95.6%0.92--MAPE: 4.4%
PURE[54]97.8%0.98--MAPE: 2.2%
COHFACE[72]96.7%0.95--MAPE: 3.3%
VIPL-HR[55]94.2%0.89--MAPE: 5.8%
UTA-RLDD[61]92.1%0.8788.4%0.82MAPE: 7.9%
MMSE-HR[56]94.8%0.91--MAPE: 5.2%
UBFC-rPPG[68]97.2%0.97--MAPE: 2.8%
UBFC-Phys[91]96.8%0.9491.2%0.85EDA: r = 0.85
LGI-PPGI[73]97.1%0.96--MAPE: 2.9%
IntensePhysio[70]89.3%0.7882.1%0.74MAPE: 10.7%
Vision for Vitals[94]95.2%0.8887.9%0.81Cross-demo: 91.2%
ECG-Fitness[71]93.6%0.88--MAPE: 6.4%
DEAP[96]94.1%0.8985.6%0.79EEG: 85.0%
MAUS[97]94.9%0.9088.7%0.82EDA: r = 0.82
VicarPPG[98]96.0%0.92--HRV: 87.3%
MMSE-HR2[99]95.5%0.9186.2%0.80Multi-task: 93.8%
VIPL-HR-V2[100]94.5%0.8887.8%0.83Large-scale: 92.1%
Oulu BioFace[101]96.4%0.9389.1%0.85Multi-modal: 94.6%
BP4D-Spontaneous[102]94.2%0.8784.3%0.78BP: 91.1%
VitalDB[103]94.7%0.89--SpO2: 97.9%
IEEE SPC 2015[54]96.8%0.95--BP: 92.2%
UBFC-BVP[90]97.0%0.96--HRV: 88.2%
BIDMC PPG[104]98.3%0.9893.8%0.94SpO2: 98.2%
MIMIC-III[105]96.0%0.9289.4%0.86BP: 93.3%
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Hassanpour, A.; Yang, B. Contactless Vital Sign Monitoring: A Review Towards Multi-Modal Multi-Task Approaches. Sensors 2025, 25, 4792. https://doi.org/10.3390/s25154792

AMA Style

Hassanpour A, Yang B. Contactless Vital Sign Monitoring: A Review Towards Multi-Modal Multi-Task Approaches. Sensors. 2025; 25(15):4792. https://doi.org/10.3390/s25154792

Chicago/Turabian Style

Hassanpour, Ahmad, and Bian Yang. 2025. "Contactless Vital Sign Monitoring: A Review Towards Multi-Modal Multi-Task Approaches" Sensors 25, no. 15: 4792. https://doi.org/10.3390/s25154792

APA Style

Hassanpour, A., & Yang, B. (2025). Contactless Vital Sign Monitoring: A Review Towards Multi-Modal Multi-Task Approaches. Sensors, 25(15), 4792. https://doi.org/10.3390/s25154792

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Article metric data becomes available approximately 24 hours after publication online.
Back to TopTop