Next Article in Journal
Simple Calibration of Three-Axis Accelerometers with Freely Rotating Vertical Bench
Previous Article in Journal
An Optical Sensor for Measuring In-Plane Linear and Rotational Displacement
Previous Article in Special Issue
Building a Gender-Bias-Resistant Super Corpus as a Deep Learning Baseline for Speech Emotion Recognition
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Emotion Recognition from rPPG via Physiologically Inspired Temporal Encoding and Attention-Based Curriculum Learning

1
Department of Human-Centered Artificial Intelligence, Sangmyung University, Seoul 03016, Republic of Korea
2
Department of Emotion Engineering, Sangmyung University, Seoul 03016, Republic of Korea
*
Author to whom correspondence should be addressed.
Sensors 2025, 25(13), 3995; https://doi.org/10.3390/s25133995
Submission received: 7 May 2025 / Revised: 16 June 2025 / Accepted: 23 June 2025 / Published: 26 June 2025
(This article belongs to the Special Issue Emotion Recognition and Cognitive Behavior Analysis Based on Sensors)

Abstract

Highlights

What are the main findings?
  • A temporal-only rPPG framework with a multi-scale CNN, sparse α-Entmax attention, and Gated Pooling achieved 66.04% accuracy and a 61.97% weighted F1-score for arousal on MAHNOB-HCI (mixed subjects).
  • The model underperformed for valence (62.26% accuracy), highlighting the physiological limits of unimodal time-series signals.
What are the implications of the main findings?
  • Temporal rPPG can rival other single-modality methods for arousal when physiologically inspired temporal modeling is applied.
  • Addressing valence requires the integration of spatial or multimodal cues, guiding future affective computing designs.

Abstract

Remote photoplethysmography (rPPG) enables non-contact physiological measurement for emotion recognition, yet the temporally sparse nature of emotional cardiovascular responses, intrinsic measurement noise, weak session-level labels, and subtle correlates of valence pose critical challenges. To address these issues, we propose a physiologically inspired deep learning framework comprising a Multi-scale Temporal Dynamics Encoder (MTDE) to capture autonomic nervous system dynamics across multiple timescales, an adaptive sparse α-Entmax attention mechanism to identify salient emotional segments amidst noisy signals, Gated Temporal Pooling for the robust aggregation of emotional features, and a structured three-phase curriculum learning strategy to systematically handle temporal sparsity, weak labels, and noise. Evaluated on the MAHNOB-HCI dataset (27 subjects and 527 sessions with a subject-mixed split), our temporal-only model achieved competitive performance in arousal recognition (66.04% accuracy; 61.97% weighted F1-score), surpassing prior CNN-LSTM baselines. However, lower performance in valence (62.26% accuracy) revealed inherent physiological limitations regarding a unimodal temporal cardiovascular analysis. These findings establish clear benchmarks for temporal-only rPPG emotion recognition and underscore the necessity of incorporating spatial or multimodal information to effectively capture nuanced emotional dimensions such as valence, guiding future research directions in affective computing.

1. Introduction

Emotion recognition is a fundamental component of affective computing and human–computer interaction, with significant implications across healthcare, education, and consumer technologies [1,2]. Traditional methods primarily rely on observable cues such as facial expressions or speech. However, these external indicators can be intentionally controlled or masked, limiting their reliability in representing genuine emotional states [3]. Physiological signals—such as heart rate, blood volume pulse (BVP), and skin conductance—regulated by the autonomic nervous system (ANS), offer a more authentic and less voluntarily modifiable reflection of emotional states, making them ideal for unobtrusive affective computing [4].
Extensive psychophysiological research has firmly established that emotional states trigger characteristic, transient changes in cardiovascular activity [4]. Heart rate variability (HRV), defined by fluctuations in intervals between heartbeats, reflects ANS activity, distinctly correlating with emotional regulation processes [5]. Additionally, pulse waveform morphology captures vascular tone variations associated directly with emotional arousal [6]. Notably, these physiological responses are transient, sparsely distributed, and exhibit non-uniform temporal patterns [4,7]—highlighting a critical gap in current approaches: effective identification and the interpretation of emotionally salient temporal segments amidst noisy physiological signals.
Recent advancements in remote photoplethysmography (rPPG) enable unobtrusive, camera-based monitoring of cardiovascular activity by measuring subtle skin color variations induced by cardiac pulse waves [8]. This technique allows scalable affective computing applications across diverse industrial contexts due to the proliferation of camera-equipped devices (e.g., smartphones, laptops, and surveillance systems), significantly broadening the practical utility of emotion recognition technology [9].
Despite promising potential, recognizing emotions exclusively from unimodal temporal rPPG signals has significant unresolved challenges. Firstly, emotional physiological responses often manifest briefly and sporadically rather than continuously, complicating effective temporal analysis [4]. Secondly, rPPG signals inherently suffer from noise and artifacts from subject motion and environmental lighting changes compared to contact-based methods, impairing robust interpretation of subtle emotional cues [8,9]. Thirdly, typical session-level annotations induce a weak-label Multiple Instance Learning (MIL) scenario [10], necessitating sophisticated models to pinpoint informative temporal segments accurately, a challenge recently tackled in rPPG with domain adaptation techniques [11]. Fourthly, recognizing valence (the pleasantness of an emotion) from physiological signals remains inherently more challenging than arousal (the intensity of an emotion) as its correlates are more complex and less directly tied to general ANS activation [4,12]. Furthermore, large-scale, deep-learning-ready rPPG corpora remain rare; most public datasets contain <50 participants and fewer than 10 sessions per subject, thereby limiting the statistical power of data-hungry models such as LSTM or Transformer variants [13,14].
Consequently, we deliberately restrict our first evaluation to the well-controlled MAHNOB-HCI benchmark—which provides a relatively large number of sessions per subject—to isolate the intrinsic capabilities and limits of temporal-only rPPG before tackling cross-dataset generalization in future work in Section 5.2.
Our study explicitly addresses these challenges by proposing a novel deep learning framework. By processing signals in short, localized temporal chunks, we effectively isolate and analyze transient physiological responses. We introduce a Multi-scale Temporal Dynamics Encoder (MTDE), physiologically motivated by multi-rate ANS response characteristics [5,15] and designed to capture subtle temporal patterns across different timescales [16]. Furthermore, an adaptive sparse attention mechanism leveraging α-Entmax and entropy regularization explicitly identifies and prioritizes temporally sparse emotional segments, emulating selective human attentional processes [17,18,19]. A novel Gated Temporal Pooling mechanism robustly aggregates chunk-level information, effectively filtering noise through joint temporal weighting and feature-level gating [20].
Critically, we employ a physiologically inspired, three-phase curriculum learning strategy—exploration, discrimination, and exploitation—mirroring human attentional refinement and developmental learning principles [21,22]. This systematic training approach addresses weak-label issues, temporal sparsity, and signal noise incrementally, enabling stable, progressive learning from complex, noisy temporal data. To robustly evaluate our method, we employ weighted F1-scores, addressing class imbalance more objectively compared to prior work such as that by Mellouk & Handouzi [23]. This work therefore establishes a transparent baseline that future research in cross-dataset generalization and domain adaptation can build upon. Our evaluations on the MAHNOB-HCI dataset show promising results: for arousal, it achieves an accuracy of 66%, an F1-score of 0.7429, and a weighted F1-score of 0.6224; for valence, it achieves an accuracy of 62.3%, an F1-score of 0.6667, and a weighted F1-score of 0.6227, underscoring our model’s effectiveness despite inherent challenges.
Our contributions explicitly bridge critical research gaps, providing clear advancements:
  • Focused Temporal Analysis of rPPG: This method establishes foundational insights into the capabilities and limitations of using exclusively temporal physiological information, providing a rigorous benchmark.
  • Multi-scale Temporal Dynamics Encoder (MTDE): This effectively captures physiologically meaningful ANS responses across multiple timescales, addressing complexity in subtle temporal emotional signals.
  • Adaptive Sparse Attention: This method precisely identifies transient, emotionally relevant physiological segments amidst noisy rPPG data, significantly enhancing robustness.
  • Gated Temporal Pooling: This method sophisticatedly aggregates emotional information across temporal chunks, effectively mitigating noise and irrelevant features.
  • Curriculum Learning Strategy: This method systematically addresses learning complexities associated with weak labels, noise, and temporal sparsity, ensuring robust, stable model learning.

2. Related Work

2.1. The Physiological Signals for Emotion Recognition

Cardiovascular activity is tightly coupled with autonomic nervous system (ANS) dynamics and therefore provides a robust, involuntary window on affective states [4,24]. Two families of descriptors are especially relevant:
  • Heart-rate variability (HRV). Spectral power in the very-low-, low- and high-frequency bands maps onto distinct sympathetic–parasympathetic rhythms and has been linked to both arousal and valence [5,15].
  • Pulse-wave morphology. Beat-to-beat amplitude and upslope changes correlate with rapid sympathetic activation, while longer-term contour variations reflect vagal tone [25,26].
Early affective computing systems relied on hand-crafted HRV indices (e.g., LF/HF ratio and SDNN). Later work introduced shallow CNNs or temporal convolutional networks (TCNs) that learn nonlinear dynamics directly from the waveform [16].
Our Multi-scale Temporal Dynamics Encoder (MTDE) builds upon this principle by operationalizing physiological insights into its architecture. Its three one-dimensional convolutional branches have effective receptive fields (≈0.2 s, 2 s, and 4 s) that are designed to intentionally cover the canonical HF, LF, and VLF HRV windows [5,15]. As validated by a kernel-size sensitivity study in Section 4.2, this physiologically aligned design yields superior performance compared to single-scale or randomly dilated variants.

2.2. Remote PPG Signal Extraction and Denoising

Remote PPG (rPPG) estimates the blood volume pulse from skin-color variations in ordinary video but is highly sensitive to illumination changes, motion, and camera noise [9,27]. Public rPPG–emotion databases are also small (<50 subjects) and heterogeneous in frame rate or lighting, limiting large-scale learning [9,28].
Recent end-to-end extractors—PhysNet [29], PhysFormer [30], RhythmFormer [31], and PhysMamba [32]—combine long-range sequence modeling with explicit noise suppression. We adopt PhysMamba [32] as a front-end because it (i) achieves the highest signal-to-noise ratio on MAHNOB-HCI, (ii) is open-source, and (iii) provides a stable, high-quality BVP signal. This approach allows for a focused analysis of the downstream emotion recognition stage as the influence of signal extraction quality is minimized.

2.3. Emotion Recognition from rPPG/PPG

Most previous rPPG affect studies fuse the pulse trace with facial appearance cues, obscuring the contribution of the temporal signal itself [9]. The few unimodal–temporal approaches can be grouped as follows:
  • CNN-LSTM. Mellouk & Handouzi [23] combine 2D convolutions with an LSTM on four-second segments and report the first subject-balanced results on MAHNOB-HCI. The network architecture and training protocol are fully specified, making it the most appropriate domain-equivalent baseline.
  • TCN and Transformer variants. Later works explored purely convolutional temporal blocks [16] or self-attention for longer contexts [30,31], but most studies used proprietary datasets or combined spatial features, preventing a direct apples-to-apples comparison on rPPG alone.
Given this landscape, CNN-LSTM serves as our primary, temporally focused reference. To further contextualize our model’s performance against non-recurrent architectures, additional various backbones including lightweight eight-layer TCNs are implemented as a comparison baseline (see Section 4.2 and Appendix E).

2.4. Comparison and Key Differences

In summary, our work is distinguished from prior studies in several key aspects. The following points highlight our main contributions:
  • Temporal-only perspective. We analyze the BVP waveform in isolation, establishing a lower-bound benchmark that future multimodal extensions can build upon.
  • Physiology-guided multi-scale encoding. MTDE hard-codes HRV-aligned receptive fields [22,30] and is empirically validated via an ablation on kernel size.
  • Interpretable sparse attention. α-Entmax [16] learns a few high-confidence temporal “spikes”; heat-map visualizations (Figure 5) coincide with sympathetic bursts in the waveform, aligning with theories of selective attention [19].
  • Noise-robust aggregation. Gated Pooling jointly weights time steps and feature channels, reducing artifacts propagated from the rPPG extractor.
  • Rigorous protocol and efficiency. A dedicated hold-out test set, weighted F1 scoring, and a 198 k-parameter model that runs at 0.66 s per two-minute video establish (Section 3.7) both statistical and practical credibility.
Together, these contributions define the first systematically evaluated baseline for temporal-only rPPG emotion recognition and lay the groundwork for future cross-dataset or domain adaptation studies.

3. Methodology

This section delineates the methodology employed in our study for emotion recognition from remote photoplethysmography (rPPG) signals. We describe the dataset utilized, the preprocessing steps, the overall framework architecture, the detailed components of our model, the physiologically inspired curriculum learning strategy, the evaluation metrics and comparative baseline, and the experimental setup.

3.1. Dataset and Experimental Protocol

3.1.1. MAHNOB-HCI Dataset

We employ the publicly available MAHNOB-HCI multimodal emotion dataset, which comprises recordings from 27 subjects who viewed 20 emotional film clips [28]. The dataset represents one of the most comprehensive multimodal emotion databases with synchronized physiological signals, facial videos, and subjective ratings. After excluding unusable data from 3 subjects, a total of 527 valid sessions were utilized for this study. The dataset provides synchronized facial videos and ground-truth physiological signals, including electrocardiogram (ECG) data recorded at a 256 Hz sampling rate.
All videos were downsampled to 30 fps, a rate validated as sufficient for capturing subtle cardiovascular dynamics and maintaining HRV analysis accuracy. Studies have demonstrated that frame rates of 30 fps or higher preserve the temporal resolution necessary for accurate heart rate variability measurements. For classification, the self-reported valence and arousal ratings (on a 1–9 scale) were binarized into “low” (ratings 1–4) and “high” (ratings 5–9) classes by thresholding at the midpoint (4.5), following established protocols in affective computing research [33].

3.1.2. Data Partitioning and Protocol Rationale

To maximize the data available for our deep learning model, the 527 sessions were partitioned into training (80%), validation (10%), and testing (10%) sets using a session-based random split. We acknowledge that a strict subject-independent split is the gold standard for evaluating generalization. However, as established in systematic reviews, the landscape of public cardiovascular-based emotion datasets is characterized by limited scale; a comprehensive survey of 18 such datasets revealed that most include fewer than 50 participants, with many offering limited session diversity per subject [6]. This data scarcity poses a significant challenge for training data-hungry deep learning models.
Consequently, we opted for this subject-mixed protocol to facilitate a thorough investigation of our proposed architecture’s capabilities on a rich set of session data. We explicitly acknowledge that this protocol may lead to an overestimation of the model’s generalization capability to entirely unseen subjects, a limitation that, along with the challenges of cross-dataset validation, is further addressed in our discussion (Section 5.1).

3.1.3. A Two-Stage Process for Optimal Method Selection

To ensure a high-quality physiological input, we conducted a rigorous, two-stage experiment to select the optimal rPPG method. The MAHNOB-HCI dataset provides ground-truth physiological signals, including ECG data; however, it does not include raw contact-based photoplethysmography (PPG) signals for direct comparison. As a ground-truth heart rate (HR), we processed the dataset’s raw ECG signals; R-peaks were detected using the Pan–Tompkins algorithm [34] with a 0.5–40 Hz band-pass filter, and the resulting inter-beat intervals (IBIs) were used to calculate HR (HR = 60/IBI), which was then resampled to 30 Hz.
First, a broad comparison of traditional unsupervised algorithms against state-of-the-art deep learning models was conducted on our target dataset. The accuracy of the extracted signals was measured against the ECG-based ground-truth using Mean Absolute Error (MAE) in beats per minute (bpm) and Root Mean Square Error (RMSE). As summarized in Table 1, deep-learning-based methods demonstrated vastly superior accuracy, with PhysFormer [30] and PhysMamba [32] (both pre-trained on UBFC-rPPG) emerging as the top two performers.
Second, to make a final selection, we conducted a head-to-head analysis focusing on signal quality (SNR) and computational efficiency (VRAM usage). This test was performed using the standard UBFC-rPPG intra-dataset protocol [35]. The results in Table 2 clearly favor PhysMamba, which achieved a significantly higher SNR (3.67 vs. 0.25) and lower VRAM usage.

3.1.4. Final Method and Implementation

While our validation on MAHNOB-HCI yielded an MAE of 5.38 bpm for HR estimation, our primary goal is to obtain a high-fidelity BVP waveform for learning subtle affective features. Research has demonstrated that the morphological characteristics of PPG signals contain essential physiological information for emotion recognition, with signal quality being more critical than absolute HR accuracy for the effective classification of affective state [6,36,37]. Therefore, based on its superior signal quality (SNR) and computation efficiency, PhysMamba [32] was selected as the optimal front-end extractor. For implementation, the selected model was applied to 128-frame chunks of the videos using the DiffNormalized scheme [38].

3.2. Overall Framework

The complete end-to-end pipeline, together with its three curriculum phases, is shown in Figure 1. Raw RGB video is segmented into 128-frame chunks (≈4 s at 30 fps) and converted to rPPG waveforms by a pretrained PhysMamba extractor. Each chunk is then encoded by the Multi-scale Temporal Dynamics Encoder (MTDE). Depending on the training phase, different modules become active: the projection head and L s u p c o n in Phase 0 (Figure 2), the chunk auxiliary classifier and Top-K α-Entmax attention in Phase 1 (Figure 3), and the Gated Pooling plus session classifier in Phase 2 (Figure 4). At inference time, the green-outlined path in Figure 1 (MTDE → α-Entmax → Gated Pooling → Main Classifier) constitutes the operational model.
Our emotion recognition model processes the 1D temporal rPPG signal, derived from the video, in fixed-length, non-overlapping temporal chunks of 128 frames (approximately 4 s at 30 fps). This specific chunk size was strategically chosen for multiple reasons. Firstly, it aligns with the temporal window used by the robust PhysMamba rPPG extractor [32] and common processing units in the rPPG-Toolbox framework [38]. Secondly, and crucially, prior work [23] has demonstrated that a 4-s segmentation size yields optimal performance for emotion classification from contactless PPG signals, reinforcing its appropriateness for capturing pertinent physiological dynamics within the temporal domain. Physiologically, a 4-s window is well-suited as it typically encompasses several cardiac cycles (e.g., approximately 4–7 heartbeats at a resting heart rate of 60–100 bpm). Analyzing physiological patterns such as heart rate variability (HRV) [5,15] or subtle pulse waveform changes, which are known to reflect autonomic nervous system dynamics [39] and have been effectively used for affective computing [6] over this duration, allows for the capture of meaningful short-term autonomic nervous system (ANS) modulations, which are widely recognized as crucial indicators of emotional states.
Each temporal chunk is processed using the Multi-scale Temporal Dynamics Encoder (MTDE), yielding a 256-dimensional feature embedding. The embeddings are scored by the attention scoring module (AttnScorer) via α-Entmax attention and, in parallel, forwarded to the Gated Pooling module (Gated Pooling). Gated Pooling integrates the attended and feature-gated embeddings into a single session representation from which the main classifier predicts the valence and arousal labels using only the aggregated temporal information. Phase-specific modules that augment this core path are illustrated in Figure 2 (Phase 0), Figure 3 (Phase 1), and Figure 4 (Phase 2) and are detailed in the curriculum learning section (Section 3.4).

3.3. Training Modules

This section provides a detailed description of the core modules constituting our emotion recognition framework.

3.3.1. rPPG Extraction Front-End (PhysMamba)

As determined by our selection process (Section 3.1.3), the UBFC-rPPG-pretrained PhysMamba model [32,38]—chosen for its superior SNR and lower VRAM footprint—serves as the rPPG extraction front-end in our framework. Its primary role is to extract a refined, denoised 1D temporal rPPG (blood- volume pulse, BVP) signal from raw facial video frames.
PhysMamba operates on non-overlapping 128-frame chunks (≈4 s at 30 fps). Each chunk is first pre-processed with the DiffNormalized scheme [32,38], which computes frame-wise ratio differences normalized by the standard deviation, thereby enhancing subtle blood flow changes. The resulting 1D BVP waveform—produced without any additional per-session rescaling or band-pass filtering—constitutes the exclusive input to the downstream emotion recognition modules.

3.3.2. Multi-Scale Temporal Dynamics Encoder (MTDE)

The MTDE’s purpose is to learn rich feature representations by capturing physiological dynamics across various temporal scales within each 128-frame BVP chunk. The design is physiologically grounded in the principle that the autonomic nervous system (ANS) modulates the cardiac pulse through multiple superimposed rhythms [40], with sympathetic and parasympathetic activities contributing to distinct frequency components of heart rate variability [15]. While these rhythms are classically studied via heart rate variability (HRV) frequency bands [41], our model does not compute HRV metrics explicitly. Instead, it learns features directly from the BVP waveform that are implicitly sensitive to the physiological processes underlying these HRV bands.
As detailed in Appendix A, the MTDE comprises two main stages: a SlimStem and a Multi-Scale Temporal Block (MSTB). The SlimStem consists of two sequential 1D convolutional layers for initial noise reduction and low-level feature extraction, conceptually akin to early sensory filtering. The subsequent MSTB implements our physiology-aligned strategy, featuring a three-branch architecture using dilated convolutions. Each branch is designed to have an effective receptive field (RF) that spans one of the key ANS timescales:
  • Short-scale branch (RF 0.2 s): This scale is sensitive to rapid beat-to-beat changes in pulse morphology, reflecting high-frequency dynamics driven by parasympathetic (vagal) fluctuations, analogous to the HF-like band of HRV [4].
  • Medium-scale branch (RF 2.2 s): This branch targets slower oscillations from the interplay between sympathetic and parasympathetic systems (e.g., baroreflex), operating in a window comparable to the LF-like band [4].
  • Long-scale branch (RF 4.3 s): This branch integrates very slow modulations across the entire chunk, reflecting hormonal or thermoregulatory influences, which are thought to contribute to the VLF-like band [41]. Here, as the chunk is limited to 128 frames, the three RFs cover only the order-of-magnitude ranges rather than exact periods.
Outputs from the three MSTB branches are concatenated and passed through a SoftMax-based temporal attention pooling layer. This final step aggregates multi-scale information into a single, fixed-size 256-dimensional embedding ( h i   R D , where D = 2 56) for each chunk.

3.3.3. AttnScorer

The AttnScorer’s function is to evaluate each chunk embedding ( h i ) and assign it a single scalar relevance score. This module is directly inspired by the physiological principle that autonomic responses to emotional stimuli are often transient and sparsely distributed in time [42]. Its purpose is therefore to learn to identify these brief, emotionally salient moments.
Architecturally, it consists of a 2-layer MLP with GELU activation (Appendix B). The raw attention scores are adaptively normalized using σ-γ scaling (a running-std gain control), a mechanism analogous to biological sensory normalization [43]. These scores are then transformed using α-Entmax attention [18,44], encouraging a sparse selection of salient chunks, which mirrors the selective focus of biological attention mechanisms [20].

3.3.4. Gated Pooling

The Gated Pooling module aggregates the sequence of chunk embeddings ( h 1 ,     ,   h T ) into a single, noise-resilient session-level representation ( h p o o l e d ). Its design is a direct response to the noisy nature of rPPG signals, where even within an emotionally salient chunk, not all feature dimensions are equally reliable [8].
It implements a learned, content-aware aggregation via two parallel mechanisms [45]. First, it uses the sparse attention scores ( a i ) from the AttnScorer as temporal weights. Second, it computes a learned gate vector g i   R D that modulates each dimension via element-wise product ( g i     h i ), The final representation is computed as
h p o o l e d   =   i = 1 T α i g i     h i
This feature-level gating is pivotal, allowing the model to selectively amplify robust feature dimensions while suppressing those potentially corrupted by noise, a process inspired by neural gating and inhibition [46].

3.3.5. Auxiliary Components

These modules are active only during specific curriculum phases (Section 3.4) to support learning objectives. The ChunkProjection module (Phase 0) is an MLP head for normalized embeddings used by the Supervised Contrastive Loss [47]. The ChunkAuxClassifier module (Phase 1) is a classifier attached before Gated Pooling module, predicting session labels from individual chunks to pretrain the MTDE for local discrimination. It is used to initialize the Main Classifier in Phase 2.

3.3.6. Main Classifier

The Main Classifier receives the final aggregated session representation h p o o l e d vector and predicts final emotion labels (low/high valence/arousal) via fully connected layers.

3.4. Training Curriculum

Our training employs a three-phase curriculum learning strategy, a methodology inspired by the developmental principle that meaningful learning occurs by progressively increasing the difficulty of training samples [21]. This structured approach is conceptually analogous to human skill acquisition, moving from broad feature perception to fine-grained discrimination and finally to integrated decision-making [22,48]. It systematically addresses the challenges of noisy, temporally sparse, and weakly labeled time-series data to achieve stable and effective learning. The specific pipeline configuration and module activations during each phase are illustrated in Figure 2, Figure 3 and Figure 4. Detailed hyperparameters for each phase are provided in Appendix C.

3.4.1. Phase 0 (Epochs 0–14): Exploration and Representation Learning

  • Cognitive/Computational Link: This initial phase corresponds to “pervasive exploration” observed in human learning, where a system first captures a wide array of sensory inputs to build a general understanding of the feature space before focusing on specific tasks. This aligns with classic models of attention where an initial, broad orientation precedes focused analysis [49].
  • Objective: The primary objective is to train the MTDE and related components (AttnScorer and ChunkProjection) to produce robust and diverse embedding representations for individual temporal chunks. During this phase, the Gated Pooling module and the Main Classifier are not used for the primary loss computation.
  • Primary Losses: The total loss in Phase 0 is a combination of the Supervised Contrastive Loss and an Entropy Regularization Loss. The Supervised Contrastive Loss ( L s u p c o n ) [47] is applied to the normalized embeddings from the ChunkProjection. This loss encourages embeddings from chunks originating from the same session (sharing the same label) to be closer in the representation space while pushing embeddings from different sessions apart.
The training in this phase is guided by a combination of two loss functions. The primary objective is the Supervised Contrastive Loss ( L s u p c o n ), defined as
L s u p c o n P = i     I 1 P i p P i l o g e x p z i   · z p / τ a A i e x p z i   · z a / τ
Here, I is the set of anchor indices in the batch, A i is the set of all indices in the batch except i , P i is the set of indices of positive samples (same label) of i , z represents the normalized embedding vectors, and τ is the temperature parameter. To complement this, an Entropy Regularization Loss ( L e n t r o p y ) is applied to the AttnScorer’s internal SoftMax attention output. Weighted by λ e n t r o p y , this loss encourages a more uniform initial attention distribution, promoting a broader exploration of temporal features. This exploration is critical for preventing the model from prematurely converging on spurious patterns—a known risk related to the memorization capacity of deep networks, especially in the presence of noisy labels [50].
The overall loss for this phase combines both objectives.
L T o t a l =   L s u p c o n   +   λ e n t r o p y L e n t r o p y

3.4.2. Phase 1 (Epochs 15–29): Chunk-Level Discrimination and Attentional Refinement

  • Cognitive/Computational Link: This second phase corresponds to the establishment of selective attention. After the initial exploration, learning resources are focused on task-relevant signals. By using the AttnScorer and Focal Loss on only the Top-K chunks, our model progressively narrows its attentional focus, mimicking how neural systems learn to prioritize information-rich stimuli over time [51].
  • Objective: The objective is to significantly enhance the discriminative capacity of the individual chunk embeddings and to refine the AttnScorer’s ability to identify emotionally salient temporal segments. During this phase, the ChunkProjection module is frozen, while the MTDE, AttnScorer, and ChunkAuxClassifier are actively trained.
  • Primary Losses: The main objective is driven by a chunk-level classification task using the ChunkAuxClassifier. To handle class imbalance and focus on challenging examples, we employ Focal Loss [52] with γ = 2.0 (Appendix C):
The primary loss function for Phase 1 is the chunk-level Focal Loss ( L F o c a l = L c h u n k C E ), calculated as
L c h u n k C E = L F o c a l p t = α t 1 p t γ l o g p t
Here, we use a formulation where p t is the model’s estimated probability for the target class, α t is a class-balancing weight, and γ is the focusing parameter.
Critically, this loss is calculated only for the Top-K chunks selected based on the AttnScorer’s α-Entmax output. The Top-K ratio, K , is strategically annealed downwards as training progresses. This forces the model to progressively focus its discriminative learning resources on the most salient segments, a process that mimics how human attention narrows onto key details within a stimulus [53].
Furthermore, this phase prepares the model for the next stage of training. The session-level cross-entropy loss ( L s e s s i o n C E ) is scheduled to be introduced from epoch 25, with its weight gradually ramping up. While the parameters for session-level components are unfrozen to allow for preparatory fine-tuning, L s e s s i o n C E is not yet included in the total loss. This staged activation facilitates a smoother transition to Phase 2.
This staged activation strategy enables the model to begin adapting the session-level representation and pooling dynamics without prematurely influencing the optimization objective, thus facilitating a smoother transition to Phase 2 training. The overall loss for this phase is L T o t a l =   L c h u n k C E .

3.4.3. Phase 2 (Epochs ≥ 30): Session-Level Exploitation and Fine-Tuning

  • Physiological/Cognitive Link: The final phase is analogous to evidence integration for decision-making [54]. The system leverages its refined representations and focused attention to weigh and combine the most reliable evidence distributed over time to form a final, consolidated judgment [55].
  • Objective: The objective is to optimize the entire end-to-end pipeline for the final session-level emotion recognition task. The ChunkAuxClassifier is removed, and the Main Classifier is initialized with its weights. The MTDE, AttnScorer, Gated Pooling, and Main Classifier are all fine-tuned.
  • Primary Loss Function: The sole objective function in this phase is applying the session-level cross-entropy loss ( L s e s s i o n C E ) to the output of the Main Classifier based on the Gated Pooling session embedding.
In this phase, the model is optimized using a session-level cross-entropy loss:
L C E y ,   y ^ = c = 1 C y c l o g ( y ^ c )
Here, y is the one-hot encoded ground-truth label for the session, and y ^ is the predicted probability distribution from the Main Classifier over C classes.
To achieve optimal session-level performance, key parameters are carefully scheduled to exploit the attentional patterns learned in the previous phase. Specifically, the AttnScorer is fine-tuned at a reduced learning rate (see schedule in Appendix C). Simultaneously, the α g value for the α-Entmax transformation within the Gated Pooling module is annealed from 1.5 (at epoch 30) to 1.8 (at epoch 50). This increase in α g promotes higher sparsity in the temporal attention weights, further refining the model’s focus onto the most crucial temporal segments for the final prediction.

3.5. Evaluation Metrics

Model performance was quantitatively evaluated using standard metrics on the independent test set. These metrics were chosen to provide a comprehensive and robust assessment, particularly considering potential class imbalances.
Performance was measured using accuracy and the weighted F1-score. The weighted F1-score is particularly valuable in the presence of class imbalances as it accounts for performance on all classes weighted by their frequency, providing a more objective measure than simple accuracy or macro-averaged metrics in such scenarios.
  • Accuracy: This is defined as the proportion of correctly classified sessions out of the total number of sessions in the test set.
A c c u r a c y   =   N u m b e r   o f   C o r r e c t   P r e d i c t i o n s T o t a l   N u m b e r   o f   P r e d i c t i o n s
  • Weighted F1-score: This metric is calculated based on the precision ( P r e c i s i o n c ), recall ( R e c a l l c ), and F1-score ( F 1 c ) for each individual class c . The formulas for these class-specific metrics are
P r e c i s i o n c   =   T P c T P c   +   F P c
R e c a l l c   =   T P c T P c   +   F N c
F 1 c   =   2 · P r e c i s i o n c     Recall c Precision c   +   Recall c
The overall weighted F1-score is then computed as the average of the class F1-scores, weighted by the number of samples in each class ( N c ):
W e i g h t e d   F 1   =   c = 1 C N c N · F 1 c
Here, N is the total number of samples in the test set, and C is the number of classes. In addition to these primary metrics, we also analyze the confusion matrix to gain insights into the model’s performance across different classes and the types of errors made.

3.6. Baseline

To contextualize the performance of our proposed framework, we selected the CNN-LSTM model by Mellouk & Handouzi [23] as our primary baseline. This choice was based on its status as the most domain-equivalent benchmark: it is a recent deep learning work that, similar to our study, uses contactless PPG for emotion recognition with approximately 4-s segments on the MAHNOB-HCI dataset.
However, a direct comparison faces significant limitation. Because Mellouk & Handouzi [23] did not make their source code available and do not disclose their specific test partition, it is impossible to run their model on our exact data split. Therefore, while we utilize their reported results for a comparative analysis, their performance should be interpreted as indicative rather than directly comparable. Our study mitigates this challenge by rigorously detailing our data split (Section 3.1) and employing the weighted F1-score (Section 3.5) to provide a more robust assessment that accounts for potential class imbalance. This approach establishes a clearer and more reproducible benchmark for future work.
Furthermore, to ensure a comprehensive evaluation and robustly validate our specific architectural choices, we implemented and compared our framework against a diverse set of alternative architectures. As detailed in Appendix E, these include a standard Temporal Convolutional Network (TCN), a 1D-CNN+LSTM hybrid model, and variants of our own model with different aggregation (Bi-LSTM) and attention (SE block) mechanisms. These extensive comparisons provide a clear justification for our design choices over various alternatives.

3.7. Experimental Setup

All experiments were conducted on a system with Ubuntu 20.04, Python 3.8, and PyTorch 2.1.2 (+CUDA 12.1), running on a single NVIDIA RTX 4080 GPU. A consistent batch size of 8 was used for all training phases of the emotion recognition model. Each item within a batch consisted of the complete sequence of temporal chunks belonging to a single session, ensuring that chunks from different sessions were never mixed within a sequence.
To manage GPU memory during the initial data preparation stage, long sessions were processed by the rPPG extractor in smaller segments (e.g., 24 chunks at a time). The resulting BVP waveforms were then concatenated to reconstruct the full session sequence before being used as input for model training.
For optimization, we employed the AdamW optimizer with a CosineAnnealingLR schedule. The learning rates were set to 3 × 10 4 , 2 × 10 4 , and 1 × 10 4 for Phases 0, 1, and 2, respectively, with weight decay set to 1 × 10 4 in Phase 0 and 5 × 10 4 for the subsequent phases.
In terms of computational resources, our proposed emotion recognition modules are lightweight. The front-end PhysMamba extractor consumes approximately 9.8 GiB of VRAM, while our entire pipeline has a maximum GPU memory consumption of 14.5 GiB under our batch configuration. This indicates that our downstream modules (MTDE, attention, etc.) add only about 4.7 GiB to the baseline load. The complete model has a total of 197,892 trainable parameters (164,996 for inference). On the aforementioned hardware, the average training time was approximately 4.1 min per epoch, and the average inference time for a 2-min session was 0.66 s, confirming the model’s efficiency and suitability for practical applications.

4. Results

4.1. Main Results

As summarized in Table 3 and Table 4, the proposed end-to-end framework demonstrates competitive performance in arousal classification using only the temporal rPPG signal from the MAHNOB-HCI dataset under a subject-mixed evaluation protocol. Specifically, the model achieves an accuracy of 66.04% and a weighted F1-score of 61.97%.
When compared against a conventional CNN-LSTM baseline [23]—evaluated using their reported confusion matrix—our model yields consistent improvements across all relevant metrics:
  • Accuracy: 66.04% vs. 61.31%;
  • Positive-class F1-score: 74.29% vs. 50.96%;
  • Weighted F1-score: 61.97% vs. 59.46%.
These improvements validate the effectiveness of our architectural choices. Specifically, the MTDE is designed to be sensitive to the rapid waveform changes characteristic of sympathetic activation, while the sparse attentional chunk selection and feature-level Gated Pooling are adept at isolating these strong, transient physiological “events” from noisy segments. This event-driven design is particularly well-suited for capturing the discriminative temporal dynamics of arousal. The performance margin over the prior deep learning baseline affirms the merit of our tailored design for temporal-only physiological modeling. The corresponding confusion matrix of this result is provided in Appendix D (Table A5).
In contrast, and consistent with the known physiological limitations of this modality, valence classification remains a more complex and challenging task. Our model achieves a 62.26% accuracy and a 62.26% weighted F1-score, which are both substantially lower than those reported by the CNN-LSTM baseline (73.50% accuracy and 73.14% weighted F1-score calculated from their reported confusion matrix) [23]. The corresponding confusion matrix of this result is provided in Appendix D (Table A6).
This performance gap can be attributed to several fundamental factors:
  • Physiological limitations: Arousal is closely associated with sympathetic nervous system (SNS) activity, which often manifests as acute, high-magnitude changes in cardiovascular patterns [56]. In contrast, valence is linked to more complex and subtle interactions involving the parasympathetic nervous system (PNS) and cortical patterns, which are not as easily captured in peripheral signals [4,57].
  • Modality Constraints: By spatially averaging the facial region into a single 1D signal, our approach cannot access fine-grained spatial information. This is a notable limitation as recent studies suggest that regional facial blood flow patterns, such as subtle temperature changes around key facial areas, may hold valuable cues related to emotional valence [58].
  • Architectural Inductive Bias: Our framework’s event-driven design, particularly its reliance on sparse attention to identify transient, high-magnitude physiological events, introduces an inductive bias. This approach is highly effective for detecting the “spikes” characteristic of SNS-driven arousal responses but is inherently less suited to integrating the subtle, sustained, and context-dependent patterns that often characterize valence.
  • Data Imbalance: A notable class imbalance in valence labels may also contribute to biased predictions.
Collectively, these findings suggest that while our temporal rPPG modeling is potent for arousal recognition, effective valence modeling may necessitate either multimodal fusion or different architectural philosophies that prioritize long-term, contextual state integration over longer temporal windows.

4.2. Ablation Studies

To quantitatively assess the contributions of our key architectural components and the phased training strategy, we conducted a series of ablation studies. As our primary findings in Section 4.1 demonstrate that the proposed framework yields significant and meaningful improvements primarily in arousal classification, all subsequent ablation studies were centered on this task. This focused approach allows for a precise dissection of the components contributing to the model’s main success. Each study was designed to isolate the impact of a specific part of our framework on arousal recognition.

4.2.1. Impact of Temporal Aggregation Strategy

First, to isolate the contribution of our pooling mechanism, we compared its performance against simpler aggregation strategies on the arousal task. All models in this test used the same MTDE backbone and were trained with the full curriculum. As shown in Table 5, replacing Gated Pooling with standard attention or average pooling resulted in a significant performance degradation of over 15% in accuracy. This finding underscores that a sophisticated, learned aggregation strategy is critical for effectively handling the noisy and complex temporal features extracted from rPPG signals.

4.2.2. Validation of Multi-Scale Architecture in MTDE

To empirically validate our choice of a three-branch multi-scale architecture, we conducted an ablation study analyzing the impact of the number of parallel branches within the MTDE. We compared our proposed three-branch model against variants with one (a standard TCN), two, and four parallel branches. The detailed architecture of the TCN baseline is described in Appendix E. For the two-branch and four-branch variants, we used specific combinations of our predefined scales: the two-branch model combined our short- and long-scale branches, while the four-branch model added an additional medium-scale branch with a different dilation factor to test the impact of increased complexity.
The results presented in Table 6 reveal a clear trend that supports our three-branch design as the optimal configuration. The four-branch model showed the lowest performance among the multi-scale variants, suggesting that simply adding more branches leads to model over-parameterization and redundancy without providing additional discriminative information from the four-second chunks.
The TCN (single-branch) model, while outperforming some overly complex configurations, still fell short of our proposed model. This indicates that although it is a strong temporal model, a single receptive field is insufficient to capture the diverse range of physiological dynamics (from rapid morphological changes to slower oscillations) present in the BVP signal.
The two-branch model performed better than the single-branch TCN, confirming the benefit of a multi-scale approach. However, it did not surpass our proposed three-branch model, suggesting that incorporating the full spectrum of short-, medium-, and long-term dynamics is crucial for achieving the best performance.
These findings provide strong empirical support that our three-branch, physiologically aligned MTDE achieves the optimal trade-off between model complexity and the capacity to represent the multi-faceted temporal dynamics essential for emotion recognition.

4.2.3. Ablation Study on Pooling and Attention Mechanisms (Arousal)

We also evaluated the impact of each core component within our three-phase curriculum learning strategy. As presented in Table 7, the full curriculum achieved the highest performance, and removing any key mechanism resulted in a clear degradation, confirming that each phase provides an essential contribution to the final model performance.
The most significant performance drop occurred with direct training (Phase 2 only), highlighting the fundamental need for a structured curriculum when dealing with noisy and weakly labeled data. This outcome empirically validates our core premise that a progressive, physiologically inspired learning strategy, analogous to human skill acquisition, is crucial for robust learning from complex physiological signals [21,22,48]. Removing the exploratory representation learning of Phase 0 (w/o SupCon) also significantly impaired the model’s ability to form a useful feature space, underscoring the importance of contrastive pre-training in such tasks. This observation aligns with the “pervasive exploration” principle in human learning, where an initial broad understanding of the feature space is foundational before focused analysis can occur [49].
Interestingly, the model trained without Top-K annealing yielded a slightly higher weighted F1-score despite its lower accuracy. A closer analysis revealed that this is not merely a statistical artifact of class imbalance but a direct consequence of the interplay between our curriculum strategy and the physiological nature of arousal. High arousal often manifests as strong, transient “events” in the BVP signal (e.g., sympathetic spikes), providing a clear and highly discriminative learning signal. Our Top-K annealing mechanism, designed to progressively focus on the most informative segments, naturally learns to prioritize these unambiguous high-arousal events. This transforms the model into a highly specialized “high-arousal detector,” achieving perfect recall for that class. This specialization is a direct computational manifestation of “selective attention,” where the model learns to prioritize information-rich stimuli akin to how neural systems focus on key details [51,53]. However, this specialization comes at the cost of a prediction bias as the model becomes more likely to misclassify the subtler, less event-like patterns of low-arousal states. The model without this refinement mechanism is less specialized, leading to a more balanced, albeit less accurate, performance across classes.
This finding provides a deep insight: our curriculum is highly effective at forcing the model to learn the most salient physiological features, but this very effectiveness can lead to a predictable bias when one class (high arousal) presents a much clearer, more “learnable” signal than the other.

4.3. Qualitative Analysis of Sparse Attention

The effectiveness and interpretability of our sparse attention mechanism are paramount, particularly its ability to prioritize physiologically meaningful regions [59]. We conducted a detailed visualization of the attention weights from our final model (with α-Entmax) and compared it against a standard attention model (with SoftMax) overlaid on a representative ECG signal from a random subject’s random session (Figure 5). This analysis provides direct insights into how our model’s attention aligns with critical physiological events.
As illustrated in Figure 5b, the standard Softmax attention produces a diffuse distribution across the temporal chunks, assigning moderate weights broadly without a clear, specific focus. This behavior suggests a lack of selective attention, where the model processes all chunks with similar emphasis, potentially diluting the significance of critical events. Quantitatively, this diffuse characteristic is reflected by its higher entropy value of 2.92. In stark contrast, Figure 5c clearly demonstrates that our α-Entmax attention learns a highly sparse distribution, evidenced by its significantly lower entropy value of 1.14, signifying a strong and concentrated allocation of attentional weight on a select few temporal chunks.
Figure 5. Qualitative analysis of sparse attention mechanism on ECG signal. This figure presents a comparative qualitative analysis of the attention distributions from a standard SoftMax model and our proposed α-Entmax sparse attention model on a representative ECG signal from a random subject’s random session. (a) displays the ECG signal with physiologically meaningful R-peaks highlighted. (b) shows the diffuse attention distribution of SoftMax (Entropy: 2.92). (c) illustrates the highly sparse attention distribution of α-Entmax (Entropy: 1.14). (d) overlays both attention distributions on the ECG signal, demonstrating how α-Entmax selectively focuses on specific physiological events, particularly R-peak clusters.
Figure 5. Qualitative analysis of sparse attention mechanism on ECG signal. This figure presents a comparative qualitative analysis of the attention distributions from a standard SoftMax model and our proposed α-Entmax sparse attention model on a representative ECG signal from a random subject’s random session. (a) displays the ECG signal with physiologically meaningful R-peaks highlighted. (b) shows the diffuse attention distribution of SoftMax (Entropy: 2.92). (c) illustrates the highly sparse attention distribution of α-Entmax (Entropy: 1.14). (d) overlays both attention distributions on the ECG signal, demonstrating how α-Entmax selectively focuses on specific physiological events, particularly R-peak clusters.
Sensors 25 03995 g005
Crucially, Figure 5d distinctly shows that α-Entmax attention correctly identifies and assigns almost all of its attentional weight to specific temporal chunks that robustly coincide with significant physiological events in the ECG signal—specifically, clusters of high-frequency R-peaks. These R-peaks are widely recognized as key indicators of cardiac electrical activity and can signify sudden sympathetic arousal responses, making them physiologically meaningful moments [59,60]. This visual comparison provides a compelling demonstration of α-Entmax’s ability to filter out less relevant information and pinpoint critical temporal chunks, directly supporting its claim to mimic biological selective attention.
This visualization provides strong evidence that our sparse attention mechanism profoundly enhances model interpretability. By learning to selectively focus on physiologically meaningful moments through a carefully designed training curriculum, α-Entmax substantiates its interpretability and alignment with physiological principles, addressing critical considerations for robust scientific claims.

5. Discussion

Our proposed framework demonstrates promising capabilities in recognizing emotions exclusively from temporal remote photoplethysmography (rPPG) signals, achieving competitive arousal classification performance (66.04% accuracy; 61.97% weighted F1-score) and outperforming the CNN-LSTM baseline [23]. This superior arousal performance stems from our physiologically inspired temporal modeling and neuro-cognitively motivated learning strategy. The Multi-scale Temporal Dynamics Encoder (MTDE) effectively captures ANS modulations across timescales [4,15,39,40,41], showing that it is adept at detecting rapid pulse morphology changes indicative of sympathetic activation. The adaptive sparse α-Entmax attention [17] mirrors human selective attention [18,19,43,51,53], precisely identifying transient, salient physiological events like R-peak clusters (Figure 5) [59,60]. Robust aggregation via Gated Pooling [20,45], inspired by neural gating principles [46], further mitigates noise. Critically, our staged curriculum learning strategy [21], analogous to human skill acquisition [22,48,49], ensured stable and effective learning by systematically addressing data challenges [50]. Our ablation studies (Section 4.2) validate each component’s significant contribution to these gains.
Conversely, valence recognition remains significantly challenging with temporal rPPG signals. Our model achieved lower valence performance (62.26% accuracy; 62.26% weighted F1-score) compared to Mellouk & Handouzi’s baseline [23] (accuracy: 73.50%; weighted F1-score: 73.14%). This discrepancy highlights fundamental physiological limitations and inherent methodological constraints in valence detection using only temporal, spatially averaged rPPG data.
Physiologically, arousal is largely driven by sympathetic nervous system (SNS) activation [4,56], manifesting as acute, high-magnitude cardiovascular changes. These “event-like” shifts are well-captured by our framework’s design. In contrast, valence involves more intricate and subtle physiological responses, including a complex interplay between the sympathetic and parasympathetic nervous systems (PNS) [4,57] and higher-order cortical processing not directly reflected in peripheral cardiovascular changes. Positive valence often involves parasympathetic activation, while negative valence can manifest as nuanced patterns of co-activation or withdrawal, less pronounced than arousal’s overt responses.
Furthermore, subtle spatial variations in facial blood flow patterns [58] may carry valence-related cues, but our single, unidimensional temporal rPPG signal inherently loses this information. Cognitively, arousal often correlates with immediate, attention-grabbing physiological “events,” whereas valence requires integrating more subtle, sustained, and context-dependent patterns over longer temporal windows [42,54,55]. Our framework’s strong inductive bias towards transient, high-magnitude physiological events, effective for arousal, is thus less suited for these nuanced valence patterns. These findings strongly suggest that while temporal rPPG robustly reflects general ANS arousal, recognizing nuanced affective dimensions like valence likely necessitates multimodal integration (e.g., facial expressions and electrodermal activity) or advanced spatial–temporal rPPG analysis.
Regarding practical applicability, our approach demonstrates notable computational efficiency. Considering that it achieves an inference speed of approximately 0.66 s for a two-minute video segment on an NVIDIA RTX 4080 GPU, our model operates faster than real time. Its compact size (164,996 parameters) further facilitates deployment in resource-constrained scenarios (e.g., mobile devices and edge computing), significantly enhancing the viability of our temporal-only rPPG approach for unobtrusive affective computing where dedicated sensors or multiple modalities may not be feasible. This highlights the potential of specialized temporal models even within unimodal signal limitations.

5.1. Limitations

Our study, despite advancing temporal rPPG-based emotion recognition, presents several limitations informing future research. Firstly, our evaluation is confined to the MAHNOB-HCI dataset. This was a deliberate choice to rigorously investigate the capabilities and limits of temporal-only rPPG in a controlled environment with sufficient per-subject session data for deep learning. However, as reviewers noted, its limited scale (527 sessions) and diversity may constrain generalization to unseen subjects or broader contexts.
Secondly, our restriction to temporal-only rPPG, the scope of this foundational investigation, inherently excludes valuable spatial rPPG information or complementary affective cues from other modalities. As evidenced by lower valence performance (Section 5), this highlights the constraints of relying solely on a spatially averaged temporal pulse signal for nuanced affective dimensions.
A further limitation involves the explicit validation of our physiologically inspired mechanisms. While qualitative analysis (Section 4.3) and ablation studies (Section 4.2) demonstrated that our attention mechanisms prioritize physiologically meaningful regions and multi-scale encoding’s effectiveness, deeper statistical correlation analyses with established physiological markers (e.g., specific HRV components) were beyond this study’s initial scope, which focused on implicit learning from the BVP waveform. Our current model thus abstracts and emulates neuro-physiological principles, and further work is needed to fully mimic and quantitatively validate these complex biological processes within its learned representations.

5.2. Future Work

Addressing identified limitations, several clear avenues exist for future research building upon our temporal processing foundations.
Firstly, to enhance generalization and robustness, validating our framework on larger, more diverse datasets (e.g., DEAP and WESAD) is crucial. This includes exploring scenarios with limited per-subject data but broad subject ranges, assessing adaptability to new individuals even without extensive calibration. We aim to develop personalization strategies (e.g., few-shot subject calibration, meta-learning, and transfer learning) for robust performance across varied populations and conditions.
Secondly, to improve nuanced affective dimensions like valence, future work should expand beyond unimodal temporal rPPG. This includes integrating spatial–temporal rPPG analyses preserving regional blood flow patterns and multimodal fusion with complementary affective cues (e.g., facial expressions and electrodermal activity). Evaluating our model on alternative emotional elicitation methods, such as large-scale VR-based datasets, would further strengthen validation and assess domain generalization.
Finally, building upon our current work that abstracts and draws inspiration from neuro-physiological mechanisms, a critical future direction is to develop models that more deeply mimic and quantitatively validate these biological processes. This involves conducting deeper analyses into learned representations, potentially exploring their statistical correlation with specific, well-defined physiological markers beyond HR estimation. Such research would provide critical insights into which physiological dynamics and temporal patterns best indicate different emotional states, fostering more interpretable, scientifically grounded affective computing models that genuinely replicate biological information processing rather than merely using biological analogies.
Collectively, our results establish foundational benchmarks and methodological insights for future advancements in unimodal physiological emotion recognition, clearly defining both the capabilities and inherent limitations of relying solely on temporal-only rPPG signals. Our approach’s comprehensive physiological and cognitive grounding, combined with rigorous evaluation, ensures robust, interpretable, and applicable outcomes, advancing the state of the art in this challenging area.

6. Conclusions

We introduced a physiologically inspired deep learning framework for recognizing emotional states exclusively from temporal remote photoplethysmography (rPPG). Our approach systematically addresses critical limitations—temporal sparsity, signal noise, and weak labeling—through the Multi-scale Temporal Dynamics Encoder (MTDE), adaptive sparse attention, Gated Temporal Pooling, and a structured three-phase curriculum learning strategy. Empirical evaluation confirmed competitive performance in arousal classification (66.04% accuracy; 61.97% weighted F1-score), surpassing previous deep learning baselines. Conversely, lower performance in valence classification (62.26% accuracy) reveals fundamental physiological constraints in using solely temporal cardiovascular signals, clearly demarcating the capability boundaries of unimodal rPPG signals.
These results establish robust methodological benchmarks and highlight promising directions for future exploration: incorporating spatially resolved rPPG analysis or multimodal integration could significantly enhance nuanced emotional inference. This study provides critical foundational insights and clear guidelines to advance affective computing towards more accurate, reliable, and interpretable physiological emotion recognition.

Author Contributions

Conceptualization, C.L., H.L. and M.W.; methodology, C.L.; software, C.L.; validation, C.L. and H.L.; formal analysis, C.L.; investigation, C.L.; resources, M.W.; data curation, C.L.; writing—original draft preparation, C.L.; writing—review and editing, H.L. and M.W.; visualization, C.L.; supervision, M.W.; project administration, M.W.; funding acquisition, M.W. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by an Electronics and Telecommunications Research Institute (ETRI) grant funded by the Korean government [25ZB1100, Core Technology Research for Self-Improving Integrated Artificial Intelligence System] and the Technology Innovation Program (or Industrial Strategic Technology Development Program-Knowledge service industrial technology development) (20023561, Development of learncation service for active senior life planning and self-development support) funded by the Ministry of Trade Industry & Energy (MOTIE, Republic of Korea).

Institutional Review Board Statement

This study uses the publicly released MAHNOB-HCI dataset, which contains de-identified recordings collected under the authors’ original institutional approval.

Informed Consent Statement

Not applicable.

Data Availability Statement

The source code and trained weights are available at https://github.com/LeeChangmin0310/ReMOTION-Temporal (accessed on 7 May 2025). The MAHNOB-HCI dataset analyzed in this study can be obtained from the original authors under the CC-BY-NC-SA license. The repository includes the training scripts, inference demo, and the raw split lists used for reproducibility. Experimental environment details are also provided in the repository’s README, consistent with the setup described in Section 3.7.

Acknowledgments

The author thank the Emotion Contents Technology Research Center members for their feedback on the experimental design. During the preparation of this work, the author(s) used ChatGPT (version 3.5, OpenAI) and Gemini (Google) for assistance with English language editing and refining scientific expressions. Additionally, GPT-4o (OpenAI) was used to assist with developing and debugging the software code for the experiments. The authors have reviewed and edited the output and take full responsibility for the content of this publication.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
AccAccuracy
ANSAutonomic Nervous System
AttnScorerAttention Scorer
BVPBlood Volume Pulse
CEConfusion Matrix
CMConvolutional Neural Network
EDAElectrodermal Activity
HRVHeart Rate Variability
MILMultiple Instance Learning
MTDEMulti-scale Temporal Dynamics Encoder
rPPGRemote Photoplethysmography
WF1Weighted F1-Score

Appendix A. Architecture Details of MTDE

Appendix A.1. Architecture Overview

The MTDE encodes each 128-frame (4 s) rPPG chunk into a 256-dimensional embedding through two stages:
  • SlimStem: Two Conv1D layers (kernel = 5, then 3 with stride = 2) for initial noise reduction and temporal downsampling (T = 128 → 64).
  • MultiScaleTemporalBlock (MSTB): Three parallel branches with different kernel sizes and dilations are used to model short-, mid-, and long-range temporal dynamics.

Appendix A.2. Physiological Rationale and Receptive Fields

Table A1. MSTB’s parameters per branch.
Table A1. MSTB’s parameters per branch.
BranchKernelDilationEffective RFApprox. DurationPhysiological Role
Short336~0.2 sParasympathetic (vagal) fluctuations, analogous to HF band of HRV.
Medium5866~2.2 sSympathetic–parasympathetic interplay (e.g., baroreflex), analogous to LF band.
Long332129~4.3 sVery slow modulations (e.g., hormonal and thermoregulatory), analogous to VLF band.
Here, the receptive field can be calculated by R F = k 1 × d + 1 .

Appendix A.3. Pooling Layer: SoftMax Pool

SoftMax over time across temporal steps:
w = s o f t m a x W a t t n x ,     h   =   t w t x t
This results in a single (B, D) embedding per chunk. No gating is applied in this layer.

Appendix B. Attention Modules: AttnScorer and Gated Pooling

Appendix B.1. AttnScorer: Phase-Aware Attention Scoring

Appendix B.1.1. Architecture

  • Two-layer MLP: Linear (D, D/2) → GELU → Linear (D/2, 1);
  • Zero-mean score normalization: s i     s i m e a n ( s ) ;
  • σ γ scaling with EMA-based adjustment:
γ   =   σ t a r g e t σ r u n + ε     [ 0.5 ,   2.0 ]
  • Raw score scaling: s i s c a l e d = γ s i .
Here, σ r u n is the running standard deviation of the raw attention scores, calculated via an exponential moving average. σ t a r g e t is a scheduled hyperparameter representing the target standard deviation (e.g., 0.8 for early epochs and 1.2 for later epochs). The value of γ is clamped to the range [0.5, 2.0].

Appendix B.1.2. Phase-Dependent Attention Mechanism

Table A2. Attention type per phase.
Table A2. Attention type per phase.
PhaseEpoch RangeAttention TypeNotes
00–14Softmax Entourage diversity
115–29 α -Entmax ()Sharp, sparse, differentiable Top-K
2≥30Raw scores onlyPassed to Gated Pooling

Appendix B.1.3. Entmax Scheduling During Phase 1

α e = 1.0 + 0.3   · e 15 5 , i f   15     e < 20 1.3 + 0.3   · e 20 5 , i f   20     e < 25 1.7 , i f   25     e < 30

Appendix B.2. Gated Pooling: Sparse Temporal Aggregation

Receives chunk embeddings h i   R D and raw attention scores. In Phase 2, the attention scores are transformed using α g -Entmax to produce temporal weights:
α i = e n t m a x α g r a w _ s c o r e i , g i   = σ M L P h i
The session-level pooled embedding is computed as
h p o o l e d   =   i = 1 T α i g i     h i
where denotes element-wise multiplication.
α g is scheduled during Phase 2 as
α g ( e ) = 1.7 + 0.3   · m i n e 30 19 ,   1.0 ,   w h e r e     e > 30
This dual mechanism (Temporal weight × Feature gate) reflects neural inhibition, enabling selective suppression of irrelevant dimensions even within salient chunks.

Appendix C. Phase-Wise Training Schedule

Table A3. Epoch-based curriculum strategy.
Table A3. Epoch-based curriculum strategy.
PhaseEpochsObjectiveActive Modules
00–14Embedding diversity (SupCon)MTDE, AttnScorer, ChunkProjection
115–29Chunk-level discriminationMTDE, AttnScorer, ChunkAuxClassifier, Gated Pooling (E ≥ 25)
230–49Session-level classificationGated Pooling, Classifier
Table A4. Hyperparameter scheduling.
Table A4. Hyperparameter scheduling.
EpochTop-K Ratio SupCon   λ CE   λ λ e n t r o p y τ (Temp) α (AttnScorer) α g (Gated)
00.01.000.00.11.2
140.00.440.00.10.7
150.60.00.50.01.01.0
250.30.00.50.01.01.7start = 1.7
300.00.70.01.0raw only1.7 → 2.0
500.01.00.01.0raw only2.0

Appendix D. Confusion Matrices

This appendix provides the confusion matrices for arousal and valence classification results from the final proposed model evaluated on the MAHNOB-HCI test set.
Table A5. Confusion matrix—arousal classification (final model).
Table A5. Confusion matrix—arousal classification (final model).
Predicted LowPredicted High
Actual Low918
Actual High026
  • Accuracy: 66.04%;
  • Weighted F1-score: 61.97%;
  • The model shows strong sensitivity to high arousal states (recall: 100%), with most misclassifications occurring in the low-arousal category.
Table A6. Confusion matrix—valence classification (final model).
Table A6. Confusion matrix—valence classification (final model).
Predicted LowPredicted High
Actual Low1310
Actual High1020
  • Accuracy: 62.26%;
  • Weighted F1-score: 62.26%;
  • The model demonstrates relatively balanced performance but reveals confusion between low and high valence categories, indicating the nuanced nature of valence detection from unimodal temporal signals.

Appendix E. Confusion Matrices

To provide a comprehensive evaluation and validate our architectural choices, we compared our proposed framework against several strong alternative architectures. Each alternative was designed to test a specific component of our model, such as the feature extractor, temporal aggregator, or attention mechanism. For a fair comparison, all models were trained using the same three-phase curriculum on the arousal classification task.

Appendix E.1. Description of Comparative Models

  • Temporal Convolutional Network (TCN): A standard TCN with a stack of dilated causal convolutions was implemented to serve as a strong, non-recurrent backbone baseline. The TCN consisted of two residual blocks, each with two layers of dilated causal convolutions (kernel size = 7, 256 channels), and dilation factors following exponential growth (1, 2, 4, 8).
  • 1D-CNN + LSTM: This is a classic hybrid architecture combining a simple 1D-CNN feature extractor with an LSTM layer to model sequential dependencies.
  • Multi-scale + Bi-LSTM: This model replaces the final attention pooling layer of our MTDE with a Bi-directional LSTM to evaluate a standard recurrent aggregation method.
  • Multi-scale + SE block: This model integrates Squeeze-and-Excitation (SE) blocks into our MTDE to assess the effect of a channel-wise attention mechanism as an alternative to our temporal attention.

Appendix E.2. Result and Analysis

The results of the comparative evaluation are summarized in Table A7.
Table A7. Performance comparison with alternative backbones (arousal classification).
Table A7. Performance comparison with alternative backbones (arousal classification).
Comparison FocusBackbone/VariantAccuracy (%)Weighted F1 (%)
Ours (3-branch MTDE)Ours (3-branch MTDE)66.0461.97
Backbone ComparisonStandard TCN58.4952.05
1D-CNN + LSTM58.4952.69
Aggregation ComparisonMulti-scale + Bi-LSTM52.8346.34
Attention ComparisonMulti-scale + SE block60.3852.50
The results consistently demonstrate that our proposed MTDE-based architecture outperforms all tested alternatives. A detailed analysis for each comparison is presented as follows:
  • Backbone Analysis (MTDE vs. TCN, 1D-CNN+LSTM): The proposed MTDE outperforms both the standard TCN and the classic hybrid model. This suggests that for this specific task, the physiologically aligned receptive fields of the MTDE are more effective at capturing the relevant, multi-scale dynamics of the BVP waveform than a generic temporal model like TCN. The relatively lower performance of the LSTM-based variant may be attributed to the nature of our input data. Recurrent models are designed for long-range dependencies, but in our framework, which processes short, 4-s chunks, the most critical emotional cues are embedded in the local waveform morphology. In such cases, the added complexity of a recurrent layer may not provide a significant advantage.
  • Aggregation and Attention Analysis (vs. Bi-LSTM, SE block): Our model also outperforms the variants with a Bi-LSTM aggregator or an SE block. This suggests that our combined temporal attention (AttnScorer) and feature-gating (Gated Pooling) mechanism is more effective than a standard recurrent aggregator. The underperformance of the SE block variant further implies that for this task, identifying when an emotionally salient event occurs (temporal attention) is more critical than re-weighting which feature channels are active (channel-wise attention).
  • Note on Transformer-based Models: While Transformer architectures are powerful, we deliberately did not include them as a primary baseline for two reasons. First, Transformers are notoriously data-hungry and are prone to overfitting on datasets of limited scale and session diversity like MAHNOB-HCI. Second, their core strength lies in modeling very long-range global dependencies, which, as observed with the LSTM variants, may be suboptimal for our event-driven, short-chunk analysis. As our ablation studies suggest, for this specific problem context, increasing architectural complexity does not necessarily guarantee better performance.
Overall, these extensive comparisons validate that our tailored, physiology-aligned, and lightweight architecture provides the most effective solution for temporal rPPG-based emotion recognition under the given data constraints.

References

  1. Picard, R.W. Affective Computing; MIT Press: Cambridge, MA, USA, 2000; ISBN 978-0-262-66115-7. [Google Scholar]
  2. Saffaryazdi, N.; Wasim, S.T.; Dileep, K.; Nia, A.F.; Nanayakkara, S.; Broadbent, E.; Billinghurst, M. Using Facial Micro-Expressions in Combination With EEG and Physiological Signals for Emotion Recognition. Front. Psychol. 2022, 13, 864047. [Google Scholar] [CrossRef] [PubMed]
  3. Mauss, I.B.; Robinson, M.D. Measures of Emotion: A Review. Cogn. Emot. 2009, 23, 209–237. [Google Scholar] [CrossRef] [PubMed]
  4. Kreibig, S.D. Autonomic Nervous System Activity in Emotion: A Review. Biol. Psychol. 2010, 84, 394–421. [Google Scholar] [CrossRef] [PubMed]
  5. Shaffer, F.; Ginsberg, J.P. An Overview of Heart Rate Variability Metrics and Norms. Front. Public Health 2017, 5, 258. [Google Scholar] [CrossRef]
  6. Yu, S.-N.; Lin, I.-M.; Wang, S.-Y.; Hou, Y.-C.; Yao, S.-P.; Lee, C.-Y.; Chang, C.-J.; Chu, C.-S.; Lin, T.-H. Affective Computing Based on Morphological Features of Photoplethysmography for Patients with Hypertension. Sensors 2022, 22, 8771. [Google Scholar] [CrossRef]
  7. Sanchez, E.; Tellamekala, M.K.; Valstar, M.; Tzimiropoulos, G. Affective processes: Stochastic modelling of temporal context for emotion and facial expression recognition. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021; pp. 9070–9080. [Google Scholar]
  8. Wang, W.; den Brinker, A.C.; Stuijk, S.; de Haan, G. Algorithmic Principles of Remote PPG. IEEE Trans. Biomed. Eng. 2017, 64, 1479–1491. [Google Scholar] [CrossRef]
  9. Pirzada, P.; Wilde, A.; Doherty, G.H.; Harris-Birtill, D. Remote photoplethysmography (rPPG): A state-of-the-art review. bioRxiv 2023. [Google Scholar] [CrossRef]
  10. Dietterich, T.G.; Lathrop, R.H.; Lozano-Pérez, T. Solving the Multiple Instance Problem with Axis-Parallel Rectangles. Artif. Intell. 1997, 89, 31–71. [Google Scholar] [CrossRef]
  11. Xie, Y.; Yu, Z.; Wu, B.; Xie, W.; Shen, L. SFDA-rPPG: Source-free domain adaptive remote physiological measurement with spatio-temporal consistency. arXiv 2024, arXiv:2401.01234. [Google Scholar]
  12. Russell, J.A. A Circumplex Model of Affect. J. Personal. Soc. Psychol. 1980, 39, 1161–1178. [Google Scholar] [CrossRef]
  13. Jemioło, P.; Storman, D.; Mamica, M.; Szymkowski, M.; Żabicka, W.; Wojtaszek-Główka, M.; Ligęza, A. Datasets for Automated Affect and Emotion Recognition from Cardiovascular Signals Using Artificial Intelligence—A Systematic Review. Sensors 2022, 22, 2538. [Google Scholar] [CrossRef] [PubMed]
  14. A Survey on Physiological Signal-Based Emotion Recognition. Available online: https://www.mdpi.com/2306-5354/9/11/688 (accessed on 14 June 2025).
  15. Akselrod, S.; Gordon, D.; Ubel, F.A.; Shannon, D.C.; Berger, A.C.; Cohen, R.J. Power Spectrum Analysis of Heart Rate Fluctuation: A Quantitative Probe of Beat-to-Beat Cardiovascular Control. Science 1981, 213, 220–222. [Google Scholar] [CrossRef] [PubMed]
  16. Vu, T.; Huynh, V.T.; Kim, S.-H. Multi-scale transformer-based network for emotion recognition from multi-physiological signals. In Lecture Notes in Computer Science; Springer: Cham, Switzerland, 2023; Volume 14408, pp. 113–122. [Google Scholar]
  17. Martins, A.F.T.; Farinhas, A.; Treviso, M.; Niculae, V.; Aguiar, P.M.Q.; Figueiredo, M.A.T. Sparse and Continuous Attention Mechanisms. arXiv 2020, arXiv:2006.07214. [Google Scholar]
  18. Corbetta, M.; Shulman, G.L. Control of Goal-Directed and Stimulus-Driven Attention in the Brain. Nat. Rev. Neurosci. 2002, 3, 201–215. [Google Scholar] [CrossRef]
  19. Briggs, F.; Mangun, G.R.; Usrey, W.M. Attention Enhances Synaptic Efficacy and the Signal-to-Noise Ratio in Neural Circuits. Nature 2013, 499, 476–480. [Google Scholar] [CrossRef]
  20. Lee, C.-Y.; Gallagher, P.; Tu, Z. Generalizing Pooling Functions in CNNs: Mixed, Gated, and Tree. IEEE Trans. Pattern Anal. Mach. Intell. 2018, 40, 863–875. [Google Scholar] [CrossRef]
  21. Bengio, Y.; Louradour, J.; Collobert, R.; Weston, J. Curriculum Learning. In Proceedings of the 26th Annual International Conference on Machine Learning, Montreal, QC, Canada, 14–18 June 2009; ACM: Montreal, QC, Canada, 2009; pp. 41–48. [Google Scholar]
  22. Elman, J.L. Learning and Development in Neural Networks: The Importance of Starting Small. Cognition 1993, 48, 71–99. [Google Scholar] [CrossRef]
  23. Mellouk, W.; Handouzi, W. CNN-LSTM for Automatic Emotion Recognition Using Contactless Photoplythesmographic Signals. Biomed. Signal Process. Control 2023, 85, 104907. [Google Scholar] [CrossRef]
  24. Lang, P.J.; Bradley, M.M. The Affect System Has Parallel and Integrative Processing Components. Available online: https://www.researchgate.net/publication/232564098_The_Affect_System_Has_Parallel_and_Integrative_Processing_Components (accessed on 14 June 2025).
  25. Zhou, K.; Schinle, M.; Stork, W. Dimensional Emotion Recognition from Camera-Based PRV Features. Methods 2023, 218, 224–232. [Google Scholar] [CrossRef]
  26. Talala, S.; Shvimmer, S.; Simhon, R.; Gilead, M.; Yitzhaky, Y. Emotion Classification Based on Pulsatile Images Extracted from Short Facial Videos via Deep Learning. Sensors 2024, 24, 2620. [Google Scholar] [CrossRef]
  27. Adisa, O.; Akinduyite, O.; Badeji-Ajisafe, B. A comprehensive review of remote photoplethysmography techniques for accurate and robust heart-rate monitoring. In Proceedings of the IEEE 5th International Conference on Electro-Computing Technologies for Humanity (NIGERCON), Minna, Nigeria, 23–25 January 2024; pp. 1–5. [Google Scholar]
  28. Soleymani, M.; Lichtenauer, J.; Pun, T.; Pantic, M. A Multimodal Database for Affect Recognition and Implicit Tagging. IEEE Trans. Affect. Comput. 2012, 3, 42–55. [Google Scholar] [CrossRef]
  29. Unke, O.T.; Meuwly, M. PhysNet: A Neural Network for Predicting Energies, Forces, Dipole Moments and Partial Charges. J. Chem. Theory Comput. 2019, 15, 3678–3693. [Google Scholar] [CrossRef]
  30. Yu, Z.; Shen, Y.; Shi, J.; Zhao, H.; Torr, P.; Zhao, G. PhysFormer: Facial Video-Based Physiological Measurement with Temporal Difference Transformer. In Proceedings of the 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18–24 June 2022; IEEE: New Orleans, LA, USA, 2022; pp. 4176–4186. [Google Scholar]
  31. Zou, B.; Guo, Z.; Chen, J.; Zhuo, J.; Huang, W.; Ma, H. RhythmFormer: Extracting Patterned rPPG Signals Based on Periodic Sparse Attention. arXiv 2025, arXiv:2402.12788. [Google Scholar] [CrossRef]
  32. Luo, C.; Xie, Y.; Yu, Z. PhysMamba: Efficient Remote Physiological Measurement with SlowFast Temporal Difference Mamba. arXiv 2024, arXiv:2409.12031. [Google Scholar]
  33. Lampier, L.C.; Caldeira, E.; Delisle-Rodriguez, D.; Floriano, A.; Bastos-Filho, T.F. A Preliminary Approach to Identify Arousal and Valence Using Remote Photoplethysmography. In Proceedings of the XXVII Brazilian Congress on Biomedical Engineering; Springer International Publishing: Cham, Switzerland, 2022; Volume 83, pp. 1659–1666. [Google Scholar]
  34. Pan, J.; Tompkins, W.J. A Real-Time QRS Detection Algorithm. IEEE Trans. Biomed. Eng. 1985, 32, 230–236. [Google Scholar] [CrossRef]
  35. Bobbia, S.; Macwan, R.; Benezeth, Y.; Mansouri, A.; Dubois, J. Unsupervised Skin Tissue Segmentation for Remote Photoplethysmography. Pattern Recognit. Lett. 2019, 124, 82–90. [Google Scholar] [CrossRef]
  36. Rinella, S.; Massimino, S.; Fallica, P.G.; Giacobbe, A.; Donato, N.; Coco, M.; Neri, G.; Parenti, R.; Perciavalle, V.; Conoci, S. Emotion Recognition: Photoplethysmography and Electrocardiography in Comparison. Biosensors 2022, 12, 811. [Google Scholar] [CrossRef]
  37. Zhu, Z.; Wang, X.; Xu, Y.; Chen, W.; Zheng, J.; Chen, S.; Chen, H. An Emotion Recognition Method Based on Frequency-Domain Features of PPG. Front. Physiol. 2025, 16. [Google Scholar] [CrossRef]
  38. Liu, X.; Narayanswamy, G.; Paruchuri, A.; Zhang, X.; Tang, J.; Zhang, Y.; Sengupta, R.; Patel, S.; Wang, Y.; McDuff, D. rPPG-Toolbox: Deep Remote PPG Toolbox; GitHub Repository: 2023. Available online: https://github.com/ubicomplab/rPPG-Toolbox (accessed on 15 January 2025).
  39. La Fountaine, M.F.; Toda, M.; Testa, A.J.; Hill-Lombardi, V. Autonomic Nervous System Responses to Concussion: Arterial Pulse Contour Analysis. Front. Neurol. 2016, 7, 13. [Google Scholar] [CrossRef]
  40. Berntson, G.G.; Cacioppo, J.T.; Quigley, K.S. Autonomic Determinism: The Modes of Autonomic Control, the Doctrine of Autonomic Space, and the Laws of Autonomic Constraint. Psychol. Rev. 1991, 98, 459–487. [Google Scholar] [CrossRef]
  41. Task Force of the European Society of Cardiology and the North American Society of Pacing and Electrophysiology. Heart-rate variability: Standards of measurement, physiological interpretation and clinical use. Circulation 1996, 93, 1043–1065. [Google Scholar] [CrossRef]
  42. Bradley, M.M.; Lang, P.J. Measuring Emotion: The Self-Assessment Manikin and the Semantic Differential. J. Behav. Ther. Exp. Psychiatry 1994, 25, 49–59. [Google Scholar] [CrossRef] [PubMed]
  43. Reynolds, J.H.; Heeger, D.J. The Normalization Model of Attention. Neuron 2009, 61, 168–185. [Google Scholar] [CrossRef] [PubMed]
  44. Peters, B.; Niculae, V.; Martins, A.F.T. Sparse Sequence-to-Sequence Models. Available online: https://arxiv.org/abs/1905.05702v2 (accessed on 15 June 2025).
  45. O’Reilly, R.C.; Munakata, Y. Computational Explorations in Cognitive Neuroscience; MIT Press: Cambridge, MA, USA, 2000. [Google Scholar]
  46. Hochreiter, S.; Schmidhuber, J. Long Short-Term Memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef]
  47. Khosla, P.; Teterwak, P.; Wang, C.; Sarna, A.; Tian, Y.; Isola, P.; Maschinot, A.; Liu, C.; Krishnan, D. Supervised Contrastive Learning. Available online: https://arxiv.org/abs/2004.11362v5 (accessed on 15 June 2025).
  48. Anderson, J.R. Acquisition of Cognitive Skill. Psychol. Rev. 1982, 89, 369–406. [Google Scholar] [CrossRef]
  49. Posner, M.I.; Petersen, S.E. The Attention System of the Human Brain. Annu. Rev. Neurosci. 1990, 13, 25–42. [Google Scholar] [CrossRef]
  50. Arpit, D.; Jastrzębski, S.; Ballas, N.; Krueger, D.; Bengio, E.; Kanwal, M.S.; Maharaj, T.; Fischer, A.; Courville, A.; Bengio, Y.; et al. A Closer Look at Memorization in Deep Networks. Available online: https://arxiv.org/abs/1706.05394v2 (accessed on 15 June 2025).
  51. Hasson, U.; Honey, C.J. Future Trends in Neuroimaging: Neural Processes as Expressed within Real-Life Contexts. Neuroimage 2012, 62, 1272–1278. [Google Scholar] [CrossRef]
  52. Lin, T.-Y.; Goyal, P.; Girshick, R.; He, K.; Dollár, P. Focal loss for dense object detection. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 2980–2988. [Google Scholar]
  53. Kumar, M.; Packer, B.; Koller, D. Self-paced learning for latent variable models. Adv. Neural Inf. Process. Syst. 2010, 23, 1189–1197. [Google Scholar]
  54. Gold, J.I.; Shadlen, M.N. The neural basis of decision making. Annu. Rev. Neurosci. 2007, 30, 535–574. [Google Scholar] [CrossRef]
  55. Ratcliff, R.; Smith, P.L. A Comparison of Sequential Sampling Models for Two-Choice Reaction Time. Psychol. Rev. 2004, 111, 333–367. [Google Scholar] [CrossRef]
  56. Thayer, J.F.; Ahs, F.; Fredrikson, M.; Sollers, J.J.; Wager, T.D. A Meta-Analysis of Heart Rate Variability and Neuroimaging Studies: Implications for Heart Rate Variability as a Marker of Stress and Health. Neurosci. Biobehav. Rev. 2012, 36, 747–756. [Google Scholar] [CrossRef] [PubMed]
  57. Kop, W.J.; Synowski, S.J.; Newell, M.E.; Schmidt, L.A.; Waldstein, S.R.; Fox, N.A. Autonomic Nervous System Reactivity to Positive and Negative Mood Induction: The Role of Acute Psychological Responses and Frontal Electrocortical Activity. Biol. Psychol. 2011, 86, 230–238. [Google Scholar] [CrossRef] [PubMed]
  58. Ioannou, S.; Ebisch, S.; Aureli, T.; Bafunno, D.; Ioannides, H.A.; Cardone, D.; Manini, B.; Romani, G.L.; Gallese, V.; Merla, A. The Autonomic Signature of Guilt in Children: A Thermal Infrared Imaging Study. PLoS ONE 2013, 8, e79440. [Google Scholar] [CrossRef] [PubMed]
  59. Brisinda, D.; Picerni, M.; Fenici, P.; Fenici, R. Automatic Calculation of the Parasympathetic, Sympathetic and Baevsky Stress Indexes Provides a More Comprehensive Assessment of Cardiac Autonomic Modulation. Eur. Heart J. 2024, 45, ehae666.3425. [Google Scholar] [CrossRef]
  60. Amhia, H.; Wadhwani, A.K. Stability and Phase Response Analysis of Optimum Reduced-Order IIR Filter Designs for ECG R-Peak Detection. J. Healthc. Eng. 2022, 2022, 9899899. [Google Scholar] [CrossRef]
Figure 1. End-to-end pipeline for rPPG-based emotion recognition. An RGB video is separated into 128-frame segments (≈4 s), converted to rPPG waveforms using a pretrained PhysMamba, encoded using the Multi-scale Temporal Dynamics Encoder (MTDE), and optimized through Phase 0, Phase 1, and Phase 2, which are delimited by blue, orange, and green outlines, respectively. The arrows indicate the direction of the data flow and processing sequence.
Figure 1. End-to-end pipeline for rPPG-based emotion recognition. An RGB video is separated into 128-frame segments (≈4 s), converted to rPPG waveforms using a pretrained PhysMamba, encoded using the Multi-scale Temporal Dynamics Encoder (MTDE), and optimized through Phase 0, Phase 1, and Phase 2, which are delimited by blue, orange, and green outlines, respectively. The arrows indicate the direction of the data flow and processing sequence.
Sensors 25 03995 g001
Figure 2. Phase 0—exploration and representation learning. MTDE embeddings are projected into a 128-dimensional space and trained with Supervised Contrastive Loss, while an internal SoftMax attention head is regularized by entropy; Gated Pooling and the session classifier remain frozen. The arrows indicate the direction of the data flow and processing sequence.
Figure 2. Phase 0—exploration and representation learning. MTDE embeddings are projected into a 128-dimensional space and trained with Supervised Contrastive Loss, while an internal SoftMax attention head is regularized by entropy; Gated Pooling and the session classifier remain frozen. The arrows indicate the direction of the data flow and processing sequence.
Sensors 25 03995 g002
Figure 3. Phase 1—chunk-level discrimination and attentional refinement. (a) Top-K α-Entmax attention selects the most informative chunks; (b) the chunk auxiliary classifier is trained with focal cross-entropy, and Gated Pooling parameters are unfrozen from epoch 25. The arrows indicate the direction of the data flow and processing sequence.
Figure 3. Phase 1—chunk-level discrimination and attentional refinement. (a) Top-K α-Entmax attention selects the most informative chunks; (b) the chunk auxiliary classifier is trained with focal cross-entropy, and Gated Pooling parameters are unfrozen from epoch 25. The arrows indicate the direction of the data flow and processing sequence.
Sensors 25 03995 g003
Figure 4. Phase 2—session-level inference and optimization. α-Entmax weights each chunk, Gated Pooling aggregates gated features into a 256-dimensional session vector, and the main classifier predicts high/low valence–arousal labels using session-level cross-entropy loss. The arrows indicate the direction of the data flow and processing sequence.
Figure 4. Phase 2—session-level inference and optimization. α-Entmax weights each chunk, Gated Pooling aggregates gated features into a 256-dimensional session vector, and the main classifier predicts high/low valence–arousal labels using session-level cross-entropy loss. The arrows indicate the direction of the data flow and processing sequence.
Sensors 25 03995 g004
Table 1. Performance comparison of rPPG extraction methods on MAHNOB-HCI’s ECG.
Table 1. Performance comparison of rPPG extraction methods on MAHNOB-HCI’s ECG.
Method CategoryModel/AlgorithmPre-Trained onMAERMSE
Deep LearningPhysFormerUBFC-rPPG5.377.32
PhysFormerPURE7.019.72
PhysMambaUBFC-rPPG5.387.43
PhysMambaPURE6.168.75
TraditionalICA-13.0216.34
POS-12.8615.96
CHROM-13.3116.80
GREEN-16.9020.44
LGI-15.7619.28
PBV-16.1219.58
OMIT-15.7819.30
Table 2. Head-to-head performance comparison on UBFC-rPPG intra-dataset test.
Table 2. Head-to-head performance comparison on UBFC-rPPG intra-dataset test.
ModelMAERMSEMAPE Pearson   ( ρ ) SNR (db)VRAM (GiB)
PhysFormer0.531.180.760.99940.2510.1/16.3
PhysMamba0.350.790.510.99973.679.8/16.3
Table 3. Main performance results of arousal classification (accuracy, F1-score for positive class, and weighted F1-score) on MAHNOB-HCI test set (53 unseen subjects) with subject-mixed split. Highlights performance solely from temporal domain. All metrics are reported as percentages.
Table 3. Main performance results of arousal classification (accuracy, F1-score for positive class, and weighted F1-score) on MAHNOB-HCI test set (53 unseen subjects) with subject-mixed split. Highlights performance solely from temporal domain. All metrics are reported as percentages.
MethodAccuracy (%)F1 of Positive (%)Weighted F1 (%)
CNN-LSTM [23]61.3150.9659.46
Ours66.0474.2961.97
Table 4. Main performance results of valence classification (accuracy, F1-score for positive class, and weighted F1-score) on MAHNOB-HCI test set (53 unseen subjects) with subject-mixed split. Highlights performance solely from temporal domain. All metrics are reported as percentages.
Table 4. Main performance results of valence classification (accuracy, F1-score for positive class, and weighted F1-score) on MAHNOB-HCI test set (53 unseen subjects) with subject-mixed split. Highlights performance solely from temporal domain. All metrics are reported as percentages.
MethodAccuracy (%)F1 of Positive (%)Weighted F1 (%)
CNN-LSTM [23]73.5076.2373.14
Ours62.2666.6762.26
Table 5. Performance comparison across pooling strategies (arousal classification).
Table 5. Performance comparison across pooling strategies (arousal classification).
MethodAccuracy (%)Weighted F1 (%)
Ours (MTDE + Gated Pooling)66.0461.97
MTDE + Attention Pooling50.9447.56
MTDE + Average Pooling50.9439.07
Table 6. Ablation study on the number of MTDE branches (arousal classification).
Table 6. Ablation study on the number of MTDE branches (arousal classification).
Feature Extractor ArchitectureAccuracy (%)Weighted F1 (%)
Ours (3-branch MTDE)66.0461.97
4-branch MTDE54.7254.72
2-branch MTDE62.2660.42
TCN (Single-branch)58.4952.05
Table 7. Ablation study on curriculum learning strategy (arousal classification).
Table 7. Ablation study on curriculum learning strategy (arousal classification).
Training StrategyAccuracy (%)Weighted F1 (%)
Full Curriculum (Phase 0→2)66.0461.97
Full Curriculum (w/o Top-K)64.1563.89
Phase 0 → 2 (w/o Auxiliary Classifier)58.4958.48
Phase 1 → 2 (w/o SupCon)54.7255.19
Phase 2 (Direct Training)50.9434.18
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Lee, C.; Lee, H.; Whang, M. Emotion Recognition from rPPG via Physiologically Inspired Temporal Encoding and Attention-Based Curriculum Learning. Sensors 2025, 25, 3995. https://doi.org/10.3390/s25133995

AMA Style

Lee C, Lee H, Whang M. Emotion Recognition from rPPG via Physiologically Inspired Temporal Encoding and Attention-Based Curriculum Learning. Sensors. 2025; 25(13):3995. https://doi.org/10.3390/s25133995

Chicago/Turabian Style

Lee, Changmin, Hyunwoo Lee, and Mincheol Whang. 2025. "Emotion Recognition from rPPG via Physiologically Inspired Temporal Encoding and Attention-Based Curriculum Learning" Sensors 25, no. 13: 3995. https://doi.org/10.3390/s25133995

APA Style

Lee, C., Lee, H., & Whang, M. (2025). Emotion Recognition from rPPG via Physiologically Inspired Temporal Encoding and Attention-Based Curriculum Learning. Sensors, 25(13), 3995. https://doi.org/10.3390/s25133995

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop