Next Article in Journal
MSDP-SAM2-UNet: A Novel Multi-Scale and Dual-Path Model for Wheat Leaf Disease Segmentation Based on SAM2-UNet
Previous Article in Journal
Ship Traffic Flow Analysis and Prediction in High-Traffic Areas Under Complex Environments
Previous Article in Special Issue
Understanding the Impact of User Engagement Metrics on the Dissemination of Traditional Culture: A Structural Equation Modeling Analysis of Cantonese Opera Videos on Bilibili
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Review

Machine Learning, Physiological Signals, and Emotional Stress/Anxiety: Pitfalls and Challenges

1
School of Computer Science and Technology, Hangzhou Dianzi University, Hangzhou 310018, China
2
The China-Spain AI Technology Joint Laboratory
3
Clinical and Experimental Neuroscience (NiCE), Biomedical Research Institute of Murcia (IMIBPascual Parrilla), Institute for Aging Research, UniWell, Campus Mare Nostrum, School of Medicine, University of Murcia, 30100 Murcia, Spain
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Appl. Sci. 2025, 15(21), 11777; https://doi.org/10.3390/app152111777
Submission received: 30 July 2025 / Revised: 13 September 2025 / Accepted: 20 September 2025 / Published: 5 November 2025
(This article belongs to the Special Issue Advances and Applications of Complex Data Analysis and Computing)

Abstract

Anxiety and emotional stress are pervasive psychological challenges that profoundly impact human health in today’s fast-paced society. Traditional assessment methods, such as self-reports and clinical interviews, often suffer from subjective biases and lack the capability for objective, real-time evaluation of mental states. However, the integration of physiological signals—including electroencephalography (EEG), heart rate (HR), electrodermal activity (EDA), and eye movements—with advanced machine learning (ML) techniques, offers a promising approach to automate and objectify mental health assessments. A systematic review was conducted to explore recent advances in the early detection of anxiety and stress by combining physiological signals and ML methods. To assess methodological quality, a specific analysis framework was designed for the 113 studies included, which identified significant deficiencies in the literature. This highlights the urgent need to adopt standardized reporting guidelines in this field. The role of these technologies in feature extraction, classification, and predictive modeling was analyzed, also addressing critical challenges related to data quality, model interpretability, and the influence of intersectional factors like gender and age. Ethical and privacy considerations in the current research were also included. Finally, potential avenues for future research were summarized, highlighting the potential of ML technologies for early detection and proactive intervention in mental disorders.

1. Introduction

Mental disorder issues refer to clinically significant disturbances in an individual’s cognition, emotional regulation, or behavior, often leading to distress or impairment in critical areas of functioning. These issues have become a major global concern, with profound impacts on society and public health systems. By 2023, the prevalence of probable mental disorders among children aged 8 to 16 in England rose to 20.3%, while it reached 23.3% for those aged 17 to 19, with anxiety and stress disorders being the most prevalent [1]. The COVID-19 pandemic, which began in 2020, has further exacerbated the mental disorder problems. The prevalence of anxiety and emotional stress has increased significantly by 26% [2]. The number of female patients is twice that of men [3], posing a major challenge to both individuals and the health care system.
Anxiety and emotional stress are two key factors affecting mental health problems and are the body’s natural response to a perceived threat or need. Appropriate levels of brief, situation-specific anxiety and emotional stress are beneficial and can provide extra strength to protect oneself or enable a quick response to avoid danger [4,5]. However, chronic anxiety or emotional stress poses significant challenges [6] that can persist over extended periods and can impair cognitive function and emotional stability [7,8], and have serious consequences for overall well-being [9], increasing the risk of various health issues, including cardiovascular disease [10], diabetes [11], and immune system dysfunction [12]. Furthermore, in addition to their direct physiological consequences, chronic anxiety and stress frequently serve as primary etiological factors for the development of specific mental disorders. These conditions impose a significant burden on both the quality of life of affected individuals and the socioeconomic resources of global healthcare systems [13,14]. Moreover, they may represent pivotal precursors to the onset of substance abuse and illicit drug use [15].
Therefore, effective monitoring and predicting of the levels of anxiety and emotional stress can prevent these harmful effects and improve the quality of life. Physiological signals, including electroencephalography (EEG), heart rate (HR), electrodermal activity (EDA), and eye movements, offer a streamlined and effective approach, providing key insights into the body’s response to stressors [16]. Unlike traditional techniques, such as self-report questionnaires and cortisol tests, which rely primarily on subjective perception, biomarkers such as EEG, HR, EDA, and eye movement, are the direct and reliable reflection of the body’s stress state and may provide an objective, non-invasive method for assessing anxiety and emotional stress levels, as well as preventing the bias of traditional methods.
Through advancements in Machine Learning (ML) applied to physiological data, the classification and prediction of anxiety and emotional stress levels have evolved considerably [17]. ML techniques excel at analyzing complex physiological data and identifying patterns that might be missed by traditional methods [18]. This capability allows for improved early detection and management of mental disorders, providing a nuanced understanding of an individual’s anxiety and emotional stress levels and overall mental health. However, despite its potential, the application of ML to assess anxiety and emotional stress based on physiological signals faces several challenges. Firstly, ensuring the quality of physiological signals is difficult due to their low signal-to-noise ratio (SNR), leading to inaccurate or incomplete data and undermining the reliability of ML models. Secondly, interpreting ML models can be complex, as they often function as “black boxes,” making it difficult to understand how they arrive at their conclusions. Thirdly, ML models often struggle to account for the impact of other important factors such as gender and age [19,20], in addition to physiological signals, on anxiety and emotional stress. Finally, intersectional factors such as race, socioeconomic status, sexual orientation, ethics, and privacy are often not incorporated into physiological signals, directly affecting the credibility of anxiety and emotional stress assessment results using ML. The primary objective of this review is to explore how ML methods are applied to classify and predict anxiety or stress-related mental disorders using physiological signals. The PRISMA 2020 guidelines [21] will be followed to assess the effectiveness of these methods, highlight their limitations, and address key challenges such as data integrity, model interpretability, demographic influences, and ethical considerations. A secondary objective is to advance early detection and intervention strategies for mental health disorders.
The contributions of this review are summarized in four aspects: (1) We quantify the recent landscape across the included studies by modality and compare machine learning and deep learning performance. (2) We identify recurrent failure modes, especially the challenge of distinguishing stress from neutral states. (3) We connect technical choices to data quality along the end-to-end pipeline from acquisition and preprocessing to feature learning and evaluation. (4) We integrate demographic, intersectional, and privacy considerations into actionable guidance for reproducible and equitable study design.

2. Materials and Methods

To improve the clarity and precision of the results, a rigorous methodological approach adhering to PRISMA guidelines was employed. This involved the implementation of a well-defined search strategy, systematic data collection, clearly delineated inclusion and exclusion criteria, and a comprehensive review and synthesis of the findings.

2.1. Search Strategy

A focused and comprehensive search strategy was employed to investigate the application of ML in the analysis of physiological signals associated with anxiety and emotional stress. This search was conducted between 1 January 2024, and 19 July 2024, to identify the relevant literature published in English within the last 5 years and available in the PubMed and Web of Science databases. The selected search terms ensured a thorough examination of the integration of physiological monitoring and ML techniques. The search terms used are listed below (for more details, please refer to Appendix A).
  • (“anxiety” [Title/Abstract]
  • OR     “blood pressure” [Title/Abstract]
  • OR     “EL” [Title/Abstract]
  • OR     “eye tracking” [Title/Abstract]
  • OR     “heart rate” [Title/Abstract]
  • OR     “skin conductance” [Title/Abstract]
  • OR     “stress” [Title/Abstract])
  • AND  (“classification” [Title/Abstract]
  • OR     “deep learning” [Title/Abstract]
  • OR     “machine learning” [Title/Abstract])
This search strategy was designed to comprehensively cover a wide range of research studies investigating the application of ML to diverse physiological signals associated with anxiety and emotional stress. By incorporating both medical and computational terminology, the search strategy facilitated the identification of studies at the intersection of computational neuroscience, encompassing the variability of physiological responses, the influence of demographic factors, and the imperative need for transparent and interpretable ML models. This strategic approach to literature selection was crucial for assembling a comprehensive body of research that aligned with the objectives of this review and provided a solid foundation for understanding the current state and future potential of ML in the study of anxiety and emotional stress.

2.2. Inclusion and Exclusion Criteria

The keyword search yielded 1871 documents. To refine the inclusion criteria in the data collection phase, a two-stage selection process was implemented. In the first stage, titles and abstracts were screened to identify studies relevant to the assessment of anxiety and/or stress using physiological signals such as EEG, electrocardiogram (ECG), EDA, HRV, or eye tracking, involving ML methods for data analysis, and reporting evaluation metrics with results demonstrating model performance. In the second stage, full texts of the selected articles were examined to confirm their true relevance to the review goals. To ensure the accuracy and consistency of the selected articles, a dual independent extraction and cross-validation method was adopted, and the screening results were compared and discussed, with no inconsistencies found. This rigorous selection process ensured the inclusion of studies that were most relevant to the study objectives. The review included only those articles that assessed anxiety or stress using physiological signals analyzed with ML methods, reported specific evaluation metrics, and provided clear findings. Conversely, studies that were unrelated to anxiety or stress assessment, were written in languages other than English, were unpublished, involved only subjects without anxiety or stress, or were specifically focused on conditions such as Post-Traumatic Stress Disorder (PTSD) were excluded.

2.3. Screening Synthesis and Results

Following the establishment of the search strategy and the definition of inclusion and exclusion criteria, a comprehensive literature search yielded a total of 1781 records from two primary databases, PubMed (n = 1262) and Web of Science (n = 519). As depicted in Figure 1, the PRISMA flow diagram outlines the study selection process, detailing the total number of records identified, the screening procedure after duplicate records were removed, and the final studies included for analysis. In the identification phase, an initial screening identified and removed 135 duplicate records. Automation tools were further employed to eliminate an additional 2 records. The remaining 1644 records underwent a title and abstract screening to assess their relevance to the study objectives. Consequently, 1169 records were excluded based on the predefined criteria. Of the 475 potentially eligible studies, 362 could not be retrieved due to reasons such as irrelevance to (ML/DL), lack of reported metrics, invasive nature of the procedures, or absence of the specified physiological signals. The remaining studies were assessed for eligibility, resulting in the final inclusion of 113 studies for analysis.
These studies underwent detailed data extraction, and information was collected on research objectives, data types, and data collection equipment, sample characteristics, preprocessing methods, ML or deep learning (DL) models employed, feature extraction techniques, evaluation metrics, and main results. This rigorous approach ensured a comprehensive analysis of relevant and high-quality literature.

2.4. Study Quality Assessment

To address the absence of standardized quality assessment instruments specifically designed for machine learning studies utilizing physiological signals, we developed and implemented a domain-specific evaluation framework to systematically assess the methodological rigor of all 113 included studies. This retrospective assessment was conducted based exclusively on information explicitly reported in the published manuscripts.

2.4.1. Assessment Framework and Evaluation Criteria

Our quality assessment framework encompassed five critical domains fundamental to the validity and reproducibility of machine learning research in physiological signal analysis. These domains included study population and sample characteristics, signal acquisition methodology, data preprocessing transparency, model validation strategies, and confounding factor consideration.
The first domain evaluated the adequacy of sample size reporting and the completeness of participant demographic information, including age distribution and gender composition, which are essential for assessing the generalizability of findings across different populations.
The second domain focused on the transparency of reporting regarding data collection procedures, examining whether studies provided sufficient detail about hardware specifications, sampling rates, electrode placement protocols, and experimental conditions to enable replication.
Data preprocessing and signal processing transparency constituted the third evaluation criterion. This assessment examined the comprehensiveness of preprocessing methodology reporting, particularly focusing on artifact removal procedures, noise reduction techniques, and signal filtering approaches that directly impact data quality and subsequent model performance.
The fourth domain addressed model validation and evaluation strategies. We evaluated the robustness of validation approaches employed, distinguishing between studies that relied solely on training set performance and those implementing more rigorous cross-validation procedures or independent test set evaluation methodologies.
Finally, we assessed each study’s consideration of potential confounding factors. This evaluation examined whether studies acknowledged and addressed confounding variables such as medication effects, baseline anxiety levels, and other demographic or clinical factors that might influence physiological responses and thereby affect model validity.

2.4.2. Quality Assessment Results

Our systematic evaluation revealed substantial heterogeneity in methodological reporting quality across the included literature. Through comprehensive analysis of the 113 studies, we identified several recurring patterns that highlight areas requiring improvement in current research practices.
The quality and completeness of methodological reporting varied considerably across studies, with many investigations providing insufficient detail to enable comprehensive quality assessment. Among studies that provided adequate methodological information for detailed analysis, we observed substantial variation in research design and implementation approaches. Sample sizes across the literature demonstrated considerable range, with some investigations involving as few as 10 participants while others included over 100 individuals, reflecting the diversity of research contexts and objectives within this field.
Examination of validation methodologies revealed inconsistent implementation of rigorous evaluation strategies across the literature. While most studies reported some form of model validation, there was considerable variation in the sophistication of these approaches. Many investigations relied primarily on cross-validation techniques, with limited implementation of independent test set evaluation or external validation procedures. This pattern raises concerns about the generalizability of reported model performance beyond the specific study populations and experimental conditions.
The reporting of data preprocessing methodologies also demonstrated significant inconsistencies. Although the majority of studies acknowledged implementing preprocessing procedures, the level of methodological detail varied markedly across publications. Some studies provided comprehensive descriptions of artifact removal, filtering, and signal conditioning procedures, while others offered only cursory mentions of these critical steps. This variability substantially limits the reproducibility of reported findings and complicates meaningful comparison of results across different investigations.
Furthermore, our analysis revealed limited systematic consideration of potential confounding variables that could influence physiological responses and, consequently, model performance. Factors such as participant medication use, pre-existing medical conditions, and baseline anxiety or stress levels were infrequently reported or systematically controlled for in the analytical approaches. This oversight represents a significant methodological limitation that may compromise the validity of reported associations between physiological signals and anxiety or stress states.
The heterogeneity in reporting quality prevented systematic quantification of methodological adequacy across all included studies. However, the patterns identified through our assessment highlight fundamental challenges in evidence synthesis within this rapidly evolving field, particularly the need for standardized reporting frameworks that would facilitate more rigorous systematic evaluation of research quality.

2.4.3. Assessment Limitations and Methodological Considerations

Several important limitations constrain the interpretation of our quality assessment findings. First, this evaluation was necessarily limited to information explicitly reported in the published manuscripts, and the absence of reporting does not definitively indicate that appropriate methodological procedures were not implemented. Second, the heterogeneity in journal requirements and article length restrictions may have contributed to variations in methodological detail across studies. Finally, the retrospective nature of this assessment precluded the application of more comprehensive quality evaluation instruments that would require access to original study protocols or datasets.
These findings collectively highlight the critical need for the development and adoption of standardized reporting guidelines specific to machine learning applications in physiological signal research, similar to established frameworks in other research domains.

3. Results

3.1. Physiological Signals

This comprehensive literature review emphasizes that physiological signals, such as EEG [22], HR [23], HRV [24], EDA [25], and eye movements [26], provide valuable data for machine learning-based early detection of mental health disorders. EEG offers high temporal resolution but is susceptible to noise, while MRI, despite high detail, is costly and less accessible [27]. HR and HRV, easily collected via wearable devices, are linked to stress but provide limited information compared to EEG. Respiratory rate is easily measured but provides less information than EEG or HRV [26]. EDA is sensitive to emotional responses but is susceptible to environmental factors [25]. Eye movement data provide insights into attention and emotion but are prone to head movement interference [26].
Among the 113 reviewed papers, approximately 51 utilized EEG, 21 used HR or HRV, 17 employed EDA, 26 adopted ECG, 14 implemented PPG, and 4 incorporated EMG.

3.1.1. EEG

EEG, a non-invasive technique, records brain electrical activity through scalp electrodes, reflecting neuronal synchronization and cognitive-emotional states [27,28]. The location of electrodes on the EEG cap always (e.g., 64, 128) follows the 10–20 system (Figure 2) [29,30]. Raw EEG data usually contain the noise derived from eye blinking and muscle activity, which introduces artifacts. To mitigate the adverse effects of these artifacts, preprocessing is essential. A typical preprocessing pipeline involves band-pass filtering and the application of independent component analysis (ICA) to identify and remove artifact-related components. In EEG analysis, feature extraction serves as a critical step bridging the signals and ML models. Handcrafted features are commonly employed to represent the time and frequency domain characteristics of EEG signals [18,19,31]. Time domain features include mean, variance, peak-to-peak amplitude, and signal-to-noise ratio of multichannel EEG, which capture the overall fluctuations of the signal. Frequency domain features always focus on the frequency bands Delta (0.5–4 Hz), Theta (4–8 Hz), Alpha (8–13 Hz), Beta (13–30 Hz), and Gamma (30–100 Hz) [27,28], which are often associated with specific emotional states. These handcrafted features offer high interpretability and are well-suited for classifiers such as Support Vector Machines (SVMs), Linear Discriminant Analysis (LDA), and Random Forests.

3.1.2. Heart Rate (HR) and Heart Rate Variability (HRV)

Heart rate (HR) and heart rate variability (HRV) are key indicators of cardiovascular function and autonomic nervous system activity, and can be recorded through photoplethysmography (PPG) or electrocardiogram (ECG) devices [32,33,34,35,36]. HRV measures the fluctuation in RR intervals (Figure 3A), reflecting the balance between sympathetic and parasympathetic activity, with decreased HRV associated with stress and anxiety [37,38,39,40,41,42]. In HR and HRV data, ectopic beats, motion artifacts, and missing intervals can distort variability measures. Preprocessing procedures include the detection and removal of ectopic beats, interpolation of missing intervals, and resampling for evenly spaced time series when spectral analysis is applied. Extracted features involve time-domain (e.g., SDNN, RMSSD) and frequency-domain (Figure 3B,C) [18,19,31]. High HRV indicates good cardiovascular health, while low HRV is associated with increased stress and health risks [34]. Table 1 summarizes key HRV metrics.

3.1.3. Electrodermal Activity (EDA)

Electrodermal Activity (EDA), reflecting autonomic nervous system activity, measures changes in skin conductance due to sweat gland activity [43,44,45,46]. EDA is typically measured using non-invasive sensors (e.g., fingertip sensors, wristbands) (Figure 4) [30]. Similarity, some artifacts arising from motion, temperature fluctuations, and sensor noise would also decrease the quality of the EDA signals. Common approaches include down-sampling, smoothing, baseline correction, and artifact removal using methods such as median filtering or adaptive thresholding. Signal decomposition is often applied to separate tonic (skin conductance level, SCL) and phasic (skin conductance response, SCR) components, which correspond to slow baseline changes and rapid event-related responses, respectively [47]. Feature extraction from EDA can be performed in multiple domains. Time-domain features include mean and variance of SCL, as well as amplitude, latency, and frequency of SCR peaks. Frequency-domain features, obtained through Fourier or wavelet analysis, characterize oscillatory patterns associated with autonomic activity. After that, the obtained EDA signals can be used for effective stress and anxiety monitoring [48,49,50].

3.1.4. Eye Movement

Eye movement refers to the motion or activity of the eyes. By analyzing fixations, saccades, and variations in pupil diameter derived from eye movements, visual attention and cognitive processes can be distinguished [50]. Eye-tracking devices can collect non-invasive, high-resolution gaze position, pupil diameter, or some blink events [51,52,53,54,55,56,57]. Preprocessing of eye movement data is necessary to reduce noise from head movements or blinks. The pipeline includes filtering, blink detection and interpolation, coordinate normalization, and calibration correction, which could ensure consistency across trials or sessions. For gaze position, smoothing or velocity thresholding are often applied to identify saccades and fixations accurately [58]. Features extracted from eye movement also include spatial and temporal parts. Spatial features such as fixation duration, dispersion, and saccade velocity reflect visual attention and search strategies, while temporal features such as blink rate, pupil dilation dynamics, and inter-saccadic intervals can relate to cognitive load or emotion arousal [59,60]. Through the preprocessing and feature extraction, it is possible to use eye movement data to predict the anxiety and stress state of subjects.

3.1.5. Other Physiological Signals

Beyond EEG and EDA, respiratory rate [24], blood pressure [59], and blood oxygen levels [60] are crucial physiological indicators of stress and anxiety. Rapid and shallow breathing patterns often accompany heightened anxiety [30]. Blood pressure fluctuations reflect stress responses, with elevated levels indicating increased anxiety and potential health risks [61]. Decreased blood oxygen saturation may signal respiratory distress or heightened anxiety. These physiological signals, in conjunction with EEG and EDA, provide a comprehensive understanding of the body’s stress response.

3.2. Machine Learning Methods in Stress and Anxiety

At the heart of the modern AI revolution lie two main approaches to machine learning: traditional ML and DL models (Figure 5). Due to the outstanding performance of conventional ML and DL methods in feature extraction and classification evaluation, many researchers have applied them to assessing and predicting stress and anxiety. A clear understanding of these methodologies is essential for enhancing predictive modeling and improving mental health interventions. Although both aim to enable computers to think, they differ significantly in their execution and applications, ML techniques excel in handling complex and high-dimensional data, and are particularly suited for feature extraction and classification tasks in physiological signal analysis [62]. This section details commonly used ML and DL methods for anxiety and stress evaluation and explains the differences between them, as well as their relationship.

3.2.1. ML Approaches for Emotional Stress and Anxiety Analysis Using Physiological Signals

Machine Learning involves separate steps of each process for feature extraction and classification, requiring manual intervention. Traditional ML methods are extensively applied in the analysis of physiological signals [63]. These methods include, but are not limited to, Support Vector Machines (SVMs), Naive Bayes, and Multilayer Perceptrons (MLPs). SVM is a supervised learning technique that classifies data by finding the optimal separating hyperplane. SVM performs well with high-dimensional data and is suitable for classifying physiological signals such as (HRV) and Skin Conductance Responses (SCR) [64]. Naive Bayes is based on Bayes’ theorem and assumes independence among features. Although this assumption may not be held in practical applications, Naive Bayes still demonstrates good performance in physiological signal classification, especially with smaller datasets [65]. MLP, a feedforward neural network with input, hidden, and output layers, uses the backpropagation algorithm to capture non-linear relationships in physiological signals [66]. Beyond SVM, Naive Bayes, and MLP, we also consider classic tabular-learning baselines and tree ensembles that are well suited to handcrafted physiological features (e.g., HRV and EDA statistics). K-Nearest Neighbor (KNN) serves as a nonparametric reference that classifies samples by their local neighborhood; it is simple and effective when features are well scaled. Logistic Regression provides a linear, interpretable baseline with calibrated decision scores, often competitive when signal-derived features are informative. Decision Trees capture non-linear interactions in a transparent rule-based form, but may be overfit without pruning. Random Forests mitigate such variance by ensembling many trees, yielding robust performance across heterogeneous feature sets. Gradient boosting implementations (LightGBM, XGBoost, CatBoost) further improve accuracy on tabular data by sequentially correcting residuals; they handle non-linearities and feature interactions with built-in regularization, albeit at the cost of careful hyperparameter tuning and higher computational demand. Together with SVM/MLP, this spectrum of models spans linear, instance-based, and ensemble methods, enabling fair comparisons across decision-boundary complexity, interpretability, and robustness for stress/anxiety detection from physiological signals. For predicting anxiety and stress from physiological signals, ML involves several key steps. Data preprocessing ensures data quality by removing noise and normalizing values. Data augmentation adds diversity to the dataset, helping prevent overfitting. Feature selection identifies relevant attributes using techniques like PCA and MI. Traditional algorithms, such as SVM and Naive Bayes, are trained to classify and predict [67]. Hyperparameter tuning refines model performance through methods like grid search. Finally, model evaluation uses metrics like accuracy and confusion matrices to assess prediction reliability.

3.2.2. Fusion Strategies of Multimodal Physiological Signals

In some studies, multimodal physiological signals are used. The fusion strategies can be broadly categorized into early fusion, intermediate fusion, and late fusion.
Early fusion, also known as feature-level fusion, which concatenates features from different modalities after feature extraction and selection, is the most widely used approach in machine learning [57,68,69,70,71,72,73]. Its main advantage lies in its simplicity and the ability to directly exploit complementary information across modalities. This makes it particularly suitable for traditional shallow machine learning models with limited representation capacity. However, early fusion is sensitive to differences while processing features with different scales or dimensions, and it is possible to introduce noise when modalities are heterogeneous. Moreover, the model cannot dynamically learn complex cross-modal relationships, which means the performance may degrade if some modalities are missing or corrupted. Intermediate and late fusion are less commonly used in traditional machine learning. Intermediate fusion, which integrates modality-specific representations at hidden layers, and late fusion, which combines decisions from individual modalities, offer greater flexibility and robustness but are more complex to implement and typically require deeper models or larger datasets. In this way, these two strategies are limited in shallow, resource-constrained machine learning settings. Existing fusion strategies have their own advantages and limitations, and their performance may vary across different datasets. Therefore, the choice of a fusion strategy should be guided by the specific characteristics and requirements of the task.

3.2.3. ML Performance in Analyzing Physiological Signals

Machine learning models have shown promising performance in predicting stress and anxiety using physiological signals, and they are widely employed by researchers. Table 2 summarizes recent studies that applied traditional machine learning methods for stress detection and reports their final classification accuracy. For each study, the models demonstrating the highest levels of accuracy were chosen. It is important to note that these accuracy values are for reference only, as different studies used different signals and datasets, making direct comparison difficult. However, these values still offer valuable reference points. As shown in Table 2, traditional machine learning methods have demonstrated excellent performance in stress and anxiety prediction based on neural signals. Despite the variations in neural signals, models, and datasets used across studies, most models achieved an accuracy of over 75%. Notably, the KNN model reached an accuracy of over 94% when classifying stress based on ECG and EDA signals [67,74], highlighting the great potential of machine learning methods in this area of research. Ensemble learning methods, which combine multiple models for classification prediction, also exhibited superior performance in predicting stress and anxiety from physiological signals [64,75]. Despite demonstrating a clear ability to distinguish between ‘stress’ and ‘relaxation’ states using HRV signals, machine learning models have shown limitations in differentiating between ‘stress’ and ‘neutral’ states [42]. This underscores the challenges machine learning models face in this field, suggesting that while they show significant potential, further refinement is needed to address these limitations.
A comprehensive analysis of Table 2 indicates that model efficacy is highly coupled with the types of physiological signals used and their specific research contexts, rather than being solely represented by isolated accuracy figures. EEG signals, which reflect central nervous system activity [66,75], demonstrate outstanding performance in well-controlled experimental settings due to their high temporal resolution and capacity to characterize neural mechanisms. For instance, Ref. [75] achieved over 97% accuracy using a large clinical sample, highlighting EEG’s strong potential for fine-grained emotion discrimination. However, such approaches generally rely on specialized equipment and constrained environments, which limit their ecological validity and practical scalability. In contrast, signals derived from the peripheral nervous system—such as ECG, EDA, and PPG—offer stronger real-world applicability due to their compatibility with wearable devices, as exemplified using Empatica E4 in Ref. [45]. Studies such as Refs. [67,68] have shown that even with small sample sizes (N = 18 and N = 10, respectively), fusion of ECG and EDA can achieve around 94% accuracy. It should be noted, however, that such high performance may be limited to distinguishing between extreme states. A fundamental challenge common across all signal types was highlighted by Ref. [42]: although HRV-based models (derived from ECG/PPG) effectively differentiated between “stress” and “relaxation” states (Random Forest, 89.2%), their performance dropped significantly to approximately 60% when identifying more nuanced distinctions such as between “stress” and “neutral” states. This suggests that current models may be overfitting to prominent physiological responses in specific experimental scenarios, rather than capturing the more complex and subtle patterns of stress encountered in real-world environments. In terms of model selection, ensemble learning methods—such as LightGBM and XGBoost [69] and Random Forest in Ref. [42]—exhibited superior and consistent performance across different signals and datasets, making them a preferred robust option. Moreover, the comparable results between Logistic Regression (76.4%) and Random Forest (76.5%) in Ref. [45] indicate that, in some applications, feature engineering and signal quality may be as critical as model selection. Future research must move beyond singular focus on accuracy metrics, adopt more rigorous cross-subject validation protocols, and focus on enhancing decoding capacity for subtle and continuous mental states in naturalistic settings.

3.3. DL Approaches for Emotional Stress and Anxiety Analysis Using Physiological Signals

Deep learning, a branch of machine learning, simplifies the process of feature extraction and classification through complex neural networks, thereby improving both efficiency and accuracy (Figure 6) [62]. In recent years, deep learning methods have achieved significant success in analyzing physiological signals such as EEG and ECG. Common architectures in deep learning include Convolutional Neural Networks (CNNs), Recurrent Neural Networks (RNNs), and Long Short-Term Memory (LSTM) networks [77]. CNNs are widely used to extract spatial features through convolutional and pooling layers, reducing the need for manual feature engineering (Figure 6A). Moreover, CNNs are not only adept at extracting spatial features, but can also capture temporal features to some extent through their layered structure and local receptive fields, making them suitable for analyzing time-series data. However, CNNs have limitations in capturing more complex long-term dependencies. To address this, RNNs (Figure 6B) use recurrent connections to retain temporal information, making them well-suited for time-series data analysis [78].
However, RNNs face challenges such as the vanishing gradient problem when processing long sequences. LSTM networks (Figure 6C), an improved version of the RNN architecture, were specifically designed to overcome the vanishing gradient issue in standard RNNs. Unlike standard RNNs, LSTMs introduce memory cells and gating mechanisms that enable them to effectively capture and maintain long-term dependencies, making them more suitable for handling long time-series data [79]. In the application of deep learning for stress and anxiety prediction, neural signal data are first preprocessed to ensure high-quality input. In some cases, data augmentation techniques are employed to expand the dataset, enhancing the model’s generalization ability and reducing the risk of overfitting. CNNs are used to capture spatial features and, to some extent, temporal features, while RNNs and LSTMs focus on capturing complex temporal patterns. Together, these three architectures can automatically learn meaningful features from the data for prediction. These foundational architectures form the basis of the most existing deep learning models.
Deep learning methods have shown promising results in predicting stress and anxiety using physiological signals, and they are increasingly adopted by researchers. Table 3 summarizes recent studies applying deep learning techniques for stress detection, listing the final classification accuracy. In the “Method” column, the model names are either the ones given by the original authors or, if not provided, the primary architecture of the model is used as the name. The “Structure” column indicates the key architectures used in the models.
An analysis of accuracy has been performed in Table 3, where based on neural signals in order to detect anxiety and stress, the different deep learning methods used demonstrate excellent performance, and despite variations in datasets, models and signal most models achieve accuracy rates exceeding 80%. Notably, the DNN model reported an accuracy of 95.51% for EEG-based stress detection [80], underscoring the ability of DNNs to capture complex patterns in large datasets, which is crucial for precise stress and anxiety evaluation. Convolutional Neural Networks (CNNs) have also played a significant role, particularly in spatial data analysis, such as EEG signal processing. The LConvNet model, a hybrid CNN-LSTM, demonstrated outstanding performance with an accuracy of 97%, which highlights the model’s effectiveness in handling both spatial and temporal features [81]. Similarly, the MuLHiTA and BLSTM models achieved accuracy rates of 91.8% and 84.13%, respectively, further validating the strength of LSTM-based models in analyzing sequential data [61]. In addition to EEG signals, the CNN and 1D-CNN models for ECG-based stress detection showed promising results, with accuracies of 93.74% and 83.55%, respectively [64,82]. The CNN-TLSTM hybrid model, which combines EEG and pulse rate data, achieved an impressive accuracy of 97.86%, suggesting that incorporating multiple physiological signals enhances model performance [69]. Although the deep learning models summarized in Table 3 generally report high accuracy rates (mostly exceeding 80%) in anxiety and stress detection, a tiered evaluation based on dataset scale, quality, and model generalizability reveals significant differences in their practical value. Studies utilizing large-scale, diverse samples with detailed acquisition information (e.g., Ref. [80] with 81 subjects and over 18,000 samples) demonstrate higher reliability and potential generalizability and should be regarded as high-quality evidence. In contrast, studies relying on small sample sizes (e.g., Ref. [69] with only 19 subjects) or insufficiently described datasets, while reporting high performance metrics, may suffer from overfitting and limited generalizability, and their results should be interpreted cautiously. In terms of model selection, the suitability of an architecture highly depends on data scale and task requirements. For low-data environments or scenarios requiring implementable deployment, lightweight architectures such as EEGNet (84.66% in [61]) and 1D-CNN (83.55% in [64]) are more practical and reliable choices due to their parameter efficiency and training stability. Conversely, although large hybrid models (e.g., CNN-LSTM and CNN-TLSTM) can achieve peak performance with sufficient data (97.86% in [69]), their vast parameter counts and complex structures make them highly prone to overfitting in limited-data settings, constituting “over-engineering” that should be avoided without justification. Furthermore, although LSTM-based architectures outperform CNNs in modeling long-range dependencies, they still face limitations such as low training efficiency and difficulties in capturing complex relationships in ultra-long sequences. Moreover, the field as a whole has not yet fully incorporated more advanced global modeling mechanisms, such as Transformers. Additionally, a fundamental methodological challenge in current research paradigms lies in the conflation of experimental design and model learning objectives. Many studies (including some in Table 2) use “task-induced states vs. resting states” as a proxy for “stress/anxiety vs. relaxation.” This approach carries an inherent risk: models may learn not the physiological fingerprints specific to stress or anxiety, but rather general physiological differences between task performance and resting states (e.g., changes due to cognitive load, muscular activity, or sensory input). Therefore, the high accuracy reported by many models may reflect the model’s ability to distinguish behavioral states rather than specific affective states. Future studies should adopt more nuanced experimental designs—such as employing multiple distinct stress-induction tasks, including neutral task controls, and utilizing ecological momentary assessment—coupled with explainability techniques to ensure that models capture core, generalizable features of stress across contexts rather than artifacts of specific experimental paradigms.
Table 3. Recent studies on stress detection using DL.
Table 3. Recent studies on stress detection using DL.
ResearchSignalMethodAccuracy%StructureDataset Details
[69] (2023)EEG, pulse rateCNN-TLSTM97.86CNN, LSTM19 subjects (15 M/4 F, 31 ± 14 years), 8 electrodes, 250 Hz sampling rate. Each participant yields 20 × 90 s EEG measurements, generating a 30 min EEG record per user.
[80] (2022)EEGDNN95.51-45 GAD patients (41.8 ± 9.4 years, 13 M/32 F); 36 HC (36.9 ± 11.3 years, 11 M/25 F). 16-channel EEG, 250 Hz. 10,273 GAD and 7773 HC samples.
[70] (2024)EEGGRU94.4LSTM24 healthy subjects; EEG acquired via ECL electrodes, filtered with a 4–400 Hz band-pass filter at a 256 Hz sampling rate.
[82] (2020)ECGCNN93.74CNN-
[61] (2022)EEGMuLHiTA91.8LSTM
[61] (2022)EEGCNN-LSTM87.57CNN, LSTM
[83] (2019)EEGCNN87.5CNN-
[61] (2022)EEGEEGNet84.66CNNMIST: 20 males, 22.54 ± 1.53 years, 500 Hz, 24 electrodes. STEW: 45 males, 18–24 years, 128 Hz, 14 electrodes. DMAT: 25 subjects (20 F/5 M), 18.60 ± 0.87 years, 500 Hz, 23 electrodes.
[61] (2022)EEGBLSTM84.13LSTM
[64] (2021)ECG1D-CNN83.55CNN
[64] (2021)ECGVGG83.09CNNDriveDB: multiparameter recordings during driving; ECG, QRS, GSR. Arachnophobia: 80 spider-fearful individuals (18–40 years); 100 Hz, 10-bit resolution; ECG, GSR, respiration.
[81] (2023)EEGLConvNet97CNN, LSTMTUEP v2.0.0: 49 epilepsy, 49 healthy sessions; ≥25 channels.
[81] (2023)EEGDeepConvNet96CNN
[83] (2019)HRV, EDA, EEGCNN90CNN-
[83] (2019)HRV, EDACNN84CNN-
[81] (2023)EEGShallowConvNet78CNN

3.4. Software and Hardware Platforms

We systematically extracted and summarized information on the software and hardware platforms used to implement the machine learning models wherever such information was available. The most commonly reported software environments were MATLAB (version 2024b) and Python 3.11.3 (with libraries such as scikit-learn, TensorFlow, Keras, and PyTorch). Hardware descriptions were infrequently provided; when specified, implementations typically ran on standard desktop workstations or laptops, and a minority explicitly reported GPU acceleration (e.g., NVIDIA GPUs) for deep learning workloads.

4. Discussion

In this review, we synthesize modality-spanning evidence via a PRISMA workflow and map pitfalls onto the end-to-end ML pipeline, yielding concrete guidance for model selection and deployment. We explicitly link demographic, intersectional, and privacy issues to technical decisions (preprocessing, feature learning, evaluation) and distill comparative observations, including advantages of multimodal fusion, the persistent difficulty of separating stress from neutral states, and the limits of small-sample generalization into testable recommendations for future work.

4.1. Comparison of the Advantages and Disadvantages of Different Models and Usage Scenarios

Recent advancements in deep learning models have substantially improved data collection methods for stress detection and prediction. These models enable the remote quantification of stress through mobile device cameras and non-contact imaging techniques, effectively reducing the need for additional hardware and lowering system complexity. This provides a scalable and efficient solution for cognitive stress assessment. Particularly noteworthy is the feasibility of real-time monitoring using low-cost wearable devices, which enhances the practicality of these methods. The integration of non-invasive techniques, such as virtual reality simulations and single-channel EEG for stress classification, further highlights the convenience and ecological validity of these approaches [84]. Despite these advancements, the application of deep learning models to stress detection is not without challenges. A critical issue is the limited dataset size in many studies, which often involve small sample groups ranging from 12 to 45 participants [85]. Such limited datasets can impair the generalizability and robustness of the models, as they may fail to adequately represent diverse populations or real-world conditions. Moreover, experimental design is pivotal in determining the effectiveness of stress detection models. However, many studies are conducted in controlled environments, which, while ensuring precision, do not fully replicate the complexity of real-life situations. This discrepancy may diminish the ecological validity of the models [86,87], as they are less likely to capture the variability and unpredictability of stress responses encountered in everyday life. Data preprocessing is another area where deep learning models exhibit both strengths and weaknesses. Advanced preprocessing techniques, such as frequency domain feature extraction and multilayer perceptron classifiers, can significantly improve classification performance and accuracy [88]. However, the reproducibility of results is often complicated by the lack of detailed descriptions of preprocessing methods, signal acquisition processes, and the equipment used. The quality of physiological data, such as PPG and EEG signals, is frequently compromised by artifacts, motion noise, and limited signal clarity, which can adversely affect model accuracy and interpretability [87,89] The complexity of high-performance models, such as CNN-TLSTM with attention mechanisms and XGBoost, also presents a double-edged sword. While these models have demonstrated exceptional classification accuracies, CNN-TLSTM averaging 98.83% and XGBoost reaching up to 99.2% in stress detection [87,90], their computational demands may limit practical deployment, especially in resource-constrained environments. Another significant challenge lies in the generalizability of these models. The reliance on specific datasets or singular types of physiological signals often results in models that do not fully capture the diversity of stress responses across various scenarios [84,85]. This limitation underscores the need for more comprehensive, large-scale studies that can enhance the generalizability of these findings. While non-invasive methods, such as remote photoplethysmography and wearable devices, offer practical solutions for continuous stress assessment [88,89], making these methods more accessible by reducing the reliance on expensive equipment, the initial costs associated with deploying advanced deep learning models and acquiring high-quality physiological data can still be prohibitive for some settings. Finally, the quality of data remains a crucial factor in the performance of stress detection models. The reliability of physiological signals, like PPG and EEG, is often undermined by artifacts, motion noise, and limited clarity [84,89], leading to potential inaccuracies in stress assessments. Additionally, the complexity of interpreting subtle physiological changes, such as those in HRV, further complicates the analysis, potentially reducing the accuracy of these models.

4.2. Pitfalls and Challenges

In the analysis of physiological signals associated with stress and anxiety, several challenges and limitations can compromise the effectiveness of predictive models. Ensuring high signal quality is critical, as environmental noise and artifacts can introduce distortions that affect interpretation. Proper feature preprocessing is required to extract meaningful indicators from raw data, while controlling model complexity is essential to avoid issues such as overfitting or underfitting. Addressing these challenges is vital to enhance the reliability and accuracy of the results, ultimately contributing to more effective interventions and support for individuals experiencing stress and anxiety.

4.2.1. Pitfalls

Signal quality and environmental interference are critical for accurate physiological data analysis, yet several pitfalls can compromise it. Environmental interference, such as background noise and electromagnetic disturbances, can distort signals, making it challenging to extract meaningful information [91]. Improper sensor placement further exacerbates this issue, as even slight misalignments can lead to weak or distorted signals [92].
Movement artifacts and sensor placement introduce noise, particularly in wearable devices, skewing the readings during data collection [93]. Physiological variability among individuals, such as differences in EDA or HR, also complicates data interpretation and can affect overall signal quality [94]. Ensuring correct sensor placement and addressing movement-related noise is essential to maintaining signal accuracy.
Signal saturation and calibration occur when the signal exceeds the sensor’s maximum range, resulting in critical information loss [95]. Regular calibration of sensors is essential to mitigate these issues, as inadequate calibration can introduce biases [96]. Proper calibration ensures reliable data collection.
Data loss and connectivity issues from intermittent connections or hardware failures can create gaps in the analysis, leading to incomplete or skewed results [97]. Addressing hardware reliability and ensuring consistent data transmission can help minimize this pitfall and enhance the overall validity of the analysis.
Noise reduction and preprocessing are crucial steps in preparing physiological data for analysis, yet they pose pitfalls. Insufficient noise reduction can result in inadequate filtering, failing to eliminate artifacts and irrelevant variations, which can mislead results [98]. Proper filtering methods must be applied to improve signal clarity.
Feature selection, model complexity, and model overfitting/underfitting are common concerns. Overfitting occurs when too many or irrelevant features are selected, causing the model to learn noise rather than meaningful patterns, reducing generalizability to new data. Underfitting, on the other hand, overlooks important features, leading to missed opportunities for enhancing model performance [71]. A balanced feature selection process is key to improving the model’s generalization. Otherwise, balancing the complexity of models is crucial for optimal performance [99,100].
Inconsistent preprocessing techniques across datasets can create discrepancies that hinder effective comparison and analysis. Improper handling of missing data is another common pitfall; failing to address gaps appropriately can skew results and lead to inaccurate interpretations [98]. Standardizing preprocessing methods is crucial for reliable comparisons.
Scaling, normalization, and bias can significantly affect the analysis. Inconsistent application of scaling and normalization can introduce bias, allowing features with vastly different scales to dominate the analysis [100]. Ensuring consistent scaling and normalization reduces the risk of bias.
Hyperparameter tuning and resource constraints also pose difficulties. Complex models require more resources, including processing power and time, which can be barriers in real-time applications or when handling large datasets [99]. Moreover, complex models may be harder to interpret, complicating the understanding of predictions, which can hinder trust and usability. Hyperparameter tuning becomes more complicated with increased model complexity, as larger models have more parameters to optimize, leading to longer training times and a higher risk of suboptimal configurations [101].
Signal quality, environmental interference, artifacts, and signal saturation are well-recognized pitfalls in physiological signal processing, which remain nearly insurmountable even under current conditions. Nonetheless, researchers continue to refine models and methods within these constraints, striving to develop optimal predictive approaches that best accommodate these difficulties.

4.2.2. Challenges

The heterogeneity in study quality identified through our systematic assessment presents a fundamental challenge for evidence synthesis in this field. Our evaluation revealed significant variability in methodological rigor across the included studies, ranging from investigations with comprehensive validation procedures and transparent reporting to those with substantial methodological limitations. This variability directly impacts the interpretability of findings and complicates the drawing of robust conclusions from the literature.
Studies demonstrating more rigorous methodological approaches, including comprehensive validation strategies and detailed reporting of preprocessing procedures, often present more conservative performance estimates compared to investigations with methodological weaknesses. This pattern suggests that apparent differences in model performance across studies may partially reflect variations in methodological quality rather than genuine differences in analytical approaches or signal processing techniques. The observed heterogeneity in research quality underscores the urgent need for standardized methodological frameworks and reporting guidelines that would facilitate more meaningful comparison of results across investigations and enable more robust evidence synthesis.
Age and physiological responses present a significant challenge in mental health assessments. Age-related differences can influence both biological sex and gender identity, affecting physiological responses to stress and anxiety due to developmental changes, hormonal fluctuations, and varying life experiences [102]. For example, younger individuals may exhibit distinct coping mechanisms compared to older adults. Adolescents might experience heightened emotional responses driven by hormonal changes, while older adults often rely on more adaptive coping strategies shaped by their life experiences [103]. This variability complicates data interpretation and the development of age-appropriate interventions, requiring a tailored approach that accounts for these differences [104].
Gender identity adds another layer of complexity to physiological research and mental health assessments. Individuals who identify outside the traditional male–female binary may face unique stressors related to societal acceptance and discrimination, which can significantly impact their physiological responses to stress and anxiety [105]. Traditional assessments often fail to capture the experiences of non-binary or transgender individuals, leading to gaps in understanding and support [106]. Researchers must develop inclusive methodologies that validate diverse gender identities. By creating tailored assessment tools, we can better understand the physiological signals associated with anxiety and stress in these populations, ultimately leading to more effective interventions [107].
Medication use among study participants was assessed. The use of medications for clinical anxiety (e.g., benzodiazepines or antidepressants) or other chronic medical conditions (antihypertensives, antiarrhythmics, lipid-lowering agents, hypoglycaemic agents, anticoagulants, etc.) can significantly alter physiological signals related to emotional stress and anxiety, thereby compromising the validity of the analytical results [108]. It is crucial to include these parameters in data collection and to consider the possibility of recalling bias or intentional omission of this information by study participants [109].
The distinction between baseline and acute anxiety, characterized by the presence of autonomic symptoms (e.g., cold sweats, dizziness, bradycardia, and fluctuations in blood pressure), presents a methodological challenge in data acquisition and analysis. Given the dynamic nature of anxiety, an individual’s anxiety level can fluctuate rapidly, transitioning between states with varying physiological manifestations, which can complicate the interpretation of assessment results. Both basal and acute anxiety significantly contribute to emotional stress, but have distinct clinical implications. Consequently, establishing a standardized baseline of anxiety for all participants is desirable, but may be difficult to achieve in practice [110].
Intersectionality and socio-economic factors also play a critical role in shaping individuals’ experiences related to biological sex and gender identity. Factors such as race, socio-economic status, and cultural background can influence access to healthcare and support systems for managing stress and anxiety [26]. Individuals from marginalized communities often face unique stressors, including systemic discrimination and economic instability, which exacerbate mental health challenges [111]. Recognizing these intersectional influences is essential for developing comprehensive strategies that support individuals across different demographics [112].
In practical applications, several challenges warrant greater attention. First, the convenience of wearing and using wearable devices remains limited, which restricts their broader adoption. For example, EEG caps typically rely on wet electrodes that require conductive gel, and prolonged electrode pressure on the scalp may cause discomfort or even dizziness. Second, achieving real-time assessment of stress or anxiety demands sufficient battery life and computational capacity. At the same time, integrating multimodal data in real time is challenging, as signals from different sensors often vary in temporal resolution and sampling rates. Moreover, users may develop resistance to long-term monitoring, particularly those already experiencing stress or anxiety [113]. They may be reluctant to remain under continuous surveillance and concerned about the risk of data leakage leading to stigma or discrimination. Collectively, these practical issues constitute key bottlenecks that hinder the translation of research outcomes into real-world applications [114,115,116].
Cross-dataset generalization remains largely under-examined. Many studies report high within-dataset accuracy on benchmarks such as WESAD or SWELL, but few evaluate models across datasets or populations. Addressing this requires standardized protocols, cross-dataset validation as a default reporting practice, and techniques such as domain adaptation or self-supervised pretraining to improve robustness. In parallel, privacy-preserving learning paradigms (e.g., federated learning, secure aggregation, differential privacy) should be explored to enable training on distributed sensitive physiological data without centralizing identifiable information. Scaling models by training on substantially larger, more diverse physiological corpora, an approach that underpins the generalization of modern foundation models, may improve out-of-distribution performance, but is currently limited by data availability and privacy constraints.
Neural rights emerge as a critical issue with the rise of terotechnology and brain–computer interfaces. Protecting rights related to mental privacy, personal identity, and free will is becoming increasingly important [117]. Future research must explore these rights and contribute to developing policies and frameworks that ensure individuals maintain control over their neural data [118]. Addressing these privacy and data protection issues will help establish a secure and ethical foundation for the use of machine learning and physiological signal analysis in mental health [118].

4.3. Future Opportunities

Signal quality and data processing are critical aspects to address for improving early detection of anxiety and stress. Enhancing sensor calibration and developing more robust noise reduction techniques are necessary to mitigate issues such as signal distortion. This can help ensure that the data collected are more accurate and reliable, which is essential for effective analysis. Improving preprocessing methods, such as artifact removal and signal filtering, will further refine the data, leading to better model performance and more reliable physiological data analysis.
Personalization of models plays a crucial role in improving detection accuracy. Accounting for individual physiological variability, such as differences related to age, gender, and intersectionality, is essential to creating models that can accurately assess stress and anxiety across diverse populations. Personalized models can address the unique physiological responses of different individuals, leading to more accurate and tailored interventions [103,105].
Multimodal data integration offers significant potential for enhancing the accuracy of stress and anxiety detection. By combining different physiological signals, such as EEG, HRV, and EDA, through deep learning models, researchers can create a more comprehensive picture of an individual’s mental state. This multimodal approach allows for the identification of stress and anxiety from multiple angles, leading to more precise and reliable detection methods [108].
Inclusivity in assessment tools is another critical opportunity. Developing inclusive tools that account for diverse gender identities, races, medical histories, and socio-economic backgrounds will help ensure that mental health assessments are equitable and effective for all individuals. Addressing the unique stressors faced by marginalized communities, such as discrimination and economic instability, is key to creating comprehensive mental health strategies that support individuals across different demographics [112,113].
Innovative sensor technologies will enhance the precision and convenience of physiological signal collection. By embracing advanced wearable devices and sensor technologies, real-time monitoring of stress and anxiety can become more practical and accessible. This will drive the future of personalized healthcare, enabling continuous and unobtrusive data collection that can inform timely interventions and improve patient outcomes.
To address these intersectional challenges, future ML model development should incorporate demographic-aware architectures that explicitly account for population-specific variations in physiological responses. This could include implementing stratified validation frameworks where model performance is evaluated separately across different demographic subgroups, ensuring equitable accuracy across age, gender, race, socio-economic status, and cultural background. Additionally, fairness-constrained optimization techniques and adversarial debiasing methods could be employed to minimize discriminatory outcomes while maintaining overall model performance.
Demographic factors in signal modeling should also be considered. It is necessary to adopt richer, comparable metadata at data collection time (including race/ethnicity, socioeconomic indicators, and recruitment channels) and conduct stratified and sensitivity analyses to quantify their effects on performance and bias. Otherwise, advancing standardized statements of data usage boundaries (permitted reuse, redistribution limits, retention policies, and cross-institution transfer conditions) alongside transparency practices to improve auditability and reproducibility are essential. We encourage subsequent primary studies and dataset initiatives to publish clear demographic and ethical metadata.
Leveraging these opportunities will facilitate the development of more accurate tools for early identification, tracking, and management of anxiety and stress, thereby enhancing patient outcomes and promoting general health and well-being.

4.4. Methodological Recommendations

Based on the patterns identified through our systematic quality assessment, several key areas emerge as priorities for methodological improvement in future investigations. The development of more rigorous research standards in this field would substantially enhance the reliability and comparability of findings across studies.
Future studies would benefit significantly from implementing adequate sample size determination procedures, ideally supported by formal power calculations that account for the complexity of machine learning model development and validation. Such approaches would help ensure sufficient statistical power for detecting meaningful associations while avoiding the limitations associated with underpowered investigations that have characterized much of the current literature.
Transparent and comprehensive reporting of data preprocessing methodologies represents another critical area for improvement. Studies should provide detailed descriptions of artifact removal procedures, filtering techniques, and signal conditioning approaches to enable replication and facilitate meaningful comparison of results across investigations. The current variability in preprocessing methodology reporting substantially limits the reproducibility of findings and complicates evidence synthesis efforts.
The implementation of robust validation strategies, particularly those incorporating independent or external validation procedures, would address fundamental concerns regarding the generalizability of reported model performance. Cross-validation approaches, while valuable, should be complemented by evaluation on truly independent datasets to provide more realistic assessments of model performance in real-world applications.
Additionally, systematic consideration and reporting of demographic and clinical confounding factors, including medication use, comorbid conditions, and baseline physiological states, would enhance the validity of reported associations between physiological signals and anxiety or stress states. The current limited attention to such factors represents a significant methodological gap that may compromise the reliability of findings.
Finally, the adoption of standardized performance reporting metrics that extend beyond simple accuracy measures would facilitate more nuanced evaluation of model performance and enable more meaningful comparison of results across different analytical approaches and study populations.

5. Conclusions

A systematic review of 113 studies employing machine learning techniques to evaluate anxiety and stress based on physiological signals highlights its potential for early detection and intervention. These studies have investigated the efficacy of capturing objective physiological indicators, elucidating their role in mental health, exploring individual differences in susceptibility, and underscoring the significance of early predictions in preventing symptom exacerbation. Collectively, these findings suggest that current machine learning methodologies offer promising avenues for continuous monitoring and early prediction, which are essential for timely therapeutic interventions. Despite encouraging performance, progress is constrained by the lack of standardized, open benchmarks for stress and anxiety detection. Existing datasets differ in sensing modalities, sampling rates, labeling schemes, and elicitation paradigms, and many studies rely on proprietary data collected under heterogeneous protocols. Advancing the field requires harmonized datasets with standardized elicitation paradigms, controlled acquisition procedures that minimize environmental interference, and comprehensive metadata (e.g., demographics, medications, task context, ethics), alongside predefined cross-subject splits and external-validation procedures. Such benchmarks are essential for fair comparison, robust generalization, and reproducible improvement of current methods. Therefore, future research should not only explore the relationships between physiological signals and mental health, but also focus on developing models that perform well across diverse individuals and enhancing metrics for cross-subject detection. Innovations should include advanced predictive models and preventive strategies, alongside integrating cutting-edge sensor technologies and real-time data analysis into clinical practice. This proactive approach will improve the practicality and effectiveness of machine learning methods, ultimately enhancing individual well-being and mitigating the impact of mental health issues.

Author Contributions

Conceptualization, M.-T.H. and H.Z.; methodology, Y.L., T.B. and Y.O.; validation, H.Z. and M.-T.H.; investigation, Y.L., M.-I.P., C.T., Y.O., Z.L., Z.W. and F.T.; writing—original draft preparation, Y.L., M.-I.P., T.B., C.T., H.Z. and M.-T.H.; writing—review and editing, M.-I.P., C.T. and M.-T.H.; visualization, M.-I.P. and M.-T.H.; supervision, H.Z. and M.-T.H.; project administration, M.-T.H.; funding acquisition, H.Z. and M.-T.H. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the NSFC with grant No. 62076083, Zhejiang Provincial Natural Science Foundation of China under Grant No. ZCLZ24F0301; and by grants from the Spanish Agencia Estatal de Investigación, Ministerio de Ciencia, Innovación y Universidades, co-funded by European Union, Next Generation EU/PRTR Ref. Number: TED2021-130942B-C21//MICIU/AEI/10.13039/501100011033; the Fundación Primafrio with the code number 39747; COST Participatory Approaches with Older Adults (PAAR-net), with the code number CA22167, and China-Spain AI Technology Joint Laboratory with Hangzhou Dianzi University.

Conflicts of Interest

The authors declare no conflicts of interest.

Appendix A

Search Terms Used for Identifying Relevant Studies
This table lists the search terms used to identify studies related to physiological signals and machine learning. Terms are categorized by physiological signal types including EEG (Electroencephalogram), eye-tracking technologies, heart rate measurements, skin conductance, and blood pressure indicators. Variants and synonyms of each term were included to ensure a comprehensive search of the literature.
  • (((“Anxiety/classification” [Mesh] OR “Anxiety/diagnosis” [Mesh] OR “Stress, Psychological/classification” [Mesh] OR “Stress, Psychological/diagnosis” [Mesh])
  • AND (“Deep Learning” [Mesh] OR “Machine Learning” [Mesh] OR “Neural Networks, Computer” [Mesh]))
  • AND (“Electroencephalography” [Mesh] OR “Eye-Tracking Technology” [Mesh]
  • OR “Heart Rate” [Mesh] OR “Galvanic Skin Response” [Mesh] OR “Blood Pressure” [Mesh] OR “Oxygen Saturation” [Mesh]
  • OR “Biomarkers” [Mesh]))
  • OR
  • (((((Angst [Title/Abstract]) OR (Nervousness [Title/Abstract]) OR (Hypervigilance [Title/Abstract]) OR (Social Anxiety [Title/Abstract]) OR (Anxieties [Title/Abstract])
  • OR (Anxiety [Title/Abstract]) OR (Social Anxieties [Title/Abstract]) OR (Anxiousness [Title/Abstract]) OR (Psychological Stresses [Title/Abstract])
  • OR (Stressor [Title/Abstract]) OR (Psychological Stressor [Title/Abstract]) OR (Psychological Stressors [Title/Abstract])
  • OR (Psychological Stress [Title/Abstract]) OR (Stress [Title/Abstract]) OR (Psychologic Stress [Title/Abstract]) OR (Life Stress [Title/Abstract])
  • OR (Life Stresses [Title/Abstract]) OR (Stress [Title/Abstract]) OR (Emotional Stress [Title/Abstract])
  • OR (Stressful Conditions [Title/Abstract]) OR (Psychological Stressful Condition [Title/Abstract])
  • OR (Psychological Stressful Conditions [Title/Abstract]) OR (Stressful Condition [Title/Abstract]) OR (Psychologically Stressful Conditions [Title/Abstract])
  • OR (Psychologically Stressful Condition [Title/Abstract])
  • OR (Stressful Condition [Title/Abstract]) OR (Stress Experience [Title/Abstract]) OR (Psychological Stress Experience [Title/Abstract])
  • OR (Psychological Stress Experiences [Title/Abstract]) OR (Stress Experiences [Title/Abstract])
  • OR (Individual Stressors [Title/Abstract]) OR (Individual Stressor [Title/Abstract]) OR (Cumulative Stress [Title/Abstract])
  • OR (Cumulative Stresses [Title/Abstract]) OR (Psychological Cumulative Stresses [Title/Abstract])
  • OR (Psychological Cumulative Stress [Title/Abstract]) OR (Stress Overload [Title/Abstract]) OR (Psychological Stress Overload [Title/Abstract])
  • OR (Psychological Stress Overloads [Title/Abstract]) OR (Stress Measurement [Title/Abstract])
  • OR (Psychological Stress Measurement [Title/Abstract]) OR (Psychological Stress Measurements [Title/Abstract])
  • OR (Stress Processes [Title/Abstract]) OR (Psychological Stress Processe [Title/Abstract]) OR (Psychological Stress Processes [Title/Abstract])
  • OR (Stress Processe [Title/Abstract]))
  • AND ((classif * [Title/Abstract]) OR (diagnos * [Title/Abstract]) OR (assess * [Title/Abstract]) OR (predict * [Title/Abstract])))
  • AND ((Deep Learning [Title/Abstract]) OR (Hierarchical Learning [Title/Abstract]) OR (Machine Learning [Title/Abstract]) OR (Computer Neural Network [Title/Abstract])
  • OR (Computer Neural Networks [Title/Abstract])
  • OR (Neural Network [Title/Abstract]) OR (Deep learning
  • Model [Title/Abstract]) OR (Model [Title/Abstract]) OR (Network Model [Title/Abstract])
  • OR (Network Models [Title/Abstract]) OR (Neural Network Model [Title/Abstract]) OR (Neural Network Models [Title/Abstract]) OR
  • (Connectionist Models [Title/Abstract])
  • OR (Connectionist Model [Title/Abstract]) OR (Perceptrons [Title/Abstract])
  • OR (Perceptron [Title/Abstract]) OR (Computational Neural Networks [Title/Abstract]) OR (Computational Neural Network [Title/Abstract])
  • OR (Neural Network [Title/Abstract])
  • AND ((EEG [Title/Abstract]) OR (Electroencephalogram [Title/Abstract])
  • OR (Electroencephalograms [Title/Abstract]) OR (Eye-Tracking Technologies [Title/Abstract]) OR (Eye Tracking Technology [Title/Abstract])
  • OR (Eyetracking Technology [Title/Abstract]) OR (Eyetracking Technologies [Title/Abstract]) OR (Gaze-Tracking Technology [Title/Abstract])
  • OR (Gaze-Tracking Technologies [Title/Abstract])
  • OR (Gaze Tracking Technology [Title/Abstract]) OR (Gaze-Tracking System [Title/Abstract]) OR (Gaze Tracking System [Title/Abstract])
  • OR (Gaze-Tracking Systems [Title/Abstract])
  • OR (Gazetracking System [Title/Abstract]) OR (Gazetracking Systems [Title/Abstract]) OR (Eye-Tracking System [Title/Abstract])
  • OR (Eye Tracking System [Title/Abstract])
  • OR (Eye-Tracking Systems [Title/Abstract]) OR (Eyetracking System [Title/Abstract]) OR (Eyetracking Systems [Title/Abstract]) OR (Eye-Tracking [Title/Abstract])
  • OR (Heart Rates [Title/Abstract]) OR (Cardiac Rate [Title/Abstract]) OR (Cardiac Rates [Title/Abstract])
  • OR (Pulse Rate [Title/Abstract]) OR (Pulse Rates [Title/Abstract]) OR (Heartbeat [Title/Abstract]) OR (Heartbeats [Title/Abstract])
  • OR (Cardiac Chronotropism [Title/Abstract]) OR (Heart Rate Control [Title/Abstract])
  • OR (Cardiac Chronotropy [Title/Abstract]) OR (Galvanic Skin Responses [Title/Abstract])
  • OR (Psychogalvanic Reflex [Title/Abstract]) OR (Skin Electric Conductance [Title/Abstract])
  • OR (Electrodermal Response [Title/Abstract]) OR (Electrodermal Responses [Title/Abstract])
  • OR (Pulse Pressure [Title/Abstract]) OR (Diastolic Pressure [Title/Abstract])
  • OR (Systolic Pressure [Title/Abstract])
  • OR (Saturation of Peripheral Oxygen [Title/Abstract])
  • OR (Peripheral Oxygen Saturation [Title/Abstract]) OR (SpO2 [Title/Abstract])))

References

  1. Newlove-Delgado, T.; Marcheselli, F.; Williams, T.; Mandalia, D.; Davis, J.; McManus, S.; Savic, M.; Treloar, W.; Ford, T. Mental Health of Children and Young People in England 2022—Wave 3 Follow up to the 2017 Survey; NHS Digital: London, UK, 2022. [Google Scholar]
  2. World Health Organization. Mental Health and COVID-19: Early Evidence of the Pandemic’s Impact: Scientific Brief, 2 March 2022. Available online: https://www.who.int/publications/i/item/WHO-2019-nCoV-Sci_Brief-Mental_health-2022.1 (accessed on 4 November 2024).
  3. American Psychiatric Association. Mental Health Disparities: Women’s Mental Health. 2017. Available online: https://www.psychiatry.org (accessed on 4 November 2024).
  4. Bienertova-Vasku, J.; Lenart, P.; Scheringer, M. Eustress and Distress: Neither Good Nor Bad, but Rather the Same? BioEssays 2020, 42, e1900238. [Google Scholar] [CrossRef]
  5. de la Fuente, J.; Paoloni, P.V.; Vera-Martínez, M.M.; Garzón-Umerenkova, A. Effect of Levels of Self-Regulation and Situational Stress on Achievement Emotions in Undergraduate Students: Class, Study and Testing. Int. J. Environ. Res. Public Health 2020, 17, 4293. [Google Scholar] [CrossRef]
  6. Roberts, B.L.; Karatsoreos, I.N. Brain-body responses to chronic stress: A brief review. Fac. Rev. 2021, 10, 83. [Google Scholar] [CrossRef]
  7. Bui, T.; Zackula, R.; Dugan, K.; Ablah, E. Workplace Stress and Productivity: A Cross-Sectional Study. Kans. J. Med. 2021, 14, 42. [Google Scholar] [CrossRef]
  8. Lau, K.K.H.; Randall, A.K.; Duran, N.D.; Tao, C. Examining the Effects of Couples’ Real-Time Stress and Coping Processes on Interaction Quality: Language Use as a Mediator. Front. Psychol. 2019, 9, 2598. [Google Scholar] [CrossRef]
  9. Ross, R.A.; Foster, S.L.; Ionescu, D.F. The Role of Chronic Stress in Anxious Depression. Chronic Stress. 2017, 1, 2470547016689472. [Google Scholar] [CrossRef]
  10. Satyjeet, F.; Naz, S.; Kumar, V.; Aung, N.H.; Bansari, K.; Irfan, S.; Rizwan, A. Psychological Stress as a Risk Factor for Cardiovascular Disease: A Case-Control Study. Cureus 2020, 12, e10757. [Google Scholar] [CrossRef]
  11. Sharma, L.D.; Bohat, V.K.; Habib, M.; Al-Zoubi, A.M.; Faris, H.; Aljarah, I. Evolutionary inspired approach for mental stress detection using EEG signal. Expert Syst. Appl. 2022, 197, 116634. [Google Scholar] [CrossRef]
  12. Seiler, A.; Fagundes, C.P.; Christian, L.M. The Impact of Everyday Stressors on the Immune System and Health. In Stress Challenges and Immunity in Space: From Mechanisms to Monitoring and Preventive Strategies; Springer International Publishing: Cham, Switzerland, 2020; pp. 71–92. [Google Scholar] [CrossRef]
  13. König, H.; König, H.-H.; Gallinat, J.; Lambert, M.; Karow, A.; Peth, J.; Schulz, H.; Konnopka, A. Excess costs of mental disorders by level of severity. Soc. Psychiatry Psychiatr. Epidemiol. 2023, 58, 973–985. [Google Scholar] [CrossRef]
  14. Müller, G.; Bombana, M.; Heinzel-Gutenbrenner, M.; Kleindienst, N.; Bohus, M.; Lyssenko, L.; Vonderlin, R. Socio-economic consequences of mental distress: Quantifying the impact of self-reported mental distress on the days of incapacity to work and medical costs in a two-year period: A longitudinal study in Germany. BMC Public Health 2021, 21, 625. [Google Scholar] [CrossRef]
  15. Smith, J.P.; Book, S.W. Anxiety and Substance Use Disorders: A Review. Psychiatr. Times 2008, 25, 19. [Google Scholar]
  16. Lazarou, E.; Exarchos, T.P. Predicting stress levels using physiological data: Real-time stress prediction models utilizing wearable devices. AIMS Neurosci. 2024, 11, 76. [Google Scholar] [CrossRef]
  17. Vos, G.; Trinh, K.; Sarnyai, Z.; Azghadi, M.R. Generalizable machine learning for stress monitoring from wearable devices: A systematic literature review. Int. J. Med. Inform. 2023, 173, 105026. [Google Scholar] [CrossRef]
  18. Hellhammer, D.H.; Wüst, S.; Kudielka, B.M. Salivary cortisol as a biomarker in stress research. Psychoneuroendocrinology 2009, 34, 163–171. [Google Scholar] [CrossRef]
  19. Tadiri, C.P.; Raparelli, V.; Abrahamowicz, M.; Kautzy-Willer, A.; Kublickiene, K.; Herrero, M.-T.; Norris, C.M.; Pilote, L. Methods for prospectively incorporating gender into health sciences research. J. Clin. Epidemiol. 2021, 129, 191–197. [Google Scholar] [CrossRef]
  20. Raparelli, V.; Norris, C.M.; Bender, U.; Herrero, M.T.; Kautzky-Willer, A.; Kublickiene, K.; El Emam, K.; Pilote, L. Identification and inclusion of gender factors in retrospective cohort studies: The GOING-FWD framework. BMJ Glob. Health 2021, 6, e005413. [Google Scholar] [CrossRef]
  21. Liberati, A.; Altman, D.G.; Tetzlaff, J.; Mulrow, C.; Gøtzsche, P.C.; Ioannidis, J.P.; Clarke, M.; Devereaux, P.J.; Kleijnen, J.; Moher, D. The PRISMA statement for reporting systematic reviews and meta-analyses of studies that evaluate healthcare interventions: Explanation and elaboration. BMJ 2009, 339, b2700. [Google Scholar] [CrossRef]
  22. Al-Jebrni, A.H.; Chwyl, B.; Wang, X.Y.; Wong, A.; Saab, B.J. AI-enabled remote and objective quantification of stress at scale. Biomed. Signal Process. Control 2020, 59, 101929. [Google Scholar] [CrossRef]
  23. Kumar, M.; Zhang, W.; Weippert, M.; Freudenthaler, B. An Explainable Fuzzy Theoretic Nonparametric Deep Model for Stress Assessment Using Heartbeat Intervals Analysis. IEEE Trans. Fuzzy Syst. 2021, 29, 3873–3886. [Google Scholar] [CrossRef]
  24. Lee, S.; Bin Hwang, H.; Park, S.; Kim, S.; Ha, J.H.; Jang, Y.; Hwang, S.; Park, H.-K.; Lee, J.; Kim, I.Y. Mental Stress Assessment Using Ultra Short Term HRV Analysis Based on Non-Linear Method. Biosensors 2022, 12, 465. [Google Scholar] [CrossRef]
  25. Jiao, Y.; Sun, Z.; Fu, L.; Yu, X.; Jiang, C.; Zhang, X.; Liu, K.; Chen, X. Physiological responses and stress levels of high-speed rail train drivers under various operating conditions—A simulator study in China. Int. J. Rail Transp. 2023, 11, 449–464. [Google Scholar] [CrossRef]
  26. Barua, S.; Begum, S.; Ahmed, M.U. Supervised machine learning algorithms to diagnose stress for vehicle drivers based on physiological sensor signals. Stud. Health Technol. Inform. 2015, 211, 241–248. [Google Scholar] [CrossRef]
  27. Micoulaud-Franchi, J.A.; Jeunet, C.; Pelissolo, A.; Ros, T. EEG Neurofeedback for Anxiety Disorders and Post-Traumatic Stress Disorders: A Blueprint for a Promising Brain-Based Therapy. Curr. Psychiatry Rep. 2021, 23, 84. [Google Scholar] [CrossRef]
  28. Rao, R.P.N. Brain-Computer Interfacing: An Introduction; Cambridge University Press: New York, NY, USA, 2013. [Google Scholar] [CrossRef]
  29. Jurcak, V.; Tsuzuki, D.; Dan, I. 10/20, 10/10, and 10/5 systems revisited: Their validity as relative head-surface-based positioning systems. Neuroimage 2007, 34, 1600–1611. [Google Scholar] [CrossRef]
  30. Yen, C.; Lin, C.L.; Chiang, M.C. Exploring the Frontiers of Neuroimaging: A Review of Recent Advances in Understanding Brain Functioning and Disorders. Life 2023, 13, 1472. [Google Scholar] [CrossRef]
  31. Jiang, X.; Bian, G.B.; Tian, Z. Removal of Artifacts from EEG Signals: A Review. Sensors 2019, 19, 987. [Google Scholar] [CrossRef]
  32. Murty, P.S.C.; Anuradha, C.; Naidu, P.A.; Balaswamy, C.; Nagalingam, R.; Jagatheesaperumal, S.K.; Ponnusamy, M. An intelligent wearable embedded architecture for stress detection and psychological behavior monitoring using heart rate variability. J. Intell. Fuzzy Syst. 2023, 45, 8203–8216. [Google Scholar] [CrossRef]
  33. Boonnithi, S.; Phongsuphap, S. Comparison of Heart Rate Variability Measures for Mental Stress Detect. 2011. Available online: https://mahidol.elsevierpure.com/en/publications/comparison-of-heart-rate-variability-measures-for-mental-stress-d (accessed on 5 November 2024).
  34. Tiwari, R.; Kumar, R.; Malik, S.; Raj, T.; Kumar, P. Analysis of Heart Rate Variability and Implication of Different Factors on Heart Rate Variability. Curr. Cardiol. Rev. 2021, 17, e160721189770. [Google Scholar] [CrossRef]
  35. Kim, K.B.; Baek, H.J. Photoplethysmography in Wearable Devices: A Comprehensive Review of Technological Advances, Current Challenges, and Future Directions. Electronics 2023, 12, 2923. [Google Scholar] [CrossRef]
  36. Peng, R.C.; Zhou, X.L.; Lin, W.H.; Zhang, Y.T. Extraction of heart rate variability from smartphone photoplethysmograms. Comput. Math. Methods Med. 2015, 2015, 516826. [Google Scholar] [CrossRef]
  37. Schiweck, C.; Piette, D.; Berckmans, D.; Claes, S.; Vrieze, E. Heart rate and high frequency heart rate variability during stress as biomarker for clinical depression. A systematic review. Psychol. Med. 2019, 49, 200–211. [Google Scholar] [CrossRef]
  38. Li, X.; Zou, L.; Li, H. Multilayer Perceptron-Based Wearable Exercise-Related Heart Rate Variability Predicts Anxiety and Depression in College Students. Sensors 2024, 24, 4203. [Google Scholar] [CrossRef]
  39. Benchekroun, M.; Velmovitsky, P.E.; Istrate, D.; Zalc, V.; Morita, P.P.; Lenne, D. Cross Dataset Analysis for Generalizability of HRV-Based Stress Detection Models. Sensors 2023, 23, 1807. [Google Scholar] [CrossRef]
  40. Ihianle, I.K.; Machado, P.; Owa, K.; Adama, D.A.; Otuka, R.; Lotfi, A. Minimising redundancy, maximising relevance: HRV feature selection for stress classification. Expert Syst. Appl. 2024, 239, 122490. [Google Scholar] [CrossRef]
  41. Khomidov, M.; Lee, D.; Kim, C.H.; Lee, J.H. The Real-Time Image Sequences-Based Stress Assessment Vision System for Mental Health. Electronics 2024, 13, 2180. [Google Scholar] [CrossRef]
  42. Bahameish, M.; Stockman, T.; Carrión, J.R. Strategies for Reliable Stress Recognition: A Machine Learning Approach Using Heart Rate Variability Features. Sensors 2024, 24, 3210. [Google Scholar] [CrossRef]
  43. Markiewicz, R.; Markiewicz-Gospodarek, A.; Dobrowolska, B. Galvanic Skin Response Features in Psychiatry and Mental Disorders: A Narrative Review. Int. J. Environ. Res. Public Health 2022, 19, 13428. [Google Scholar] [CrossRef]
  44. Klimek, A.; Mannheim, I.; Schouten, G.; Wouters, E.J.; Peeters, M.W.H. Wearables measuring electrodermal activity to assess perceived stress in care: A scoping review. Acta Neuropsychiatr. 2023, 37, e19. [Google Scholar] [CrossRef]
  45. Campanella, S.; Altaleb, A.; Belli, A.; Pierleoni, P.; Palma, L. A Method for Stress Detection Using Empatica E4 Bracelet and Machine-Learning Techniques. Sensors 2023, 23, 3565. [Google Scholar] [CrossRef]
  46. Schumm, J.; Bächlin, M.; Setz, C.; Arnrich, B.; Roggen, D.; Tröster, G. Effect of movements on the electrodermal response after a startle event. Methods Inf. Med. 2008, 47, 186–191. [Google Scholar] [CrossRef]
  47. Bari, D.S.; Aldosky, H.Y.Y.; Tronstad, C.; Martinsen, Ø.G. The correlations among the skin conductance features responding to physiological stress stimuli. Skin Res. Technol. 2021, 27, 582–588. [Google Scholar] [CrossRef]
  48. van der Mee, D.J.; Gevonden, M.J.; Westerink, J.H.D.M.; de Geus, E.J.C. Validity of electrodermal activity-based measures of sympathetic nervous system activity from a wrist-worn device. Int. J. Psychophysiol. 2021, 168, 52–64. [Google Scholar] [CrossRef]
  49. Bornoiu, I.V.; Grigore, O. Kohonen neural network stress detection using only electrodermal activity features. Adv. Electr. Comput. Eng. 2014, 14, 71–78. [Google Scholar] [CrossRef]
  50. Sánchez-Reolid, R.; Martínez-Rodrigo, A.; López, M.T.; Fernández-Caballero, A. Deep Support Vector Machines for the Identification of Stress Condition from Electrodermal Activity. Int. J. Neural Syst. 2020, 30, 2050031. [Google Scholar] [CrossRef]
  51. Armstrong, T.; Olatunji, B.O. Eye tracking of attention in the affective disorders: A meta-analytic review and synthesis. Clin. Psychol. Rev. 2012, 32, 704–723. [Google Scholar] [CrossRef]
  52. Bar-Haim, Y.; Lamy, D.; Pergamin, L.; Bakermans-Kranenburg, M.J.; Van Ijzendoorn, M.H. Threat-related attentional bias in anxious and nonanxious individuals: A meta-analytic study. Psychol. Bull. 2007, 133, 1–24. [Google Scholar] [CrossRef]
  53. Cisler, J.M.; Koster, E.H.W. Mechanisms of attentional biases towards threat in anxiety disorders: An integrative review. Clin. Psychol. Rev. 2010, 30, 203–216. [Google Scholar] [CrossRef]
  54. Rojas-Thomas, F.; Artigas, C.; Wainstein, G.; Morales, J.-P.; Arriagada, M.; Soto, D.; Dagnino-Subiabre, A.; Silva, J.; Lopez, V. Impact of acute psychosocial stress on attentional control in humans. A study of evoked potentials and pupillary response. Neurobiol. Stress 2023, 25, 100551. [Google Scholar] [CrossRef]
  55. Ahmadi, N.; Sasangohar, F.; Yang, J.; Yu, D.; Danesh, V.; Klahn, S.; Masud, F. Quantifying Workload and Stress in Intensive Care Unit Nurses: Preliminary Evaluation Using Continuous Eye-Tracking. Hum. Factors 2024, 66, 714–728. [Google Scholar] [CrossRef]
  56. Pedrotti, M.; Mirzaei, M.A.; Tedesco, A.; Chardonnet, J.-R.; Mérienne, F.; Benedetto, S.; Baccino, T. Automatic Stress Classification with Pupil Diameter Analysis. Int. J. Hum. Comput. Interact. 2014, 30, 220–236. [Google Scholar] [CrossRef]
  57. Rakhmatulin, I. A Review of the Low-Cost Eye-Tracking Systems for 2010–2020. SSRN Electron. J. 2020. [Google Scholar] [CrossRef]
  58. Yousefi, M.S.; Reisi, F.; Daliri, M.R.; Shalchyan, V. Stress Detection Using Eye Tracking Data: An Evaluation of Full Parameters. IEEE Access 2022, 10, 118941–118952. [Google Scholar] [CrossRef]
  59. Zanetti, M.; Mizumoto, T.; Faes, L.; Fornaser, A.; De Cecco, M.; Maule, L.; Valente, M.; Nollo, G. Multilevel assessment of mental stress via network physiology paradigm using consumer wearable devices. J. Ambient. Intell. Humaniz. Comput. 2021, 12, 4409–4418. [Google Scholar] [CrossRef]
  60. Xu, C.; Song, Y.; Sempionatto, J.R.; Solomon, S.A.; Yu, Y.; Nyein, H.Y.Y.; Tay, R.Y.; Li, J.; Heng, W.; Min, J.; et al. A physicochemical-sensing electronic skin for stress response monitoring. Nat. Electron. 2024, 7, 168–179. [Google Scholar] [CrossRef]
  61. Xia, L.; Feng, Y.; Guo, Z.; Ding, J.; Li, Y.; Li, Y.; Ma, M.; Gan, G.; Xu, Y.; Luo, J.; et al. MuLHiTA: A Novel Multiclass Classification Framework with Multibranch LSTM and Hierarchical Temporal Attention for Early Detection of Mental Stress. IEEE Trans. Neural Netw. Learn. Syst. 2023, 34, 9657–9670. [Google Scholar] [CrossRef]
  62. Saeed, S.M.U.; Anwar, S.M.; Majid, M.; Awais, M.; Alnowami, M. Selection of Neural Oscillatory Features for Human Stress Classification with Single Channel EEG Headset. Biomed. Res. Int. 2018, 2018, 1049257. [Google Scholar] [CrossRef]
  63. Udhayakumar, R.; Rahman, S.; Gopakumar, S.; Karmakar, C. Nonlinear Features from Multi-Modal Signals for Continuous Stress Monitoring. In Proceedings of the 2023 45th Annual International Conference of the IEEE Engineering in Medicine & Biology Society (EMBC), Sydney, NSW, Australia, 24–27 July 2023. [Google Scholar] [CrossRef]
  64. Tzevelekakis, K.; Stefanidi, Z.; Margetis, G. Real-Time Stress Level Feedback from Raw Ecg Signals for Personalised, Context-Aware Applications Using Lightweight Convolutional Neural Network Architectures. Sensors 2021, 21, 7802. [Google Scholar] [CrossRef]
  65. Tutunji, R.; Kogias, N.; Kapteijns, B.; Krentz, M.; Krause, F.; Vassena, E.; Hermans, E.J. Detecting Prolonged Stress in Real Life Using Wearable Biosensors and Ecological Momentary Assessments: Naturalistic Experimental Study. J. Med Internet Res. 2023, 25, e39995. [Google Scholar] [CrossRef]
  66. Tsai, Y.H.; Wu, S.K.; Yu, S.S.; Tsai, M.H. Analyzing Brain Waves of Table Tennis Players with Machine Learning for Stress Classification. Appl. Sci. 2022, 12, 8052. [Google Scholar] [CrossRef]
  67. Mohammadi, A.; Fakharzadeh, M.; Baraeinejad, B. An Integrated Human Stress Detection Sensor Using Supervised Algorithms. IEEE Sens. J. 2022, 22, 8216–8223. [Google Scholar] [CrossRef]
  68. Can, Y.S.; Chalabianloo, N.; Ekiz, D.; Fernandez-Alvarez, J.; Repetto, C.; Riva, G.; Iles-Smith, H.; Ersoy, C. Real-Life Stress Level Monitoring Using Smart Bands in the Light of Contextual Information. IEEE Sens. J. 2020, 20, 8721–8730. [Google Scholar] [CrossRef]
  69. Mukherjee, P.; Roy, A.H. A deep learning-based approach for distinguishing different stress levels of human brain using EEG and pulse rate. Comput. Methods Biomech. Biomed. Eng. 2023, 27, 2303–2324. [Google Scholar] [CrossRef]
  70. Gonzalez-Vazquez, J.J.; Bernat, L.; Ramon, J.L.; Morell, V.; Ubeda, A. A Deep Learning Approach to Estimate Multi-Level Mental Stress from EEG Using Serious Games. IEEE J. Biomed. Health Inform. 2024, 28, 3965–3972. [Google Scholar] [CrossRef]
  71. Cui, J.; Wilke, M.; Logothetis, N.K.; Leopold, D.A.; Liang, H. Visibility states modulate microsaccade rate and direction. Vis. Res. 2009, 49, 228–236. [Google Scholar] [CrossRef]
  72. Cho, H.M.; Park, H.; Dong, S.Y.; Youn, I. Ambulatory and Laboratory Stress Detection Based on Raw Electrocardiogram Signals Using a Convolutional Neural Network. Sensors 2019, 19, 4408. [Google Scholar] [CrossRef]
  73. Rakhmatulin, I.; Dao, M.S.; Nassibi, A.; Mandic, D. Exploring Convolutional Neural Network Architectures for EEG Feature Extraction. Sensors 2024, 24, 877. [Google Scholar] [CrossRef]
  74. Karthikeyan, P.; Murugappan, M.; Yaacob, S. Analysis of stroop colorword test-based human stress detection using electrocardiography and heart rate variability signals. Arab. J. Sci. Eng. 2012, 39, 1835–1847. [Google Scholar] [CrossRef]
  75. Luo, X.; Zhou, B.; Fang, J.; Cherif-Riahi, Y.; Li, G.; Shen, X. Integrating EEG and Ensemble Learning for Accurate Grading and Quantification of Generalized Anxiety Disorder: A Novel Diagnostic Approach. Diagnostics 2024, 14, 1122. [Google Scholar] [CrossRef]
  76. Betti, S.; Lova, R.M.; Rovini, E.; Acerbi, G.; Santarelli, L.; Cabiati, M.; Del Ry, S.; Cavallo, F. Evaluation of an integrated system of wearable physiological sensors for stress monitoring in working environments by using biological markers. IEEE Trans. Biomed. Eng. 2018, 65, 1748–1758. [Google Scholar] [CrossRef]
  77. Tiwari, S.; Agarwal, S. A Shrewd Artificial Neural Network-Based Hybrid Model for Pervasive Stress Detection of Students Using Galvanic Skin Response and Electrocardiogram Signals. Big Data 2021, 9, 427–442. [Google Scholar] [CrossRef]
  78. Taye, M.M. Understanding of Machine Learning with Deep Learning: Architectures, Workflow, Applications and Future Directions. Computers 2023, 12, 91. [Google Scholar] [CrossRef]
  79. Taskasaplidis, G.; Fotiadis, D.A.; Bamidis, P.D. Review of Stress Detection Methods Using Wearable Sensors. IEEE Access 2024, 12, 38219–38246. [Google Scholar] [CrossRef]
  80. Shen, Z.; Li, G.; Fang, J.; Zhong, H.; Wang, J.; Sun, Y.; Shen, X. Aberrated Multidimensional EEG Characteristics in Patients with Generalized Anxiety Disorder: A Machine-Learning Based Analysis Framework. Sensors 2022, 22, 5420. [Google Scholar] [CrossRef]
  81. Omar, S.M.; Kimwele, M.; Olowolayemo, A.; Kaburu, D.M. Enhancing EEG signals classification using LSTM-CNN architecture. Eng. Rep. 2024, 6, e12827. [Google Scholar] [CrossRef]
  82. Huang, J.; Luo, X.; Peng, X. A Novel Classification Method for a Driver’s Cognitive Stress Level by Transferring Interbeat Intervals of the ECG Signal to Pictures. Sensors 2020, 20, 1340. [Google Scholar] [CrossRef] [PubMed]
  83. Masood, K.; Alghamdi, M.A. Modeling Mental Stress Using a Deep Learning Framework. IEEE Access 2019, 7, 68446–68454. [Google Scholar] [CrossRef]
  84. Quitadamo, L.R.; Cavrini, F.; Sbernini, L.; Riillo, F.; Bianchi, L.; Seri, S.; Saggio, G. Support vector machines to detect physiological patterns for EEG and EMG-based human-computer interaction: A review. J. Neural Eng. 2017, 14, 011001. [Google Scholar] [CrossRef]
  85. Subhani, A.R.; Mumtaz, W.; Saad, M.N.B.M.; Kamel, N.; Malik, A.S. Machine learning framework for the detection of mental stress at multiple levels. IEEE Access 2017, 5, 13545–13556. [Google Scholar] [CrossRef]
  86. Perpetuini, D.; Chiarelli, A.M.; Cardone, D.; Filippini, C.; Rinella, S.; Massimino, S.; Bianco, F.; Bucciarelli, V.; Vinciguerra, V.; Fallica, P.; et al. Prediction of state anxiety by machine learning applied to photoplethysmography data. PeerJ 2021, 9, e10448. [Google Scholar] [CrossRef]
  87. Ströhle, A.; Gensichen, J.; Domschke, K. The Diagnosis and Treatment of Anxiety Disorders. Dtsch. Arztebl. Int. 2018, 155, 611–620. [Google Scholar] [CrossRef]
  88. Sun, M.; Cao, X. A Mental Stress Classification Method Based on Feature Fusion Using Physiological Signals. J. Circuits Syst. Comput. 2023, 33, 2450016. [Google Scholar] [CrossRef]
  89. Shishido, E.; Ogawa, S.; Miyata, S.; Yamamoto, M.; Inada, T.; Ozaki, N. Application of eye trackers for understanding mental disorders: Cases for schizophrenia and autism spectrum disorder. Neuropsychopharmacol. Rep. 2019, 39, 72–77. [Google Scholar] [CrossRef] [PubMed]
  90. Salankar, N.; Koundal, D.; Qaisar, S.M. Stress Classification by Multimodal Physiological Signals Using Variational Mode Decomposition and Machine Learning. J. Healthc. Eng. 2021, 2021, 2146369. [Google Scholar] [CrossRef] [PubMed]
  91. Fernández, J.R.M.; Anishchenko, L. Mental stress detection using bioradar respiratory signals. Biomed. Signal Process. Control 2018, 43, 244–249. [Google Scholar] [CrossRef]
  92. Fang, J.; Li, G.; Xu, W.; Liu, W.; Chen, G.; Zhu, Y.; Luo, Y.; Luo, X.; Zhou, B. Exploring Abnormal Brain Functional Connectivity in Healthy Adults, Depressive Disorder, and Generalized Anxiety Disorder through EEG Signals: A Machine Learning Approach for Triple Classification. Brain Sci. 2024, 14, 245. [Google Scholar] [CrossRef]
  93. Elgendi, M.; Menon, C. Machine Learning Ranks ECG as an Optimal Wearable Biosignal for Assessing Driving Stress. IEEE Access 2020, 8, 34362–34374. [Google Scholar] [CrossRef]
  94. Dong, Y.; Xu, L.; Zheng, J.; Wu, D.; Li, H.; Shao, Y.; Shi, G.; Fu, W. A Hybrid EEG-Based Stress State Classification Model Using Multi-Domain Transfer Entropy and PCANet. Brain Sci. 2024, 14, 595. [Google Scholar] [CrossRef]
  95. Ding, Y.; Liu, J.; Zhang, X.; Yang, Z. Dynamic Tracking of State Anxiety via Multi-Modal Data and Machine Learning. Front. Psychiatry 2022, 13, 757961. [Google Scholar] [CrossRef]
  96. Díaz, H.M.; Cid, F.M.; Otárola, J.; Rojas, R.; Alarcón, O.; Cañete, L. EEG Beta band frequency domain evaluation for assessing stress and anxiety in resting, eyes closed, basal conditions. Procedia Comput. Sci. 2019, 162, 974–981. [Google Scholar] [CrossRef]
  97. Dai, R.; Lu, C.; Yun, L.; Lenze, E.; Avidan, M.; Kannampallil, T. Comparing stress prediction models using smartwatch physiological signals and participant self-reports. Comput. Methods Programs Biomed. 2021, 208, 106207. [Google Scholar] [CrossRef]
  98. Dahal, K.; Bogue-Jimenez, B.; Doblas, A. Global Stress Detection Framework Combining a Reduced Set of HRV Features and Random Forest Model. Sensors 2023, 23, 5220. [Google Scholar] [CrossRef]
  99. Cho, D.; Ham, J.; Oh, J.; Park, J.; Kim, S.; Lee, N.-K.; Lee, B. Detection of stress levels from biosignals measured in virtual reality environments using a kernel-based extreme learning machine. Sensors 2017, 17, 2435. [Google Scholar] [CrossRef] [PubMed]
  100. Chorlian, D.B.; Porjesz, B.; Cohen, H.L. Measuring Electrical Activity of the Brain ERP Mapping in Alcohol Research. Alcohol Health Res. World 1995, 19, 315–320. [Google Scholar] [PubMed]
  101. Can, Y.S.; Chalabianloo, N.; Ekiz, D.; Ersoy, C. Continuous Stress Detection Using Wearable Sensors in Real Life: Algorithmic Programming Contest Case Study. Sensors 2019, 19, 1849. [Google Scholar] [CrossRef] [PubMed]
  102. Budzynski, T.H.; Budzynski, H.K.; Evans, J.R.; Abarbanel, A. Introduction to Quantitative EEG and Neurofeedback: Advanced Theory and Applications, 2nd ed.; Elsevier: Amsterdam, The Netherlands, 2009. [Google Scholar] [CrossRef]
  103. Brezinova, V.; Kendell, R.E. Smooth pursuit eye movements of schizophrenics and normal people under stress. Br. J. Psychiatry 1977, 130, 59–63. [Google Scholar] [CrossRef]
  104. Bhavani, T.; Vamseekrishna, P.; Chakraborty, C.; Dwivedi, P. Stress Classification and Vital Signs Forecasting for IoT-Health Monitoring. IEEE/ACM Trans. Comput. Biol. Bioinform. 2024, 21, 652–659. [Google Scholar] [CrossRef]
  105. Bevilacqua, F.; Engström, H.; Backlund, P. Game-Calibrated and User-Tailored Remote Detection of Stress and Boredom in Games. Sensors 2019, 19, 2877. [Google Scholar] [CrossRef]
  106. Battiti, R. Using Mutual Information for Selecting Features in Supervised Neural Net Learning. IEEE Trans. Neural Netw. 1994, 5, 537–550. [Google Scholar] [CrossRef]
  107. Banoczi, W.R. How some drugs affect the electroencephalogram (EEG). Am. J. Electroneurodiagn. Technol. 2005, 45, 118–129. [Google Scholar] [CrossRef]
  108. Althubaiti, A. Information bias in health research: Definition, pitfalls, and adjustment methods. J. Multidiscip. Healthc. 2016, 9, 211–217. [Google Scholar] [CrossRef]
  109. Ko, K.; Jones, A.; Francis, D.; Robidoux, S.; McArthur, G. Physiological correlates of anxiety in childhood and adolescence: A systematic review and meta-analysis. Stress Health 2024, 40, e3388. [Google Scholar] [CrossRef]
  110. Awada, M.; Gerber, B.B.; Lucas, G.M.; Roll, S.C. Stress appraisal in the workplace and its associations with productivity and mood: Insights from a multimodal machine learning analysis. PLoS ONE 2024, 19, e0296468. [Google Scholar] [CrossRef]
  111. Attaran, N.; Puranik, A.; Brooks, J.; Mohsenin, T. Embedded Low-Power Processor for Personalized Stress Detection. IEEE Trans. Circuits Syst. II Express Briefs 2018, 65, 2032–2036. [Google Scholar] [CrossRef]
  112. Arpaia, P.; Moccaldi, N.; Prevete, R.; Sannino, I.; Tedesco, A. A Wearable EEG Instrument for Real-Time Frontal Asymmetry Monitoring in Worker Stress Analysis. IEEE Trans. Instrum. Meas. 2020, 69, 8335–8343. [Google Scholar] [CrossRef]
  113. AlShorman, O.; Masadeh, M.; Bin Heyat, B.; Akhtar, F.; Almahasneh, H.; Ashraf, G.M.; Alexiou, A. Frontal lobe real-time EEG analysis using machine learning techniques for mental stress detection. J. Integr. Neurosci. 2022, 21, 20. [Google Scholar] [CrossRef]
  114. Al-Shargie, F.; Al-Ameri, S.; Al-Hammadi, A.; Vladimirovna, S.D.; Tariq, U.; Al-Nashash, H. Detection of Astronaut’s Stress Levels During 240-Day Confinement using EEG Signals and Machine Learning. In Proceedings of the 2023 45th Annual International Conference of the IEEE Engineering in Medicine & Biology Society (EMBC), Sydney, NSW, Australia, 24–27 July 2023. [Google Scholar] [CrossRef]
  115. Al-shargie, F.; Tang, T.B.; Badruddin, N.; Kiguchi, M. Towards multilevel mental stress assessment using SVM with ECOC: An EEG approach. Med. Biol. Eng. Comput. 2018, 56, 125–136. [Google Scholar] [CrossRef]
  116. Al-Ezzi, A.; Al-Shargabi, A.A.; Al-Shargie, F.; Zahary, A.T. Complexity Analysis of EEG in Patients with Social Anxiety Disorder Using Fuzzy Entropy and Machine Learning Techniques. IEEE Access 2022, 10, 39926–39938. [Google Scholar] [CrossRef]
  117. Al-Ezzi, A.; Yahya, N.; Kamel, N.; Faye, I.; Alsaih, K.; Gunaseli, E. Severity Assessment of Social Anxiety Disorder Using Deep Learning Models on Brain Effective Connectivity. IEEE Access 2021, 9, 86899–86913. [Google Scholar] [CrossRef]
  118. Akella, A.; Singh, A.K.; Leong, D.; Lal, S.; Newton, P.; Clifton-Bligh, R.; Mclachlan, C.S.; Gustin, S.M.; Maharaj, S.; Lees, T.; et al. Classifying Multi-Level Stress Responses from Brain Cortical EEG in Nurses and Non-Health Professionals Using Machine Learning Auto Encoder. IEEE J. Transl. Eng. Health Med. 2021, 9, 2200109. [Google Scholar] [CrossRef]
Figure 1. PRISMA flow diagram for study selection.
Figure 1. PRISMA flow diagram for study selection.
Applsci 15 11777 g001
Figure 2. International 10–20 system brain electrode distribution map. Source: author’s own work.
Figure 2. International 10–20 system brain electrode distribution map. Source: author’s own work.
Applsci 15 11777 g002
Figure 3. Heart rate variability (HRV) analysis via PPG and RR Intervals. (A). Data obtain. (B). Analysis of RR interval. (C). Calculation and Analysis oh heart rate variability Author’s own figure.
Figure 3. Heart rate variability (HRV) analysis via PPG and RR Intervals. (A). Data obtain. (B). Analysis of RR interval. (C). Calculation and Analysis oh heart rate variability Author’s own figure.
Applsci 15 11777 g003
Figure 4. Examples of wearable Electrodermal Activity (EDA) devices include commercial products such as the Iamjoy® Smart Health Wristband and Paetae® Smart Ring. Source: author’s own work.
Figure 4. Examples of wearable Electrodermal Activity (EDA) devices include commercial products such as the Iamjoy® Smart Health Wristband and Paetae® Smart Ring. Source: author’s own work.
Applsci 15 11777 g004
Figure 5. The main differences between Machine Learning and Deep Learning. Source: author’s own work.
Figure 5. The main differences between Machine Learning and Deep Learning. Source: author’s own work.
Applsci 15 11777 g005
Figure 6. Fundamental architectures of neural networks: CNN (A), RNN (B), and LSTM (C). Source: author’s own work.
Figure 6. Fundamental architectures of neural networks: CNN (A), RNN (B), and LSTM (C). Source: author’s own work.
Applsci 15 11777 g006
Table 1. HRV indicators in the time domain.
Table 1. HRV indicators in the time domain.
IndicatorsUnitsFormula
mRR 1ms i = 1 N R R i N
mHR 2Bpm 1 N i = 1 N 60,000 R R i
SDRR 3ms i = 1 N ( R R i m R R ) 2 N 1
SDHR 4Bpm i = 1 N ( 600,000 R R i m H R ) 2 N 1
CVRR 5- S D R R × 100 m R R
RMSSD 6ms m e a n ( ( R R i + 1 R R i ) 2 )
pRR20 7% c o u n t ( R R i + 1 R R i > 20 ) N 1 × 100
pRR50 8% c o u n t ( R R i + 1 R R i > 50 ) N 1 × 100
1 mRR: Mean of RR intervals. 2 mHR: Mean heart rate. 3 SDRR: Standard deviation of RR intervals. 4 SDHR: Standard deviation of heart rate. 5 CVRR: Coefficient of variation of RR intervals. 6 RMSSD: Root mean square of successive differences between RR intervals. 7 pRR20: Percentage of pairs of successive RR intervals differing by more than 20 ms. 8 pRR50: Percentage of pairs of successive RR intervals differing by more than 50 ms.
Table 2. Recent studies on stress detection using ML.
Table 2. Recent studies on stress detection using ML.
ResearchSignalMethodAccuracy%Dataset Details
[74] (2012)ECGKNN94.5810 healthy females, ECG at 1 kHz.
[76] (2018)ECG, EDA, EEGSVM8615 participants (12 usable data), mean age 40.8. 250 Hz, 51.2 Hz, 512 Hz.
[68] (2020)HRV, EDAMLP82.596 females, 9 males, avg. age 28. 8-day data collection.
[67] (2022)ECG, EDAKNN94.418 participants, 10 min of data (6 stress, 4 relaxation). 800 Hz, 4 Hz.
[45] (2023)EDA, PPGRandom Forest76.5Initial 3 min baseline rest, five tasks (10 min, 5 min, 3 min, unlimited, 1 min), 2 min rest post-task. Data collected using Empatica E4.
[45] (2023)EDA, PPGSVM74.5
[45] (2023)EDA, PPGLogistic Regression76.4
[42] (2024)
(stress vs. relax)
HRVRandom Forest89.238 participants, 5 min HRV (PPG). WESAD: 15 participants, 700 Hz. SWELL: 25 participants, 2048 Hz ECG.
[42] (2024)
(stress vs. relax)
HRVLogistic Regression87.2
[42] (2024)
(stress vs. neutral)
HRVRandom Forest61.2
[42] (2024)
(stress vs. neutral)
HRVDecision Trees59.5
[75] (2024)EEGLightGBM96.5119 participants (39 controls, 80 GAD patients), age 22–68, diagnosed per DSM-5. EEG: 10 min resting-state data, eyes closed, awake, relaxed. 16 electrodes, 250 Hz sampling.
[75] (2024)EEGXGboost95.7
[75] (2024)EEGCatboost97.2
[75] (2022)EEGRandom Forest80.1326 participants (12 males), mean age 20 ± 1.3 years. 40-channel EEG system, 1024 Hz sampling. 60 s resting state, 30 s test duration, 20 s between tests.
[75] (2022)EEGSVM78.64
[75] (2022)EEGXGboost86.49
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Liu, Y.; Palacio, M.-I.; Bikki, T.; Toledo, C.; Ouyang, Y.; Li, Z.; Wang, Z.; Toledo, F.; Zeng, H.; Herrero, M.-T. Machine Learning, Physiological Signals, and Emotional Stress/Anxiety: Pitfalls and Challenges. Appl. Sci. 2025, 15, 11777. https://doi.org/10.3390/app152111777

AMA Style

Liu Y, Palacio M-I, Bikki T, Toledo C, Ouyang Y, Li Z, Wang Z, Toledo F, Zeng H, Herrero M-T. Machine Learning, Physiological Signals, and Emotional Stress/Anxiety: Pitfalls and Challenges. Applied Sciences. 2025; 15(21):11777. https://doi.org/10.3390/app152111777

Chicago/Turabian Style

Liu, Yu, María-Itatí Palacio, Taha Bikki, Cesar Toledo, Yu Ouyang, Zhongzheng Li, Zhengyi Wang, Francisco Toledo, Hong Zeng, and María-Trinidad Herrero. 2025. "Machine Learning, Physiological Signals, and Emotional Stress/Anxiety: Pitfalls and Challenges" Applied Sciences 15, no. 21: 11777. https://doi.org/10.3390/app152111777

APA Style

Liu, Y., Palacio, M.-I., Bikki, T., Toledo, C., Ouyang, Y., Li, Z., Wang, Z., Toledo, F., Zeng, H., & Herrero, M.-T. (2025). Machine Learning, Physiological Signals, and Emotional Stress/Anxiety: Pitfalls and Challenges. Applied Sciences, 15(21), 11777. https://doi.org/10.3390/app152111777

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Article metric data becomes available approximately 24 hours after publication online.
Back to TopTop