You are currently viewing a new version of our website. To view the old version click .
Sensors
  • Communication
  • Open Access

29 October 2025

Non-Contact Fatigue Estimation in Healthy Individuals Using Azure Kinect: Contribution of Multiple Kinematic Features

and
1
National Institute of Technology, Tsuyama College, 624-1 Numa, Tsuyama 708-8509, Japan
2
Kyoto Seisakusho Co., Ltd., 377-1 Yodomizu-cho, Fushimi-ku, Kyoto 613-0916, Japan
*
Author to whom correspondence should be addressed.
This article belongs to the Special Issue Human Activity Recognition Based on Sensors: Challenges and Perspectives

Abstract

Monitoring exercise-induced fatigue is important for maintaining the effectiveness of training and preventing injury. We evaluated a non-contact approach that estimates perceived fatigue from full-body kinematics captured by an Azure Kinect depth camera. Ten healthy young adults repeatedly performed simple, reproducible whole-body movements, and 3D skeletal coordinates from 32 joints were recorded. After smoothing, 24 kinematic features (joint angles, angular velocities, and cycle timing) were extracted. Fatigue labels (Low, Medium, and High) were obtained using the Borg CR10 scale at 30-s intervals. A random forest classifier was trained and evaluated with leave-one-subject-out cross-validation, and class imbalance was addressed by comparing no correction, class weighting, and random oversampling within the training folds. The model discriminated fatigue levels with high performance (overall accuracy 86%; macro ROC AUC 0.98 (LOSO point estimate) under oversampling), and feature importance analysis indicated distributed contributions across feature categories. These results suggest that simple camera-based kinematic analysis can feasibly estimate perceived fatigue during basic movements. Future work will expand the cohort, diversify tasks, and integrate physiological signals to improve generalization and provide segment-level interpretability.

1. Introduction

Regular physical activity is widely recommended to extend healthy life expectancy and prevent lifestyle-related diseases [1]. However, training that is too light has limited benefits, whereas excessive intensity can lead to accumulated fatigue and a higher risk of injury [2]. Consequently, there is a growing need for objective, timely monitoring of exercise-induced fatigue to provide individualized feedback on appropriate training intensity.
Existing approaches often rely on wearable or physiological sensors such as heart-rate/heart-rate variability (HR/HRV) monitors, inertial measurement units (IMUs), and electromyography (EMG) [3,4,5]. Although conventional wearable approaches are effective, body attachment can interfere with natural movement. By contrast, recent microwearables (e.g., skin-interfaced microfluidics and ultrathin bioelectronics) are designed to be skin-conformal and low burden, with wireless communication, thereby substantially reducing motion interference [6,7]. Nevertheless, many platforms remain at the prototype stage, and their cost and availability are not yet competitive with commodity, image-based, non-contact technologies. In this study, we position a single-camera, skeleton-based pipeline as a camera-only reference baseline that complements microwearables and multimodal systems (see Section 5).
In parallel, advances in depth cameras and image-based body tracking (e.g., Microsoft Azure Kinect) have enabled non-contact acquisition of full-body kinematics for motion analysis and fatigue assessment [8,9,10]. Prior studies suggest that vision-based sensing can provide unobtrusive measures of fatigue; however, feasibility and reliability remain underexplored for simple, reproducible exercises performed by healthy individuals, especially in small, class-imbalanced cohorts.
Against this background, we investigated whether perceived fatigue can be estimated from non-contact full-body kinematics captured by an Azure Kinect depth camera and analyzed using a standard machine learning pipeline. To improve reproducibility and generalizability, we used a simple, familiar whole-body movement that induces fatigue quickly and is reliably tracked by Kinect. Perceived fatigue was labeled with the Borg CR10 and categorized into three levels (Low, Medium, High; abbreviated L/M/H in figures and tables). We chose CR10 for its practicality in short trials. We also acknowledge alternative criteria (e.g., physiological measures, talk test) and discuss their implications in Section 5.
Our contributions are threefold:
  • A feasibility evaluation of non-contact fatigue estimation in healthy young adults performing a reproducible movement;
  • The definition of 24 kinematic features from 32 joints;
  • An assessment of class imbalance handling strategies and their impact on performance.

3. Materials and Methods

3.1. Participants

Ten healthy young adults (eight males, two females) participated in this study. None reported any musculoskeletal or neurological disorders. The study was approved by the Ethics Committee of the National Institute of Technology, Tsuyama College (approval No. 2023-03). All participants were briefed on the study purpose, procedures, potential risks (transient fatigue), confidentiality, and data handling; participation was voluntary, and they could discontinue at any time without penalty. Written informed consent was obtained prior to data collection. Participant characteristics are summarized in Table 1 (mean ± SD; ranges); anthropometric measurements (age, height, mass, BMI) were recorded for all ten participants, and no imputation was required.
Table 1. Participant characteristics. Values are mean ± SD; n indicates the number of available records (ranges in parentheses).
Accordingly, we adopted a small pilot cohort (n = 10), consistent with guidance for feasibility studies in which ≈10–12 participants are commonly used to assess procedures and obtain preliminary variance/effect-size estimates for subsequent sample-size calculations, rather than to test hypotheses [14,15,16]. Accordingly, we adopted a small pilot cohort (n = 10), consistent with common practice in feasibility studies; we acknowledge the limited generalizability and will use these estimates to plan a larger, sex-balanced follow-up with an a priori sample-size calculation.

3.2. Experimental Protocol and Labeling

A standardized instruction was given: participants alternated between a neutral stance and an arms-abducted posture at a comfortable, self-selected pace for up to 3 min, without external pacing or feedback. Borg CR10 ratings were verbally reported every 30 s. An investigator was present throughout to ensure safety and to terminate the task on request or if any discomfort occurred. Figure 1 illustrates the repeated movement task (start and spread postures) and stance width (≈0.80 m). We adopted the Borg CR10 to reflect perceived fatigue because it is widely used in exercise settings, requires no additional hardware, and can be administered repeatedly during short trials without interrupting the task. We prioritized a device-free protocol that would not interrupt short in-task windows; therefore, a talk test was not included at this feasibility stage. The participants could discontinue earlier if desired. A standardized instruction was given: participants alternated between a neutral stance and an arms-abducted posture at a comfortable, self-selected pace for up to 3 min, without external pacing or feedback. Borg CR10 ratings were verbally reported every 30 s. An investigator was present throughout to ensure safety and to terminate the task on request or if any discomfort occurred.
Figure 1. Repeated whole-body movement used for data collection: participants alternated between a neutral stance and an arms-abducted posture at a self-selected pace; stance width was approximately 0.80 m.
For the analysis, each CR10 report was used to label the immediately preceding 30-s movement segment with three levels: Low (CR10 0–3), Medium (CR10 4–6), and High (CR10 ≥ 7). Labels were derived exclusively from CR10; no talk test or speech-based criteria were administered in this feasibility study, and no additional physiological measures were used for labeling. Alternative criteria include heart rate (HR) and heart rate variability (HRV), surface electromyography (EMG; e.g., median-frequency shifts), blood lactate, and oxygen-consumption-based measures, as well as the talk test. These were intentionally not used here to minimize burden in a camera-only feasibility study and to align labels with subjective experience. The advantages and limitations of CR10-only labeling and our plan to integrate objective checks (e.g., HR/HRV, surface EMG, talk test) are discussed in Section 5.

3.3. Data Acquisition (RGB-D Capture and Body Tracking)

Full-body kinematic data were captured as RGB-D (color + depth) frames using the Azure Kinect DK (Microsoft, Redmond, WA, USA). RGB-D frames were recorded with the Azure Kinect Sensor SDK and processed with the Azure Kinect Body Tracking SDK (v1.1.x) to estimate 3D positions of 32 anatomical joints using a deep learning model [17,18,19,20]. The camera was mounted 1.15 m above the floor and positioned 1.47 m in front of the participant (frontal view) to minimize occlusion (Figure 2a). Joint naming and hierarchy followed the Azure Kinect Body Tracking SDK (Figure 2b). To illustrate the tracking output, Figure 2c shows an example frame with the 32-joint skeleton overlaid on the depth image, and Supplementary Video S1 provides a short clip presented sequentially—RGB, then depth, then depth with skeleton overlay (≈5 s each).
Figure 2. RGB-D capture setup and body-tracking reference: (a) measurement setup—Azure Kinect DK positioned 1.47 m in front of the participant at a height of 1.15 m (frontal view); (b) joint hierarchy defined by the Azure Kinect Body Tracking SDK (32 joints); (c) example frame from the study task with the 32-joint skeleton overlaid on the depth image (circles = joints; lines = bones). Panel (b) reproduced from Microsoft Learn, “Body joints—Joint hierarchy,” © Microsoft, licensed under the Creative Commons Attribution 4.0 International (CC BY 4.0) license [21]. See Supplementary Video S1 for a real-time demonstration presented sequentially—RGB, then depth, then depth with the 32-joint skeleton overlay.
Joint coordinates returned by Azure Kinect Body Tracking SDK are expressed in millimeters in the depth-camera coordinate system; coordinates were converted to meters (divided by 1000) prior to computing angles and velocities. The time series were smoothed before differentiation to reduce high-frequency noise (Section 3.4). The timing of the Borg CR10 reports (every 30 s) and label assignments are described in Section 3.2.

3.4. Feature Extraction

We adopted a feature-level representation because fatigue manifests as subtle coordination and timing shifts across joints rather than the on/off presence of a coarse action label. In this feasibility stage, we deliberately chose a single, simple cyclic movement instead of multiple “normal actions” (e.g., jumping, running, squatting) to control task-related variability and safety burden and to ensure reproducibility under LOSO in a small cohort. Fast, high-impact actions introduce larger inter-subject variability and occasional tracking dropouts/occlusion in depth-based skeleton tracking; a controlled task allowed us to isolate fatigue-related changes.
Accordingly, we defined 24 kinematic features spanning three categories—joint angles, angular velocities, and cycle timing—as a sensitive, task-agnostic representation for fatigue estimation. Joint coordinates returned by the Azure Kinect Body Tracking SDK are expressed in millimeters in the depth-camera coordinate system; we converted them to meters (divided by 1000) before computing angles and velocities. The time series were smoothed before differentiation to reduce high-frequency noise. Table 2 lists all feature definitions.
Table 2. Kinematic features * used for classification (Features 1–24).
Features were computed on a per-cycle basis; unless stated otherwise, results aggregate three consecutive cycles by averaging to reduce variance. Windows were non-overlapping, and Borg CR10 labels recorded every 30 s were aligned to the nearest timestamp. Figure 3 summarizes the end-to-end processing pipeline—from RGB-D capture and body tracking to preprocessing, feature extraction, and classification/evaluation. Note that the 24 kinematic features constitute the model input representation; evaluation is based on classification metrics—accuracy and macro ROC AUC (one-vs-rest)—reported under LOSO.
Figure 3. Processing pipeline: RGB-D capture and body tracking → preprocessing (interpolation, smoothing) → 24 kinematic features (angles, angular velocities, cycle timing) → random forest classification under LOSO → evaluation (accuracy, macro ROC AUC; one-vs-rest).

3.5. Classification and Evaluation

We trained a random forest (RF) and evaluated it with leave-one-subject-out cross-validation (LOSO). We preferred RF over a single decision tree because, in this small-cohort LOSO setting, single-tree models exhibited fold-to-fold instability (high variance). RF reduces variance through bootstrap aggregation and random feature subsampling (here, max_features = “sqrt”), providing more stable generalization while retaining straightforward interpretability via feature importances. In each fold, one subject was held out for testing and the remaining subjects formed the training set. All preprocessing and any resampling were fit on the training data only, and the learned parameters were applied to the held-out subject to prevent information leakage. Unless stated otherwise, RF hyperparameters were n_estimators = 500, max_features = “sqrt”, max_depth = None, min_samples_leaf = 1, and random_state = 42.
To address class imbalance, we compared three strategies within each training fold only: (i) no correction; (ii) class weighting (class_weight = “balanced” in scikit-learn), where the weight for class c was set to w c = N K n c with N the number of training samples in the fold, K the number of classes (=3), and n c the count of class c in that fold; and (iii) uniform random oversampling with replacement by duplicating minority-class samples until an exactly balanced 1:1:1 ratio (Low/Medium/High) was reached. In each fold, the target per-class size equaled the majority-class count in that training fold. All resampling used random_state = 42 for reproducibility and was applied only to the training folds; the held-out subject’s data were never reweighted or resampled. No synthetic sample generation (e.g., SMOTE) was used.
For interpretability, we summarized impurity-based RF feature importances and examined their distribution after aggregating features by anatomical groups (e.g., trunk; hip–knee–ankle). Unless stated otherwise, importances were computed from the RF trained under the oversampling setting. Group contributions were computed by summing the importances of features associated with each segment within each fold and normalizing to 100%. We report accuracy and macro ROC AUC (one-vs-rest) as LOSO point estimates; confidence intervals are not provided because per-subject prediction logs were not preserved. Summary point estimates are provided in Table 3. Analyses were performed in Python (version 3.10) using standard, widely available machine-learning libraries (e.g., scikit-learn 1.3, NumPy 1.25, pandas 2.0). Exact package versions and hardware details were not recorded in this feasibility study. Scripts are available from the corresponding author upon reasonable request.
Table 3. Classification performance of the random forest under three imbalance-handling strategies. Metrics: accuracy and macro ROC AUC (one-vs-rest). Oversampling was applied within training folds only; the held-out subject’s data were never reweighted or resampled.

4. Results

4.1. Classification Performance

Under LOSO, Table 3 summarizes point estimates (accuracy and macro ROC AUC; one-vs-rest) across the three class imbalance settings. Without correction, overall accuracy was 0.61 with weak discrimination for the Medium class. Class weighting improved balance but remained below oversampling, which achieved the best performance (accuracy 0.86; macro ROC AUC 0.98). As per Methods (Section 3.5), we report LOSO point estimates only; confidence intervals are not provided. These results should be interpreted in light of the small, sex-imbalanced cohort, a single simple task, and CR10-only labeling (no talk test), as detailed in Section 5.
These results should be interpreted with several caveats. Performance depended on class imbalance handling—oversampling outperformed no correction or class weighting—and misclassifications occurred primarily between adjacent classes (Low vs. Medium and Medium vs. High), as seen in the confusion matrix. The values are LOSO point estimates without confidence intervals in a small (n = 10) cohort performing a single simple task from a single front-view camera; generalization to other tasks or viewpoints remains uncertain. In addition, depth-based skeleton tracking can produce short dropouts; although we smoothed/interpolated these (Section 3.4), such preprocessing may attenuate rapid, high-frequency changes.

4.2. ROC Analysis

The ROC curves for the three conditions are shown in Figure 4. The curve corresponding to oversampling clearly outperformed the other two, confirming the effectiveness of imbalance correction in improving classification reliability across Low, Medium, and High fatigue levels.
Figure 4. ROC curves of the random forest classifier under three imbalance correction strategies: original (no correction), class weighting, and oversampling.

4.3. Confusion Matrices

Confusion matrices are presented in Figure 5. Without correction, frequent misclassification occurred between Low and Medium fatigue levels. Class weighting reduced this tendency, while oversampling yielded a more balanced and accurate classification across all three classes.
Figure 5. Confusion matrices of the random forest classifier: (a) no correction, (b) class weighting, (c) oversampling. Cell values are counts; color intensity corresponds to the count (darker = larger).

4.4. Feature Importance

The relative importance of kinematic features is shown in Figure 6 (see Table 2 for definitions). Importances were computed from the random forest trained with oversampling, were normalized within each fold to sum to 1, and were then averaged across folds. No single feature dominated; joint angles, angular velocities, and cycle timing each contributed modestly. This pattern indicates that fatigue is better captured by distributed changes across feature types and segments than by any isolated marker. For interpretability, we also summarize importances at the body-segment level, which corroborates this distributed pattern.
Figure 6. Feature importance of the 24 kinematic features (see Table 2 for feature definitions) after oversampling, calculated using the random forest classifier.

5. Discussion

This study demonstrated the feasibility of estimating subjective fatigue in healthy young adults using a non-contact approach with Azure Kinect and machine learning. A random forest classifier, trained on 24 kinematic features extracted from a simple whole-body task, achieved robust classification performance after imbalance correction (accuracy 0.86; macro ROC AUC 0.98, LOSO point estimate). These findings support the potential of Kinect-based systems as unobtrusive alternatives to wearable or physiological sensors for fatigue monitoring.
As described in Section 3.5, we computed impurity-based random forest feature importances and qualitatively inspected their distribution. To relate model behavior to anatomy, we grouped features into coarse anatomical regions (e.g., trunk; hip–knee–ankle; upper limb). This inspection suggested that importances were distributed across regions rather than concentrated in a single segment. In future work, we will apply permutation importance and SHAP to windowed features to quantify segment-wise, time-resolved contributions while keeping the non-contact pipeline unchanged.
Previous research on fatigue monitoring has often relied on wearable devices, such as heart rate sensors, EMG, or inertial measurement units [2,3,4,5,22,23,24]. While effective, these methods require direct attachment to the body, which may interfere with natural movement. Recent studies have explored vision-based sensing for motion and fatigue analysis, reporting encouraging results in controlled settings [8,9,10,11,12,13]. The present study extends this literature by demonstrating that even a simple, reproducible exercise performed by healthy individuals can yield reliable fatigue estimates from kinematic data alone.
Compared with prior work that often targets athlete or patient cohorts, specialized tasks, and binary fatigue labels using wearable signals (e.g., HR/HRV, EMG, IMUs) [2,3,4,5,22,23,24], camera-only approaches have shown feasibility in controlled settings [8,9,10,11,12,13] but typically adopt simpler label schemes or more complex pipelines. Table 4 provides a compact, cross-modal summary (camera-based and physiological) including modality, participants, task, labels, and the top metric (as reported), positioning our method as a camera-only reference baseline. While camera-only pipelines reduce burden, wearable or multimodal systems can capture cardio-respiratory or neuromuscular signals (e.g., HR/HRV, EMG) that may complement kinematic cues; benchmarking and fusion will be pursued in follow-up studies. Our study differs by (i) using a single commodity depth camera with off-the-shelf skeleton tracking, (ii) deliberately selecting a simple, reproducible whole-body movement that induces fatigue quickly and is reliably tracked, (iii) labeling three levels of subjective fatigue (CR10) without additional instrumentation, and (iv) explicitly evaluating class-imbalance handling under LOSO. These choices position our results as a camera-only reference baseline for non-contact deployments, clarifying when a camera-only setup can provide actionable fatigue cues relative to wearable or multimodal systems. For example, Aguirre et al. (sit-to-stand with Kinect + HR) [25] and Zhang et al. (IMU-based walking) [26] reported strong performance under different settings, complementing our camera-only setup.
Table 4. Compact comparison of representative fatigue-estimation studies across modalities. Rows list studies; columns summarize modality, participants, task, labels, and the top metric (as reported) 1.
Several limitations warrant mention. First, the cohort was small (n = 10; 8 male/2 female), limiting statistical power and generalizability; we will expand to a larger, sex-balanced cohort and, in that validation, explicitly examine whether anthropometrics (e.g., BMI and body mass) modulate kinematic patterns by quantifying their associations with kinematic feature distributions and classification performance. Second, we evaluated a single simple exercise with a single front-view depth camera; performance may vary with other tasks, speeds, viewpoints, or occlusion. Future work will extend beyond a single movement by including functional tasks (e.g., sit-to-stand, squats, walking/running) and testing whether the present feature set improves generalization across tasks or whether higher-capacity temporal/graph models do so. Third, labels relied solely on CR10; a talk test or physiological criteria were not administered. In follow-up validation, we will integrate a brief, standardized talk test (e.g., a 20–30 s passage) aligned with the labeling windows and report concordance with CR10 (e.g., Cohen’s κ and ROC analysis against a CR10 ≥ 7 threshold), together with HR/HRV and surface EMG as objective checks. Fourth, we report LOSO point estimates only because per-subject prediction logs were not preserved; to quantify uncertainty, future work will retain foldwise predictions and compute bootstrap confidence intervals for LOSO. Fifth, although oversampling improved class balance, it can bias performance estimates; we will validate on larger, sex-balanced datasets and explore cost-sensitive learning. Finally, the model scope was limited to a random forest; benchmarking higher-capacity sequence models (e.g., LSTM/GRU/temporal CNN/Transformers) and skeleton-graph networks (e.g., ST-GCN/GCN) is deferred to larger datasets, and we will extend interpretability with TreeSHAP, permutation importance, and PDP/ALE, including segment-level and cycle-phase-specific attributions.
From a deployment standpoint, microwearables and our camera-based, non-contact pipeline serve complementary needs and can be used alone or in combination. Our earlier note about motion interference pertains to conventional wearables; smart microwearables are skin-conformal and minimize such interference while providing access to physiological signals. Conversely, image-based sensing offers low setup burden and preserves natural movement; however, many microwearable platforms remain at the prototype stage with outstanding issues of cost and availability [6,7]. Depending on context, the two modalities may be deployed separately or jointly; benchmarking and fusion will be pursued in follow-up studies.
Future work will expand to larger and more heterogeneous populations, including older adults and clinical groups, and will explore integration with multimodal signals (e.g., HR/HRV, surface EMG, and microwearables) to improve accuracy and robustness [23]. Such developments could support personalized training, workplace ergonomics, and rehabilitation monitoring, where unobtrusive fatigue assessment is highly valuable.
Beyond research settings, the same non-contact pipeline could support (i) rehabilitation/physical therapy by aligning tasks with therapist-defined movements and flagging fatigue-related coordination changes during simple home or clinic exercises, with clinical oversight and objective checks (e.g., HR/HRV, surface EMG), and (ii) pets, where wearable attachment is challenging, by substituting an animal pose estimator (species-specific skeletal keypoints) and using veterinary-grounded labels. Both scenarios will require viewpoint-appropriate camera placement, occlusion handling, and broader, domain-specific datasets; model retraining and validation will be needed to address domain shift (assistive devices, altered ranges of motion, species differences).

6. Conclusions

This study demonstrated that subjective fatigue can be estimated in healthy young adults using a non-contact approach that combines Azure Kinect with machine learning. A random forest classifier trained on 24 kinematic features extracted from a simple, reproducible whole-body task achieved high performance after imbalance correction (macro ROC AUC = 0.98, LOSO point estimate). Feature importance analysis revealed that contributions from multiple joint motions were more informative than any single marker, underscoring the value of distributed kinematic changes in fatigue estimation.
By focusing on a small, homogeneous sample and a single movement task, this study clearly established the feasibility of the proposed approach under controlled conditions. These deliberate choices highlight the method’s strength in yielding reliable results from minimal experimental requirements. Building on this foundation, future work will expand to larger and more diverse populations, incorporate multiple exercise tasks, and integrate multimodal signals to further enhance accuracy and robustness. Such developments could pave the way for widespread applications in personalized training, workplace ergonomics, and rehabilitation monitoring, where unobtrusive fatigue assessment is particularly valuable. Taken together, the proposed non-contact approach demonstrates feasibility and points toward clinical rehabilitation (with objective checks and clinical oversight) and pet monitoring (via animal pose estimators and veterinary-grounded labels), alongside multimodal fusion with microwearables and richer interpretability (e.g., class-wise attributions and cycle-phase-specific analyses) as key next steps.

Supplementary Materials

The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/s25216633/s1, Video S1: Real-time demonstration of Azure Kinect body tracking during the study task, presented sequentially—RGB, then depth, then depth with 32-joint skeleton overlay (H.264, 30 fps; faces blurred for privacy).

Author Contributions

Conceptualization, K.K. and T.Y.; Methodology, K.K.; Software, K.K.; Validation, K.K. and T.Y.; Formal Analysis, K.K.; Investigation, K.K.; Resources, T.Y.; Data Curation, K.K.; Writing—Original Draft, T.Y.; Writing—Review & Editing, T.Y.; Visualization, K.K.; Supervision, T.Y.; Project Administration, T.Y. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

The study was conducted in accordance with the Declaration of Helsinki and approved by the Ethics Committee of the National Institute of Technology, Tsuyama College (approval number: 2023-03; approval date: 4 December 2023).

Data Availability Statement

The data presented in this study are available on request from the corresponding author. The data are not publicly available due to privacy restrictions.

Acknowledgments

The authors would like to thank the members of the track and field club at the National Institute of Technology, Tsuyama College, for their voluntary participation in this study. We are also deeply grateful to Kanji Harada, and Kenji Takagi, for their valuable support and coordination.

Conflicts of Interest

Author Kai Kondo conducted this research while he was a student at National Institute of Technology, Tsuyama College, and is currently employed by Kyoto Seisakusho Co., Ltd. The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
AUCarea under the ROC curve
BMIbody mass index
CR10Borg Category-Ratio 10 scale
DKDeveloper Kit (Azure Kinect DK)
EMGelectromyography
LOSOleave-one-subject-out cross-validation
RFrandom forest
RGB-Dred–green–blue plus depth
ROCreceiver operating characteristic
SDKSoftware Development Kit

References

  1. Booth, F.W.; Roberts, C.K.; Laye, M.J. Lack of exercise is a major cause of chronic diseases. Compr. Physiol. 2012, 2, 1143–1211. [Google Scholar] [CrossRef] [PubMed]
  2. Shaffer, F.; Ginsberg, J.P. An overview of heart rate variability metrics and norms. Front. Public Health 2017, 5, 258. [Google Scholar] [CrossRef] [PubMed]
  3. Gualtieri, L.; Rauch, N.; Piciullo, A.; Melnyk, A.; Lamberti, F.; Prattichizzo, D. Wearable sensors for biomechanical analysis: A review of current applications and future directions in ergonomics. Sensors 2021, 21, 4728. [Google Scholar] [CrossRef]
  4. Chowdhury, R.H.; Reaz, M.B.I.; Ali, M.A.B.M.; Bakar, A.A.A.; Chellappan, K.; Chang, T.G. Surface electromyography signal processing and classification techniques. Sensors 2013, 13, 12431–12466. [Google Scholar] [CrossRef] [PubMed]
  5. Patel, S.; Park, H.; Bonato, P.; Chan, L.; Rodgers, M. A review of wearable sensors and systems with application in rehabilitation. J. Neuroeng. Rehabil. 2012, 9, 21. [Google Scholar] [CrossRef] [PubMed]
  6. Merces, L.; Ferro, L.M.M.; Thomas, A.; Karnaushenko, D.D.; Luo, Y.; Egunov, A.I.; Zhang, W.; Bandari, V.K.; Lee, Y.; McCaskill, J.S.; et al. Bio-Inspired Dynamically Morphing Microelectronics toward High-Density Energy Applications and Intelligent Biomedical Implants. Adv. Mater. 2024, 36, e2313327. [Google Scholar] [CrossRef]
  7. Bandodkar, A.J.; Choi, J.; Lee, S.P.; Jeang, W.J.; Agyare, P.; Gutruf, P.; Wang, S.; Sponenburg, R.A.; Reeder, J.T.; Schon, S.; et al. Soft, Skin-Interfaced Microfluidic Systems with Passive Galvanic Stopwatches for Precise Chronometric Sampling of Sweat. Adv. Mater. 2019, 31, 1902109. [Google Scholar] [CrossRef]
  8. Shotton, J.; Fitzgibbon, A.; Cook, M.; Sharp, T.; Finocchio, M.; Moore, R.; Kipman, A.; Blake, A. Real-time human pose recognition in parts from single depth images. In Proceedings of the 2011 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Colorado Springs, CO, USA, 20–25 June 2011; pp. 1297–1304. [Google Scholar] [CrossRef]
  9. Clark, R.A.; Pua, Y.H.; Fortin, K.; Ritchie, C.; Webster, K.E.; Denehy, L.; Bryant, A.L. Validity of the Microsoft Kinect for assessment of postural control. Gait Posture 2012, 36, 372–377. [Google Scholar] [CrossRef] [PubMed]
  10. Pfister, A.; West, A.M.; Bronner, S.; Noah, J.A. Comparative abilities of Microsoft Kinect and Vicon 3D motion capture for gait analysis. J. Med. Eng. Technol. 2014, 38, 274–280. [Google Scholar] [CrossRef]
  11. Albert, J.A.; Arnrich, B. A computer vision approach to continuously monitor fatigue during resistance training. Biomed. Signal Process. Control 2024, 89, 105701. [Google Scholar] [CrossRef]
  12. Zhao, X.; Zheng, Y.; Yu, Z.; Song, A. Construction of Kinect-based measuring and monitoring system for the degree of employee’s fatigue. J. Adv. Comput. Intell. Intell. Inform. 2014, 18, 56–63. [Google Scholar] [CrossRef]
  13. Wakimoto, K.; Hashimoto, Y.; Kamata, M.; Hori, Y. Fatigue estimation based on Laban movement analysis using Kinect. Syst. Integr. 2012, 5, 109–116. [Google Scholar]
  14. Julious, S.A. Sample size of 12 per group rule of thumb for a pilot study. Pharm. Stat. 2005, 4, 287–291. [Google Scholar] [CrossRef]
  15. Lancaster, G.A.; Dodd, S.; Williamson, P.R. Design and analysis of pilot studies: Recommendations for good practice. J. Eval. Clin. Pract. 2004, 10, 307–312. [Google Scholar] [CrossRef]
  16. Hertzog, M.A. Considerations in determining sample size for pilot studies. Res. Nurs. Health 2008, 31, 180–191. [Google Scholar] [CrossRef] [PubMed]
  17. Microsoft. Azure Kinect Body Tracking SDK v1.1.x—API Reference. Available online: https://microsoft.github.io/Azure-Kinect-Body-Tracking/release/1.1.x/index.html (accessed on 20 September 2025).
  18. Microsoft. Azure Kinect Body Tracking SDK v1.1.2—Download Page. Available online: https://www.microsoft.com/en-us/download/details.aspx?id=104221 (accessed on 20 September 2025).
  19. Microsoft. Azure Kinect DK—Product Page. Available online: https://azure.microsoft.com/en-us/products/kinect-dk (accessed on 20 September 2025).
  20. Microsoft. Azure Kinect Sensor SDK—API Reference. Available online: https://microsoft.github.io/Azure-Kinect-Sensor-SDK/master/index.html (accessed on 20 September 2025).
  21. Microsoft. Azure Kinect Body Tracking—Body Joints (Joint Hierarchy). Microsoft Learn. Available online: https://learn.microsoft.com/en-us/previous-versions/azure/kinect-dk/body-joints (accessed on 13 October 2025).
  22. Vøllestad, N.K. Measurement of human muscle fatigue. J. Neurosci. Methods 1997, 74, 219–227. [Google Scholar] [CrossRef]
  23. Gandevia, S.C. Spinal and supraspinal factors in human muscle fatigue. Physiol. Rev. 2001, 81, 1725–1789. [Google Scholar] [CrossRef] [PubMed]
  24. Merletti, R.; Parker, P.A. Electromyography: Physiology, Engineering, and Non-Invasive Applications; Wiley-IEEE Press: Hoboken, NJ, USA, 2004. [Google Scholar] [CrossRef]
  25. Aguirre, A.; Pinto Bernal, M.J.; Cifuentes, C.A.; Perdomo, O.; Díaz, C.A.R.; Múnera, M. Machine Learning Approach for Fatigue Estimation in Sit-to-Stand Exercise. Sensors 2021, 21, 5006. [Google Scholar] [CrossRef] [PubMed]
  26. Zhang, J.; Lockhart, T.E.; Soangra, R. Classifying Lower Extremity Muscle Fatigue during Walking Using Machine Learning and Inertial Sensors. Ann. Biomed. Eng. 2014, 42, 600–612. [Google Scholar] [CrossRef] [PubMed]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.