Next Article in Journal
Augmenting Orbital Debris Identification with Neo4j-Enabled Graph-Based Retrieval-Augmented Generation for Multimodal Large Language Models
Next Article in Special Issue
A Superpixel-Based Algorithm for Detecting Optical Density Changes in Choroidal Optical Coherence Tomography Images of Diabetic Patients
Previous Article in Journal
A Multi-Sensor Fusion Approach Combined with RandLA-Net for Large-Scale Point Cloud Segmentation in Power Grid Scenario
Previous Article in Special Issue
Assistance Device Based on SSVEP-BCI Online to Control a 6-DOF Robotic Arm
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Retrospective Frailty Assessment in Older Adults Using Inertial Measurement Unit-Based Deep Learning on Gait Spectrograms

by
Julius Griškevičius
1,
Kristina Daunoravičienė
1,*,
Liudvikas Petrauskas
1,
Andrius Apšega
2 and
Vidmantas Alekna
2
1
Department of Biomechanical Engineering, Vilnius Gediminas Technical University, LT-10105 Vilnius, Lithuania
2
Faculty of Medicine, Vilnius University, LT-03101 Vilnius, Lithuania
*
Author to whom correspondence should be addressed.
Sensors 2025, 25(11), 3351; https://doi.org/10.3390/s25113351
Submission received: 17 March 2025 / Revised: 21 May 2025 / Accepted: 23 May 2025 / Published: 26 May 2025

Abstract

Frailty is a common syndrome in the elderly, marked by an increased risk of negative health outcomes such as falls, disability and death. It is important to detect frailty early and accurately to apply timely interventions that can improve health results in older adults. Traditional evaluation methods often depend on subjective evaluations and clinical opinions, which might lack consistency. This research uses deep learning to classify frailty from spectrograms based on IMU data collected during gait analysis. The study retrospectively analyzed an existing IMU dataset. Gait data were categorized into Frail, PreFrail, and NoFrail groups based on clinical criteria. Six IMUs were placed on lower extremity segments to collect motion data during walking activities. The raw signals from accelerometers and gyroscopes were converted into time–frequency spectrograms. A convolutional neural network (CNN) trained solely on raw IMU-derived spectrograms achieved 71.4 % subject-wise accuracy in distinguishing frailty levels. Minimal preprocessing did not improve subject-wise performance, suggesting that the raw time–frequency representation retains the most salient gait cues. These findings suggest that wearable sensor technology combined with deep learning provides a robust, objective tool for frailty assessment, offering potential for clinical and remote health monitoring applications.

1. Introduction

Frailty refers to a clinical syndrome characterized by diminished physiological reserves that increase the risk of adverse health outcomes such as deteriorating mobility, disability, falls, hospital admissions, and death [1]. Although frailty can be reversed, especially in its initial phases [2,3,4], its detection often relies on subjective evaluations, such as Fried’s five criteria (weight loss, exhaustion, inactivity, slowness and weakness) [1]. These criteria may be prone to bias and lack feasibility for routine care, as self-reported measures can be inconsistent [5,6,7,8]. Objective methods for frailty early detection are therefore highly sought after [9,10].
A growing body of research suggests that gait parameters, such as stride time variability, step length and double support time, may offer early markers for frailty [11,12]. However, such studies typically require laboratory-based systems, such as camera systems, force platforms, or computerized walkways [13,14,15], which are often costly and impractical in daily clinical settings. Wearable devices like inertial measurement units (IMUs) provide a promising alternative for continuous gait monitoring in real-world environments, especially combined with recent technological advancements in deep learning algorithms for signal classification [16,17,18]. IMUs can capture acceleration and angular velocity at multiple body segments, potentially revealing subtle mobility alterations prior to clinical manifestations of frailty [19,20,21,22]. Unlike traditional optical motion capture systems that require multiple cameras, studies use one or more IMUs on different body parts to capture gait dynamics. Data from a single IMU are effective for activity recognition via mobile phones or smartwatches [23,24,25,26]. In gait analysis, even a single IMU on each foot can monitor gait asymmetry and other parameters effectively [27,28,29,30]. Using multiple sensors improves frailty evaluation by providing quantitative data to differentiate levels of frailty [31,32].
Although IMUs offer numerous advantages, the analyzing IMU data can be complex, involving robust algorithms for gait events detection and intricate signal processing methods like filtering and noise reduction, which might produce varying outcomes regarding the accuracy of feature extraction [32,33,34,35]. IMU-based human movement analysis can be performed in time and frequency domains. Combining both analyses can enhance understanding of gait characteristics, particularly in pathological conditions like Parkinson’s disease, stroke, or frailty [35,36,37,38].
The integration of machine learning (ML) algorithms and convolution neural network (CNN) techniques by converting IMU time-domain data into time–frequency representations (e.g., spectrograms, Gramian Angular Field GAF, Recurrence Plot and Markov Transition Field images) has further enhanced the predictive accuracy for frailty identification, classifying Parkinson’s disease patients, healthy controls and gymnast actions [36,39,40,41,42,43]. Arshad et al. successfully combined short time Fourier transform (STFT) and GAF representations of IMU gait signals for frailty assessment [44]. These image-based approaches leverage machine vision techniques as an alternative to optical motion capture systems, enabling markerless gait classification through CNNs [36]. Beyond CNNs, other time series classification algorithms (e.g., InceptionTime) have also proven effective for frailty classification, achieving an 81% accuracy on test data [45,46]. Given that frailty datasets are frequently small and heterogenous, there remains a need for optimized strategies to harness IMU data—particularly minimally processed signals—without extensive feature engineering [47,48,49,50,51,52].
This study aims to evaluate the feasibility and effectiveness of using raw or minimally processed IMU-derived spectrograms for frailty classification via deep learning. Unlike prior investigations that rely heavily on custom-engineered features or highly processed signals, we focus on determining whether raw IMU data retains sufficient discriminatory power to differentiate among Frail, PreFrail and NoFrail older adults. Additionally, we compare multiple signal preprocessing approaches (e.g., baseline gravitational offset removal, low-pass filtering) to assess their impact on classification performance. By reducing the need for complex signal processing and leveraging spectrogram-based representations, we seek to streamline frailty assessment for potential clinical and real-world applications, thereby filling the gap in the existing literature where many proposed methods remain resource-intensive or inconsistent across different populations.

2. Materials and Methods

This study was conducted retrospectively using an existing IMU dataset originally collected from the study by Apšega et al., which investigated the association between frailty levels and gait parameters derived from wearable sensors [32]. Although the original publication focused on correlational analysis and descriptive statistics, the present work extends those findings by performing a deep learning–based classification of frailty status.
Frailty Gait Database (FRGaitDB) comprises clinical screening data and IMU-based recordings from 133 participants with an average age ± SD of 75.1 ± 8 and a body mass index of 27.6 ± 5.8. The frailty level was categorized following Fried et al. [1]: NoFrail = no criteria; PreFrail = one or two criteria; and Frail = three or more criteria out of the five components, consisting of weight loss, weakness, exhaustion, slowness and low physical activity. Based on this evaluation, participants were assigned to Frailty (n = 35), PreFrail (n = 67) and NoFrail (n = 31) groups. Six wireless inertial measurement units (IMUs, Shimmer Research, Dublin, Ireland) were secured to the thighs, shins and feet using straps. Subjects were instructed to walk 4 m at their normal speed, and data from three separate trials were used for further processing and analysis. The sensors’ data were collected through a Bluetooth wireless connection at a sampling rate of 256 Hz.
A multimodal deep learning model was developed to classify participants into three frailty categories: Frail, PreFrail, and NoFrail. The data processing and classification model architecture consisted of two parallel branches—image-based and clinical-based— and it is shown in Figure 1.
Data processing steps are described in subsequent sections.

2.1. IMU Data Acquisition and Preprocessing

The dataset comprised IMU sensor recordings from participants during gait trials. Each participant had inertial measurement units attached to both lower limbs (thigh, shank and foot). The IMU sensors recorded tri-axial accelerometer and gyroscope data across three walking trials per participant.

2.1.1. IMU Data Preprocessing and Spectrogram Generation

The data were processed into three distinct datasets to evaluate the impact of preprocessing on classification performance:
  • Raw signals (no processing). Unaltered tri-axial accelerometer and gyroscope signals were used to evaluate the baseline capability of the model to learn directly from unfiltered sensor data. Examples of raw accelerometer and gyroscope signal are shown in Figure 2.
  • Preprocessed signals (baseline processing). In the baseline processing, accelerometer data were detrended to remove gravitational offset [50]. For each gait trial, a baseline gravitational component was estimated as the mean acceleration over samples 5 to 30, corresponding to an initial standing phase, which was then subtracted from the entire accelerometer signal along each axis, effectively isolating dynamic acceleration related to movement (Figure 3). Gyroscope data were retained in their original form.
Figure 3. Baseline processed accelerometer signals of IMU fixed on thigh, shank and foot segments.
Figure 3. Baseline processed accelerometer signals of IMU fixed on thigh, shank and foot segments.
Sensors 25 03351 g003
  • Filtered signals (low pass filtering). The gravity-corrected accelerometer signals (from the baseline processed dataset) and raw gyroscope signals were further processed using a 2nd-order low-pass Butterworth filter with a cut-off frequency of 8 Hz, applied separately to each axis (Figure 4). The selection of the 8 Hz cutoff frequency was informed by the existing literature and gait frequency analysis. Typical human walking involves frequency components predominantly below 10 Hz, with higher frequencies generally attributed to noise or extraneous motion artifacts [52,53].
Figure 4. Low pass filtered gyroscope and accelerometer signals of IMU fixed on thigh segment.
Figure 4. Low pass filtered gyroscope and accelerometer signals of IMU fixed on thigh segment.
Sensors 25 03351 g004
Accelerometer and gyroscope signals collected from IMUs were transformed into time–frequency spectrograms to serve as input for deep learning models. The Short-Time Fourier Transform (STFT) was applied to the raw, preprocessed and filtered IMU signals using a sampling frequency of 256 Hz and a Hanning window of size 256 samples with 50% overlap, generating a time–frequency representation of the signal’s power spectral density. To improve the visibility of spectral features, the power spectral density was log-transformed. Each spectrogram was saved as a color image (RGB) with axes and labels removed to ensure a clean input for the convolutional neural network (CNN). Figure 5 shows example spectrograms generated from accelerometer and gyroscope signals from the foot segment. Despite the short walking distances (~4 m), we included the entire recorded sequence, including acceleration and deceleration phases, to mirror a realistic clinical setup and maintain trial consistency. Clinical studies indicate that relevant movement features can emerge during transitional gait phases.

2.1.2. Data Partitioning and Leakage Control

All experiments used a subject-wise split: the three walking trials recorded for a given participant were always assigned to the same training, validation or test partition to avoid data leakage. The resulting fold sizes were 106/13/14 subjects (training/validation/test).

2.2. Clinical Data

Clinical data contains the evaluation of participants based on frailty assessment methodology and clinical features that were imported from an Excel spreadsheet and then processed by normalization. Clinical features are organized into grip strength, physical activity, timed tests, balance tests, gait assessment, demographics and gait parameters (Table 1). Descriptive statistics are reported as mean ± standard deviation, and differences among the three frailty categories for each clinical or gait-related feature were evaluated with an omnibus test. When the omnibus test reached significance (threshold p < 0.05), post hoc pair-wise comparisons were carried out (Bonferroni-adjusted independent-samples t-tests after ANOVA). Principal Component Analysis (PCA) was conducted within each feature group to reduce dimensionality while maintaining 95% of the variance.

2.3. Subject-Based Data Splitting

To ensure realistic generalization and prevent data leakage between training and testing, a subject-wise data split was applied. In total, 80% of participants were divided into training, 10% to validation and 10% to testing subsets. There were 108 spectrograms per subject (3 axes of accelerometer and 3 axes of gyroscope per sensor, 6 sensors on 6 lower leg segments and 3 trials), totaling 14364 samples per dataset. Each spectrogram file name encoded the participant ID, sensor location (thigh, shank, foot), side (left, right), and signal channel. These IDs were used to group all spectrograms belonging to the same participant. Stratified random sampling based on frailty categories (Frail, PreFrail and NoFrail) was implemented to ensure proportional allocation by frailty class, thereby preserving the original class distribution within each subset. Participants were independently assigned to training, validation and test subsets, ensuring that all spectrograms from a specific participant were grouped within the same subset.

2.4. Deep Learning Model Architecture

The image branch (Figure 1) consisted of a pre-trained ResNet-50 convolutional neural network (CNN), which was pre-trained on ImageNet, comprising 50 convolutional layers organized into five main stages. ResNet-50 was chosen for its demonstrated effectiveness in capturing intricate spatial hierarchies, including those in non-natural image domains such as spectrograms [54,55]. Its deep residual architecture facilitates the efficient learning of time–frequency patterns that indicate gait abnormalities and frailty-specific spectral signatures. The initial convolutional layers employed a 7 × 7 kernel size, followed by multiple 3 × 3 kernels in subsequent residual blocks. Early convolutional layers were frozen, while layers 3 and 4 were fine-tuned during training to adapt to spectrogram data. The final fully connected layer of ResNet-50 was replaced with a customized classification head that features batch normalization, dropout regularization (with rates of 0.2 and 0.3), and ReLU activation functions to reduce overfitting on relatively small datasets and enhance model generalizability. The clinical branch consisted of a two-layer feedforward neural network that processed the reduced clinical feature vector obtained after principal component analysis (PCA). This network consisted of an input layer (with the same number of neurons as the PCA-reduced feature set size), followed by hidden layers with 64 and 32 neurons, each activated by ReLU and regularized with dropout (with a rate of 0.3). Outputs from both branches were concatenated and passed through a fully connected fusion layer containing 256 neurons, a ReLU activation function and dropout regularization (rate of 0.5). The spectrogram and clinical features were concatenated and passed through a fully connected fusion layer, followed by a SoftMax output layer for final classification into the three frailty categories.

2.5. Training Process

Model training and evaluation were conducted using GPU acceleration with CUDA. Due to imbalanced class sizes, class weights were assigned inversely proportional to the frequency of each class, i.e., 1.2 for Frail, 0.7 for PreFrail and 1.3 for NoFrail. During the training, the cross-entropy loss function was used as the objective, and the Adam optimizer was utilized with a learning rate of 0.0005. Early stopping with a patience of 5 epochs was applied to avoid overfitting, monitoring the validation accuracy. A batch size of 32 was employed. Each of the three spectrogram datasets—raw, baseline-processed and low-pass filtered—was trained independently to assess the effect of signal conditioning on model performance. This resulted in three separate trained models, each optimized under identical hyperparameters and evaluation protocols. A 5-fold Stratified Cross-Validation was used to ensure balanced train, validation and test splits, maintaining proportional class representation.

2.6. Evaluation Metrics

Assessment of classification accuracy is performed at both the image and subject scales. Model performance was assessed at two levels:
  • Spectrogram-Level Evaluation. Classification accuracy, confusion matrix, sensitivity, specificity and balanced accuracy were computed at the individual spectrogram level.
  • Subject-Level Evaluation. Each participant was assigned a final predicted label using a majority vote across all their spectrograms. Subject-level confusion matrices and accuracy were also computed to reflect better real-world performance, where clinical decisions would be based on whole-subject classification.
To determine whether the observed differences in classification accuracy and other metrics among the raw, baseline and low-pass filtered methods were statistically significant, a Friedman test was performed on the results obtained from 5-fold cross-validation. Each fold’s performance metric (accuracy, F1-score, etc.) was treated as a repeated measure. Where the Friedman test indicated significance (p < 0.05), a post hoc test (e.g., Nemenyi test) was applied to identify which methods significantly differed.
Separate training on clinical data was performed to gauge the benefit of the clinical features only, and accuracy and macro-F1 were averaged over the three cross-validation folds (same splits as in the main experiment).
Additionally, to assess the possible effects of data leakage, an analysis was conducted by comparing three evaluation scenarios: (a) subject-wise classification (where all spectrograms from a subject are included in either the training, validation or test sets); (b) random split classification (where spectrograms from the same subject could appear in several dataset partitions) and (c) cross-evaluation classification (validating across different subjects).

3. Results

This study evaluated the classification performance of raw, processed, and low-pass filtered IMU signals combined with clinical features, as well as an alternative approach based solely on spectrogram representations of IMU-derived gait data. The primary objective of this study was to evaluate whether raw or minimally processed IMU signals, represented as spectrograms, could effectively differentiate between Frail, PreFrail, and NoFrail individuals. The classification performance was assessed using precision, recall, F1-score and overall accuracy, comparing the ability of different data processing techniques to distinguish between Frail, PreFrail and NoFrail individuals.
Table 2 presents the classification performance across these datasets using a multimodal classification model.
The findings reveal that a combination of raw IMU data and clinical features reached an average accuracy of 85.19%, highlighting that even minimally processed signals hold valuable information for classifying frailty. The precision and recall metrics for the Frail and PreFrail categories were notably high (0.945 and 0.881 for precision, 0.925 and 0.811 for recall, respectively), implying that unprocessed sensor data alone forms a solid foundation for classification. Nonetheless, the NoFrail category demonstrated lower precision (0.703) and recall (0.825), pointing to a higher chance of incorrect classification for NoFrail individuals. The macro-averaged F1-score for this method was 0.842, indicating a balanced performance across all categories. Implementing preprocessing strategies on IMU data, specifically stripping the gravitational element and normalizing signals, led to enhanced classification efficacy. The baseline processed IMU dataset saw a rise in accuracy to 88.15%, with significant gains in identifying Frail individuals, where the F1-score went up to 0.995. The PreFrail category experienced a moderate rise in F1-score (0.879), while NoFrail classification remained unchanged (F1-score = 0.754). These results indicate that preprocessing signals boost deep learning models’ capability to differentiate between frailty categories, especially for the most critical cases. Further enhancement via low-pass filtering of IMU signals kept the overall classification accuracy at 88.15%, with slight variations in the performance of individual categories. The F1-score for NoFrail classification slightly improved to 0.773, showing that filtering reduced noise and improved the model’s accuracy in identifying NoFrail individuals. However, a slight drop in Frail classification (F1-score = 0.979) suggests that removing high-frequency elements could have removed essential gait-related data for the most severe frailty cases. Regardless, the macro-averaged F1-score stayed consistent at 0.877, affirming the overall stability of both processing techniques.
The statistical significance tests using the Friedman test were performed on macro-level performance (averaged across all classes) and per-class metrics for Frail, PreFrail, and NoFrail. Recall in the Frail class had the largest Friedman statistic (3.50) but did not reach significance (p = 0.1738). Precision and F1 also showed no significant differences (p = 0.9460 and p = 0.8187, respectively). No metric demonstrated significance in the PreFrail class (all p-values > 0.05). Similarly, Friedman tests showed non-significant differences for precision, recall and F1-score (highest statistic = 1.60) in the NoFrail class.
To further evaluate the feasibility of spectrogram-based frailty classification, models were trained solely on time–frequency representations of IMU signals without explicit clinical features. The dataset was structured as follows: a subject-level dataset containing n = 106 for training, n = 13 for validation, and n = 14 for testing subjects; a spectrogram-level dataset containing Training n = 11448, Validation n = 1404, and Test n = 1512 spectrogram images.
Using a strict subject-wise split, the Raw spectrogram model reached 71.43% accuracy, substantially higher than the Processed (64.29%) and Filtered spectrogram models (64.29%) (Figure 6). These results suggest that aggressive filtering may remove subtle high-frequency gait components informative for frailty staging.
After analyzing the top-performing spectrogram models through confusion matrix analysis, the following test accuracies were observed: Raw spectrograms achieved 93.11%, Processed spectrograms reached 97.22%, and Low-pass filtered spectrograms attained 92.14% (Figure 7).
The numerical performance of the subject-level and spectrogram-level splits of spectrogram-only models is summarized in Table 3.
These findings demonstrate that deep learning models trained solely on spectrogram data can achieve high accuracy rates. Notably, the processed spectrograms provided the highest classification performance with 97.22% accuracy. This indicates that the time–frequency representations of IMU signals are effective for classifying frailty, even without the inclusion of specifically designed clinical features.
The clinical feature block achieved an accuracy of 65.38%, lower than the 71.4% accuracy obtained using raw spectrograms. Combining these two inputs increased the accuracy to 72.75% (Δ = 1.35%).
Data leakage analysis shows that random splitting artificially inflates performance by 15.8% (Figure 8).
Results confirm a significant overestimation of classification performance when random splits are used instead of subject-wise separation. In particular, the random split test accuracy for the raw spectrogram model was 86.08%, whereas its subject-wise accuracy was only 62.31%, revealing an overestimation factor of 1.38. This suggests that models trained on spectrograms may memorize subject-specific patterns rather than generalizable gait characteristics, leading to inflated performance in cross-validation scenarios that do not enforce subject separation. The findings indicate that random distribution of spectrograms across the training, validation, and test sets results in exaggerated classification performance, likely due to overfitting and feature redundancy when images from the same subject are included in multiple partitions.

4. Discussion

The findings of this research offer insights into the possibility of categorizing frailty status using deep learning applied to IMU-derived gait signals and spectrogram-based representations. The results indicate that both raw and processed IMU signals have an adequate ability to differentiate among Frail, PreFrail, and NoFrail individuals. However, signal preprocessing is essential for improving classification performance, especially for detecting Frail individuals. Comparing raw, baseline processed and low-pass filtered IMU signals revealed that preprocessing increased classification accuracy from 85.19% (raw) to 88.15% (processed and filtered). Spectrogram-based classification accuracy is similar to the time-series classification using the InceptionTime model (81%); however, in a study by Amjad et al., they did not perform classification on a subject-based level [46]. Arshad achieved similar classification performance by combining STFT and GAF representations of IMU gait signals of frailty subjects [43,44]. Although the processed and baseline filtered methods produced slightly higher average performance metrics than raw in some cases (e.g., accuracy, F1-score), the observed differences did not reach statistical significance.
The most significant enhancement was noted in the Frail category, where the F1-score rose from 0.933 (raw) to 0.995 (processed). This implies that preprocessing boosts the model’s capability to detect movement deficits linked to frailty, likely by decreasing sensor noise and eliminating irrelevant signal variations. Nonetheless, classification performance for NoFrail individuals was consistently lower across all datasets (F1-score: 0.752 for raw, 0.754 for processed, 0.773 for filtered), suggesting that distinguishing healthy older adults from PreFrail individuals is still difficult. This is consistent with previous studies that showed frailty transitions are gradual rather than binary, and gait changes may be more subtle in early frailty stages [1,2]. Looking at the clinical features alone, in our previous study we have determined that gait speed was the most sensitive parameter for the identification of frailty and stride time was sensitive for discriminating against PreFrail or Frail from NoFrail [32]. Moreover, increased double support time was expressed in PreFrail and Frail subjects. Further optimization, such as combining biomechanical parameters or incorporating additional motion-based descriptors, could enhance NoFrail classification. Notably, low-pass filtering did not significantly outperform baseline processing, indicating that eliminating high-frequency components does not necessarily improve classification performance. While filtering slightly enhanced NoFrail classification, it somewhat reduced Frail classification accuracy, possibly due to the loss of some high-frequency gait features. This underscores the significance of preserving signal integrity in the design of gait classification models. To determine whether performance differences among the raw, baseline processed, and low-pass filtered methods were statistically significant, Friedman tests on the classification metrics (accuracy, precision, recall and F1-score) were conducted. None of the methods showed statistically superior macro-level and per-class metrics of accuracy, precision, recall or F1-score. These findings suggest that the choice of signal preprocessing does not systematically alter classification performance when combined with clinical features, at least under our dataset and experimental conditions.
The assessment of spectrogram-based deep learning models, especially combining them with clinical features, looks promising as an alternative strategy for classifying frailty. The models relying solely on spectrograms demonstrated high accuracy in classification, with the top-performing model (processed spectrograms) attaining an accuracy of 97.22%; however, due to the still small subject number for machine learning applications, there is a risk of overfitting and poor generalization. Time–frequency representations adequately capture pertinent gait features necessary for frailty classification, establishing them as a feasible alternative to time-series methods. Nonetheless, a key issue in spectrogram-based classification is data leakage, where spectrograms from the same individual might be split across training, validation, and test sets. Our analysis showed that random spectrogram splitting artificially increased test accuracy, causing overclassification bias. For instance, the test accuracy for raw spectrogram classification rose from 62.31% (subject-wise split) to 86.08% (random split), resulting in an overestimation factor of 1.38. This underscores the need for subject-wise data separation to ensure models can generalize to new individuals instead of memorizing individual-specific patterns. Despite this challenge, subject-wise spectrogram classification still produced encouraging results, with raw spectrograms achieving 71.43% accuracy. Combining with clinical features only slightly increases accuracy to 72.75 %, indicating a small but non-decisive gain. This implies that IMU-derived spectrograms alone offer valuable insights for frailty assessment, even without clinical features. Future research may investigate other spectrogram transformations, such as wavelet decomposition, recurrence plots or deep feature embeddings, to enhance classification robustness.
The results of this research have significant implications for the screening of frailty and its practical application in clinical settings. The capability to identify frailty with minimally processed IMU data endorses the potential for automated, wearable-based assessments in both clinical and home environments. In contrast to conventional frailty evaluations that depend on clinical assessments or gait laboratories, our method shows that machine learning can derive meaningful features from raw sensor data, which may lessen the need for manual feature engineering. Furthermore, the high accuracy of spectrogram-based models indicates that deep learning can effectively interpret time–frequency patterns in IMU signals. This could lead to the incorporation of spectrogram-based frailty assessment into mobile health applications or remote patient monitoring systems. However, our results also highlight the importance of stringent evaluation protocols to maintain subject-wise separation, thereby preventing the overestimation of classification performance.
There are multiple limitations to consider. First, the dataset utilized in this research was gathered in controlled laboratory settings, which might not accurately reflect real-world gait evaluations where external factors, such as uneven terrain and obstacles, add further variability. Second, treating the three trials per participant as separate inputs enlarges the training set and exposes the network to stride-to-stride variability. At the same time, the subject-wise split guarantees that no information from an individual leaks into the validation or fold. The subject-wise partition and inverse-frequency class weighting in the loss function prevent information leakage but cannot substitute for a larger, more diverse cohort. Future studies should therefore enroll more participants to improve external validity. Third, we analyzed only the accelerometer and gyroscope channels. Future investigations could integrate additional sensor types, like force sensors or electromyography (EMG), to capture more detailed movement characteristics. Fourth, examining self-supervised learning or transformer-based models could improve classification accuracy by capturing extended gait dynamics. Finally, despite the high performance of frailty classification, the ability to clinically interpret deep learning models remains an issue. Upcoming work might prioritize explainable AI approaches, like saliency maps or attention mechanisms, to determine which spectrogram areas or IMU signal patterns are most influential in classification decisions. In future research, validating models on larger [46,56] and maybe more diverse cohorts would be beneficial to ensure reliability and applicability. In parallel, exploiting advanced time–frequency methods could also help to capture subtle gait adjustments and patterns, thus enhancing early-stage frailty detection. Finally, sensitivity analyses that rank the contribution of each sensor location could identify a minimal, patient-friendly IMU configuration and reveal segment-specific gait signatures of frailty.

5. Conclusions

This research demonstrates that preprocessing signals enhance the precision of frailty assessments, particularly in identifying Frail individuals. Although low-pass filtering slightly benefits NoFrail classification, it may filter out significant high-frequency gait details. Raw IMU-derived spectrograms enabled a subject-wise frailty classification accuracy of 71.4%, confirming that minimally processed wearable signals contain discriminative gait patterns. Additional pre-processing did not improve and occasionally reduced performance. While augmenting the model with clinical features can raise the headline accuracy, that benefit vanishes when strict subject separation is applied, underscoring the importance of leakage-free evaluation protocols. Taken together, our results show that a fully passive, wearable-only pipeline is feasible for frailty screening, and that careful control of signal preprocessing is critical for reliable deployment. These findings highlight the promise of IMU-based frailty evaluations in clinical settings.

Author Contributions

Conceptualization, J.G.; methodology, J.G., K.D., A.A. and V.A.; software, J.G.; formal analysis, J.G. and A.A.; investigation, J.G. and A.A.; resources, J.G.; data curation, J.G., A.A. and L.P.; writing—original draft preparation, J.G. and K.D.; writing—review and editing, J.G. and K.D.; visualization, J.G.; supervision, J.G. and V.A. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

The study was conducted in accordance with the Declaration of Helsinki and approved by the Vilnius Regional Ethics Board for Biomedical Research.

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

Dataset available on request from the authors.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
IMUInertial Measurement Unit
CNNConvolutional Neural Network
STFTShort-Time Fourier Transform

References

  1. Fried, L.P.; Tangen, C.M.; Walston, J.; Newman, A.B.; Hirsch, C.; Gottdiener, J.; Seeman, T.; Tracy, R.; Kop, W.J.; Burke, G.; et al. Frailty in Older Adults: Evidence for a Phenotype. J. Gerontol. Ser. A Biol. Sci. Med. Sci. 2001, 56, M146–M157. [Google Scholar] [CrossRef] [PubMed]
  2. Rockwood, K.; Mitnitski, A. Frailty in Relation to the Accumulation of Deficits. J. Gerontol. Ser. A Biol. Sci. Med. Sci. 2007, 62, 722–727. [Google Scholar] [CrossRef] [PubMed]
  3. Gill, T.M.; Gahbauer, E.A.; Allore, H.G.; Han, L. Transitions between Frailty States among Community-Living Older Persons. Arch. Intern. Med. 2006, 166, 418–423. [Google Scholar] [CrossRef] [PubMed]
  4. Lang, P.O.; Michel, J.P.; Zekry, D. Frailty Syndrome: A Transitional State in a Dynamic Process. Gerontology 2009, 55, 539–549. [Google Scholar] [CrossRef]
  5. Kiely, D.K.; Cupples, L.A.; Lipsitz, L.A. Validation and Comparison of Two Frailty Indexes: The MOBILIZE Boston Study. J. Am. Geriatr. Soc. 2009, 457, 1532–1539. [Google Scholar] [CrossRef]
  6. da Câmara, S.M.A.; Alvarado, B.E.; Guralnik, J.M.; Guerra, R.O.; Maciel, Á.C.C. Using the Short Physical Performance Battery to Screen for Frailty in Young-Old Adults with Distinct Socioeconomic Conditions. Geriatr. Gerontol. Int. 2013, 13, 421–428. [Google Scholar] [CrossRef]
  7. Melzer, D.; Lan, T.Y.; Tom, B.D.M.; Deeg, D.J.H.; Guralnik, J.M. Variation in Thresholds for Reporting Mobility Disability between National Population Subgroups and Studies. J. Gerontol. Ser. A Biol. Sci. Med. Sci. 2004, 59, 1295–1303. [Google Scholar] [CrossRef]
  8. Tudor-Locke, C.E.; Myers, A.M. Challenges and Opportunities for Measuring Physical Activity in Sedentary Adults. Sports Med. 2001. [Google Scholar] [CrossRef]
  9. McDonagh, J.; Prichard, R.; Ferguson, C.; Phillips, J.L.; Davidson, P.M.; Macdonald, P.S.; Newton, P.J. Clinician Estimates of Frailty Compared to Formal Frailty Assessment in Adults with Heart Failure: A Cross-Sectional Analysis. Heart Lung Circ. 2022, 31, 1241–1246. [Google Scholar] [CrossRef]
  10. Sharma, H.; Devkaran, B.; Kapoor, K.; Odugoudar, A.; Koufopoulos, G.; Mehra, S. WTP2.2 Subjective Assessment of Frailty Compared to Validated Frailty Assessment in Adults with Chronic Renal Failure Presenting for Dialysis-Access/Transplant Workup Has Poor Estimation of Agreement. Br. J. Surg. 2023, 110 (Suppl. 6), znad241.158. [Google Scholar] [CrossRef]
  11. Amorim, A.B.; Pappas, E.; Simic, M.; Ferreira, M.L.; Tiedemann, A.; Jennings, M.; Ferreira, P.H. Integrating Mobile Health and Physical Activity to Reduce the Burden of Chronic Low Back Pain Trial (IMPACT): A Pilot Trial Protocol. BMC Musculoskelet. Disord. 2016, 17, 36. [Google Scholar] [CrossRef] [PubMed]
  12. Panhwar, Y.N.; Naghdy, F.; Naghdy, G.; Stirling, D.; Potter, J. Assessment of Frailty: A Survey of Quantitative and Clinical Methods. BMC Biomed. Eng. 2019, 1, 7. [Google Scholar] [CrossRef] [PubMed]
  13. Oliveira, C.B.; Franco, M.R.; Maher, C.G.; Tiedemann, A.; Silva, F.G.; Damato, T.M.; Nicholas, M.K.; Christofaro, D.G.D.; Pinto, R.Z. The Efficacy of a Multimodal Physical Activity Intervention with Supervised Exercises, Health Coaching and an Activity Monitor on Physical Activity Levels of Patients with Chronic, Nonspecific Low Back Pain (Physical Activity for Back Pain (PAyBACK) Trial): Study Protocol for a Randomised Controlled Trial. Trials 2018, 19, 40. [Google Scholar] [CrossRef] [PubMed]
  14. Janevic, M.R.; Shute, V.; Murphy, S.L.; Piette, J.D. Acceptability and Effects of Commercially Available Activity Trackers for Chronic Pain Management Among Older African American Adults. Pain Med. 2020, 21, e68–e78. [Google Scholar] [CrossRef]
  15. Patel, M.S.; Asch, D.A.; Volpp, K.G. Wearable Devices as Facilitators, Not Drivers, of Health Behavior Change. JAMA 2015, 313, 459. [Google Scholar] [CrossRef]
  16. Kuduz, H.; Kaçar, F. A Deep Learning Approach for Human Gait Recognition from Time-Frequency Analysis Images of Inertial Measurement Unit Signal. Int. J. Appl. Methods Electron. Comput. 2023, 11, 165–173. [Google Scholar] [CrossRef]
  17. Tran, L.; Hoang, T.; Nguyen, T.D.; Kim, H.; Choi, D. Multi-Model Long Short-Term Memory Network for Gait Recognition Using Window-Based Data Segment. IEEE Access 2021, 9, 23826–23839. [Google Scholar] [CrossRef]
  18. Ruiz-Ruiz, L.; Seco, F.; Jimenez, A.R.; Garcia, S.; Garcia, J.J. Evaluation of Gait Parameter Estimation Accuracy: A Comparison between Commercial IMU and Optical Capture Motion System. In Proceedings of the 2022 IEEE International Symposium on Medical Measurements and Applications (MeMeA), Messina, Italy, 22–15 June 2022; pp. 1–2. [Google Scholar] [CrossRef]
  19. Razjouyan, J.; Naik, A.D.; Horstman, M.J.; Kunik, M.E.; Amirmazaheri, M.; Zhou, H.; Sharafkhaneh, A.; Najafi, B. Wearable Sensors and the Assessment of Frailty among Vulnerable Older Adults: An Observational Cohort Study. Sensors 2018, 18, 1336. [Google Scholar] [CrossRef]
  20. Chen, M.; Wang, H.; Yu, L.; Yeung, E.H.K.; Luo, J.; Tsui, K.-L.; Zhao, Y. A Systematic Review of Wearable Sensor-Based Technologies for Fall Risk Assessment in Older Adults. Sensors 2022, 22, 6752. [Google Scholar] [CrossRef]
  21. Zhou, H.; Park, C.; Shahbazi, M.; York, M.K.; Kunik, M.E.; Naik, A.D.; Najafi, B. Digital Biomarkers of Cognitive Frailty: The Value of Detailed Gait Assessment Beyond Gait Speed. Gerontology 2022, 68, 224–233. [Google Scholar] [CrossRef]
  22. Fan, S.; Ye, J.; Xu, Q.; Peng, R.; Hu, B.; Pei, Z.; Yang, Z.; Xu, F. Digital Health Technology Combining Wearable Gait Sensors and Machine Learning Improve the Accuracy in Prediction of Frailty. Front. Public Health 2023, 11, 1169083. [Google Scholar] [CrossRef] [PubMed]
  23. Zebin, T.; Scully, P.; Ozanyan, K. Human Activity Recognition With Inertial Sensors Using a Deep Learning Approach. In Proceedings of the 2016 IEEE SENSORS, Orlando, FL, USA, 30 October–3 November 2016. [Google Scholar] [CrossRef]
  24. Mukhopadhyay, S.C. Wearable Sensors for Human Activity Monitoring: A Review. IEEE Sens. J. 2015, 15, 1321–1330. [Google Scholar] [CrossRef]
  25. Hassan, M.M.; Uddin, M.Z.; Mohamed, A.; Almogren, A. A Robust Human Activity Recognition System Using Smartphone Sensors and Deep Learning. Future Gener. Comput. Syst. 2018, 81, 307–313. [Google Scholar] [CrossRef]
  26. Ahmed, N.; Rafiq, J.I.; Islam, M.R. Enhanced Human Activity Recognition Based on Smartphone Sensor Data Using Hybrid Feature Selection Model. Sensors 2020, 20, 317. [Google Scholar] [CrossRef]
  27. Son, M.; Han, S.H.; Lyoo, C.H.; Lim, J.A.; Jeon, J.; Hong, K.-B.; Park, H. The Effect of Levodopa on Bilateral Coordination and Gait Asymmetry in Parkinson’s Disease Using Inertial Sensor. NPJ Park. Dis. 2021, 7, 42. [Google Scholar] [CrossRef]
  28. Nagano, H.; Sarashina, E.; Sparrow, W.; Mizukami, K.; Begg, R. General Mental Health Is Associated with Gait Asymmetry. Sensors 2019, 19, 4908. [Google Scholar] [CrossRef]
  29. García-de-Villa, S.; Neira, G.G.-V.; Álvarez, M.N.; Huertas-Hoyas, E.; Ruiz, L.R.; del-Ama, A.J.; Sánchez, M.C.R.; Jiménez, A.R. A Database with Frailty, Functional and Inertial Gait Metrics for the Research of Fall Causes in Older Adults. Sci. Data 2023, 10, 566. [Google Scholar] [CrossRef]
  30. Hirano, Y.; Yamada, Y.; Akiyama, Y.; Nakamura, H.; Matsui, Y. Preliminary Gait Analysis of Frail versus Older Adults. J. Phys. Ther. Sci. 2024, 36, 2023–2105. [Google Scholar] [CrossRef]
  31. Apsega, A.; Petrauskas, L.; Alekna, V.; Daunoraviciene, K.; Sevcenko, V.; Mastaviciute, A.; Vitkus, D.; Tamulaitiene, M.; Griskevicius, J. Wearable Sensors Technology as a Tool for Discriminating Frailty Levels During Instrumented Gait Analysis. Appl. Sci. 2020, 10, 8451. [Google Scholar] [CrossRef]
  32. Song, S.-M.; Fernandes, N.; Nordin, A.D. Characterizing Bodyweight-Supported Treadmill Walking on Land and Underwater Using Foot-Worn Inertial Measurement Units and Machine Learning for Gait Event Detection. Sensors 2023, 23, 7945. [Google Scholar] [CrossRef]
  33. Zrenner, M.; Küderle, A.; Roth, N.; Jensen, U.; Dümler, B.; Eskofier, B.M. Does the Position of Foot-Mounted IMU Sensors Influence the Accuracy of Spatio-Temporal Parameters in Endurance Running? Sensors 2020, 20, 5705. [Google Scholar] [CrossRef] [PubMed]
  34. Fatima, R.; Khan, M.H.; Nisar, M.A.; Doniec, R.; Farid, M.S.; Grzegorzek, M. A Systematic Evaluation of Feature Encoding Techniques for Gait Analysis Using Multimodal Sensory Data. Sensors 2023, 24, 75. [Google Scholar] [CrossRef] [PubMed]
  35. Aubin, P.M.; Serackis, A.; Griskevicius, J. Support Vector Machine Classification of Parkinson’s Disease, Essential Tremor and Healthy Control Subjects Based on Upper Extremity Motion. In Proceedings of the 2012 International Conference on Biomedical Engineering and Biotechnology, Macau, China, 28–30 May 2012; pp. 900–904. [Google Scholar] [CrossRef]
  36. Liu, Y.; He, X.; Wang, R.; Teng, Q.; Hu, R.; Qing, L.; Wang, Z.; He, X.; Yin, B.; Mou, Y.; et al. Application of Machine Vision in Classifying Gait Frailty Among Older Adults. Front. Aging Neurosci. 2021, 13, 757823. [Google Scholar] [CrossRef] [PubMed]
  37. Espinoza Bernal, V.C.; Hiremath, S.V.; Wolf, B.; Riley, B.; Mendonca, R.J.; Johnson, M.J. Classifying and Tracking Rehabilitation Interventions through Machine-Learning Algorithms in Individuals with Stroke. J. Rehabil. Assist. Technol. Eng. 2021, 8, 205566832110446. [Google Scholar] [CrossRef]
  38. Panwar, M.; Biswas, D.; Bajaj, H.; Jobges, M.; Turk, R.; Maharatna, K.; Acharyya, A. Rehab-Net: Deep Learning Framework for Arm Movement Classification Using Wearable Sensors for Stroke Rehabilitation. IEEE Trans. Biomed. Eng. 2019, 66, 3026–3037. [Google Scholar] [CrossRef]
  39. Romijnders, R.; Warmerdam, E.; Hansen, C.; Schmidt, G.; Maetzler, W. A Deep Learning Approach for Gait Event Detection from a Single Shank-Worn IMU: Validation in Healthy and Neurological Cohorts. Sensors 2022, 22, 3859. [Google Scholar] [CrossRef]
  40. Park, C.; Mishra, R.; Golledge, J.; Najafi, B. Digital Biomarkers of Physical Frailty and Frailty Phenotypes Using Sensor-Based Physical Activity and Machine Learning. Sensors 2021, 21, 5289. [Google Scholar] [CrossRef]
  41. Amjad, A.; Qaiser, S.; Błaszczyszyn, M.; Szczęsna, A. The Evolution of Frailty Assessment Using Inertial Measurement Sensor-based Gait Parameter Measurements: A Detailed Analysis. WIREs Data Min. Knowl. Discov. 2024, 14, e1557. [Google Scholar] [CrossRef]
  42. Carvajal-Castaño, H.A.; Pérez-Toro, P.A.; Orozco-Arroyave, J.R. Classification of Parkinson’s Disease Patients—A Deep Learning Strategy. Electronics 2022, 11, 2684. [Google Scholar] [CrossRef]
  43. Zhao, Y.; Dong, F.; Sun, T.; Ju, Z.; Yang, L.; Shan, P.; Li, L.; Lv, X.; Lian, C. Image Expression of Time Series Data of Wearable IMU Sensor and Fusion Classification of Gymnastics Action. Expert Syst. Appl. 2024, 238, 121978. [Google Scholar] [CrossRef]
  44. Arshad, M.Z.; Jung, D.; Park, M.; Shin, H.; Kim, J.; Mun, K.-R. Gait-Based Frailty Assessment Using Image Representation of IMU Signals and Deep CNN. arXiv 2021. [Google Scholar] [CrossRef]
  45. Ismail Fawaz, H.; Lucas, B.; Forestier, G.; Pelletier, C.; Schmidt, D.F.; Weber, J.; Webb, G.I.; Idoumghar, L.; Muller, P.-A.; Petitjean, F. InceptionTime: Finding AlexNet for Time Series Classification. Data Min. Knowl. Discov. 2020, 34, 1936–1962. [Google Scholar] [CrossRef]
  46. Amjad, A.; Szczeęsna, A.; Błaszczyszyn, M.; Sacha, J.; Sacha, M.; Feusette, P.; Wolański, W.; Konieczny, M.; Borysiuk, Z.; Khan, B. Deep Learning for Frailty Classification Using IMU Sensor Data: Insights From FRAILPOL Database. IEEE Sens. J. 2025, 25, 3974–3981. [Google Scholar] [CrossRef]
  47. Patoz, A.; Lussiana, T.; Breine, B.; Gindre, C.; Malatesta, D. Comparison of Different Machine Learning Models to Enhance Sacral Acceleration-Based Estimations of Running Stride Temporal Variables and Peak Vertical Ground Reaction Force. Sports Biomech. 2023, 1–17. [Google Scholar] [CrossRef]
  48. Trabassi, D.; Serrao, M.; Varrecchia, T.; Ranavolo, A.; Coppola, G.; De Icco, R.; Tassorelli, C.; Castiglia, S.F. Machine Learning Approach to Support the Detection of Parkinson’s Disease in IMU-Based Gait Analysis. Sensors 2022, 22, 3700. [Google Scholar] [CrossRef]
  49. Gaffoor, Z.; Pietersen, K.; Jovanovic, N.; Bagula, A.; Kanyerere, T.; Ajayi, O.; Wanangwa, G.J. A Comparison of Ensemble and Deep Learning Algorithms to Model Groundwater Levels in a Data-Scarce Aquifer of Southern Africa. Hydrology 2022, 9, 125. [Google Scholar] [CrossRef]
  50. Janssen, W.; Bussmann, J.; Horemans, H.; Stam, H. Analysis and decomposition of accelerometric signals of trunk and thigh obtained during the sit-to-stand movement. Med. Biol. Eng. Comput. 2005, 43, 265–272. [Google Scholar] [CrossRef]
  51. Tan, T.; Shull, P.B.; Hicks, J.L.; Uhlrich, S.D.; Chaudhari, A. Self-Supervised Learning Improves Accuracy and Data Efficiency for IMU-Based Ground Reaction Force Estimation. bioRxiv 2023, 10, 564057. [Google Scholar] [CrossRef]
  52. Wang, S.; Pitts, J.; Purohit, R.; Shah, H. The Influence of Low-Pass Filtering Methods on Motion Data in Machine Learning Models. Appl. Sci. 2025, 15, 2177. [Google Scholar] [CrossRef]
  53. Jiang, M.; Shang, H.; Wang, Z.; Li, H.; Wang, Y. A method to deal with installation errors of wearable accelerometers for human activity recognition. Physiol. Meas. 2011, 32, 347. [Google Scholar] [CrossRef]
  54. Gupta, S.; Agrawal, M.; Deepak, D. Gammatonegram based triple classification of lung sounds using deep convolutional neural network with transfer learning. Biomed. Sign. Proc. Control 2011, 70, 102947. [Google Scholar] [CrossRef]
  55. Alnuaim, A.A.; Zakariah, M.; Shashidhar, C.; Hatamleh, W.A.; Tarazi, H.; Shukla, P.K.; Ratna, R. Speaker Gender Recognition Based on Deep Neural Networks and ResNet50. Wirel. Commun. Mob. Comput. 2022, 2022, 4444388. [Google Scholar] [CrossRef]
  56. Álvarez, M.N.; Ruiz, A.R.; Neira, G.G.; Huertas-Hoyas, E.; Cerda, M.T.; Delgado, L.P.; Robles, E.R.; Del-Ama, A.J.; Ruiz-Ruiz, L.; García-de-Villa, S.; et al. Assessing falls in the elderly population using G-STRIDE foot-mounted inertial sensor. Sci. Rep. 2023, 13, 9208. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Multimodal deep learning model architecture.
Figure 1. Multimodal deep learning model architecture.
Sensors 25 03351 g001
Figure 2. Raw gyroscope and accelerometer signals of IMU fixed on thigh segment.
Figure 2. Raw gyroscope and accelerometer signals of IMU fixed on thigh segment.
Sensors 25 03351 g002
Figure 5. Spectrograms generated from accelerometer signals.
Figure 5. Spectrograms generated from accelerometer signals.
Sensors 25 03351 g005
Figure 6. Subject-level classification confusion matrices comparing raw, processed, and filtered spectrogram models. Color intensity indicates classification accuracy (darker colors represent higher values).
Figure 6. Subject-level classification confusion matrices comparing raw, processed, and filtered spectrogram models. Color intensity indicates classification accuracy (darker colors represent higher values).
Sensors 25 03351 g006
Figure 7. Spectrogram-level classification confusion matrices comparing raw, processed and filtered spectrogram models. Color intensity indicates classification accuracy (darker colors represent higher values).
Figure 7. Spectrogram-level classification confusion matrices comparing raw, processed and filtered spectrogram models. Color intensity indicates classification accuracy (darker colors represent higher values).
Sensors 25 03351 g007
Figure 8. Effect of data leakage on test accuracy.
Figure 8. Effect of data leakage on test accuracy.
Sensors 25 03351 g008
Table 1. Clinical evaluation features.
Table 1. Clinical evaluation features.
FeatureFrailPreFrailNoFrailp-Value
Grip strength
Right 1 Grip Force20.51 ± 8.8227.58 ± 8.7728.35 ± 7.580.00013
Left 1 Grip Force20.31 ± 9.4526.55 ± 9.8327.90 ± 7.640.00142
Right 2 Grip Force20.54 ± 9.0928.19 ± 9.0830.13 ± 7.390.00002
Left 2 Grip Force20.57 ± 9.3327.46 ± 10.0528.71 ± 7.410.00049
Right 3 Grip Force20.71 ± 9.0528.73 ± 8.7930.13 ± 7.580.00001
Left 3 Grip Force20.60 ± 9.4527.61 ± 9.5928.58 ± 7.730.00039
Max Grip Force23.06 ± 8.7731.22 ± 8.7331.81 ± 7.910.00001
Mobility
TUG time15.67 ± 4.8912.69 ± 4.537.64 ± 1.840.00000
Tinneti Test score20.54 ± 3.9022.28 ± 3.4925.48 ± 2.710.00000
TT balance11.40 ± 2.3912.24 ± 1.9914.23 ± 1.910.00000
TT gait9.06 ± 2.0410.03 ± 1.8711.16 ± 1.370.00004
BERG balance37.77 ± 6.8140.91 ± 5.8849.16 ± 7.370.00000
Dynamic Gait Index14.26 ± 4.2216.15 ± 3.5019.32 ± 3.340.00000
Anthropometrics
Age78.80 ± 7.5074.00 ± 8.4773.19 ± 6.340.00479
Height (cm)160.86 ± 6.43165.79 ± 9.59162.74 ± 9.600.02418
Weight (kg)68.43 ± 15.9675.85 ± 15.6976.13 ± 16.270.00604
BMI26.47 ± 6.2927.68 ± 5.6728.72 ± 5.520.29165
Spatiotemporal Gait Parameters
Gait Velocity (m/s)0.51 ± 0.130.63 ± 0.130.97 ± 0.160.00000
Gait Time (s)8.42 ± 2.516.69 ± 1.614.25 ± 0.790.00000
Right Swing (s)0.50 ± 0.080.51 ± 0.080.44 ± 0.060.00025
Left Swing (s)0.53 ± 0.070.54 ± 0.060.48 ± 0.070.00007
Right Stance (s)0.97 ± 0.210.84 ± 0.130.63 ± 0.110.00000
Left Stance (s)0.93 ± 0.210.81 ± 0.130.59 ± 0.100.00000
Right Stride (s)1.47 ± 0.251.35 ± 0.161.07 ± 0.150.00000
Left Stride (s)1.46 ± 0.231.35 ± 0.161.06 ± 0.150.00000
Double Support (s)0.21 ± 0.080.15 ± 0.070.08 ± 0.040.00000
Cadence84.54 ± 12.9891.32 ± 11.83116.75 ± 13.170.00000
Step Time (s)0.73 ± 0.130.67 ± 0.080.52 ± 0.060.00000
Right Swing (%)34.51 ± 5.0237.64 ± 4.3841.28 ± 3.300.00000
Left Swing (%)36.55 ± 4.7140.35 ± 4.0744.98 ± 3.500.00000
Right Stance (%)65.49 ± 5.0262.36 ± 4.3858.72 ± 3.300.00000
Left Stance (%)63.45 ± 4.7159.65 ± 4.0755.02 ± 3.500.00000
Table 2. Classification performance results.
Table 2. Classification performance results.
ClassPrecisionRecallF1-ScoreSupportAverage Accuracy (%)
Raw IMU data + clinical features
Frail0.945 ± 0.0750.925 ± 0.0810.933 ± 0.070756
PreFrail0.881 ± 0.0770.811 ± 0.1080.841 ± 0.0811512
NoFrail0.703 ± 0.1010.825 ± 0.0990.752 ± 0.07264885.19
macro avg0.843 ± 0.0710.854 ± 0.0710.842 ± 0.0712916
weighted avg0.858 ± 0.0690.844 ± 0.0730.845 ± 0.0732916
Baseline processed IMU data + clinical features
Frail0.993 ± 0.0070.996 ± 0.0060.995 ± 0.003756
PreFrail0.907 ± 0.0370.854 ± 0.0580.879 ± 0.0471512
NoFrail0.715 ± 0.1060.801 ± 0.0700.754 ± 0.08764888.15
macro avg0.871 ± 0.0450.884 ± 0.0430.876 ± 0.0452916
weighted avg0.886 ± 0.0410.879 ± 0.0450.881 ± 0.0442916
Low-pass filtered IMU data + clinical features
Frail0.980 ± 0.0350.979 ± 0.0410.979 ± 0.024756
PreFrail0.915 ± 0.0420.846 ± 0.0560.878 ± 0.0431512
NoFrail0.719 ± 0.0780.838 ± 0.0660.773 ± 0.07064888.15
macro avg0.871 ± 0.0410.888 ± 0.0410.877 ± 0.0412916
weighted avg0.888 ± 0.0380.879 ± 0.0420.881 ± 0.0412916
Table 3. Classification performance results of spectrogram-only models.
Table 3. Classification performance results of spectrogram-only models.
SpectrogramsAccuracy (%)Macro PrecisionMacro RecallMacro F1
Subject-level
Raw71.430.740.680.70
Baseline64.290.670.620.63
Filtered64.290.680.610.63
Spectrogram-level
Raw86.080.860.850.85
Baseline97.220.970.970.97
Filtered92.140.920.920.92
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Griškevičius, J.; Daunoravičienė, K.; Petrauskas, L.; Apšega, A.; Alekna, V. Retrospective Frailty Assessment in Older Adults Using Inertial Measurement Unit-Based Deep Learning on Gait Spectrograms. Sensors 2025, 25, 3351. https://doi.org/10.3390/s25113351

AMA Style

Griškevičius J, Daunoravičienė K, Petrauskas L, Apšega A, Alekna V. Retrospective Frailty Assessment in Older Adults Using Inertial Measurement Unit-Based Deep Learning on Gait Spectrograms. Sensors. 2025; 25(11):3351. https://doi.org/10.3390/s25113351

Chicago/Turabian Style

Griškevičius, Julius, Kristina Daunoravičienė, Liudvikas Petrauskas, Andrius Apšega, and Vidmantas Alekna. 2025. "Retrospective Frailty Assessment in Older Adults Using Inertial Measurement Unit-Based Deep Learning on Gait Spectrograms" Sensors 25, no. 11: 3351. https://doi.org/10.3390/s25113351

APA Style

Griškevičius, J., Daunoravičienė, K., Petrauskas, L., Apšega, A., & Alekna, V. (2025). Retrospective Frailty Assessment in Older Adults Using Inertial Measurement Unit-Based Deep Learning on Gait Spectrograms. Sensors, 25(11), 3351. https://doi.org/10.3390/s25113351

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop