Next Article in Journal
Analysis of Surface Deformation and Its Relationship with Land Use in the Reclaimed Land of Tianjin Based on Time Series InSAR
Previous Article in Journal
Dual Illumination and Detection Photoacoustic Tomography of Hollow Metal Cylinders
Previous Article in Special Issue
Beyond VO2 Peak: Hemodynamic Profiling After Bariatric Surgery Using the Fick Principle
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Self-Supervised Gait Event Detection from Smartphone IMUs for Human Performance and Sports Medicine

by
Andreea Maria Mănescu
1 and
Dan Cristian Mănescu
2,*
1
Doctoral School, Bucharest University of Economic Studies, 010374 Bucharest, Romania
2
Department of Physical Education and Sport, Bucharest University of Economic Studies, 010374 Bucharest, Romania
*
Author to whom correspondence should be addressed.
Appl. Sci. 2025, 15(22), 11974; https://doi.org/10.3390/app152211974
Submission received: 23 October 2025 / Revised: 9 November 2025 / Accepted: 10 November 2025 / Published: 11 November 2025
(This article belongs to the Special Issue Exercise, Fitness, Human Performance and Health: 2nd Edition)

Abstract

Background: Gait event detection from inertial sensors offers scalable insights into locomotor health, with applications in clinical monitoring and mobile health. However, supervised methods are limited by scarce annotations, device variability, and sensor placement shifts. This in silico study evaluates self-supervised learning (SSL) as a resource-efficient strategy to improve robustness and generalizability. Methods: Six public smartphone and wearable inertial measurements unit (IMU) datasets (WISDM, PAMAP2, KU-HAR, mHealth, OPPORTUNITY, RWHAR) were harmonized within a unified deep learning pipeline. Models were pretrained on unlabeled windows using contrastive SSL with sensor-aware augmentations, then fine-tuned with varying label fractions. Experiments systematically assessed included (1) pretraining scale, (2) label efficiency, (3) augmentation contributions, (4) device/placement shifts, (5) sampling-rate sensitivity, and (6) backbone comparisons (CNN, TCN, BiLSTM, Transformer). Results: SSL consistently outperformed supervised baselines. Pretraining yielded accuracy gains of ΔF1 +0.08–0.15 and reduced stride-time error by −8 to −12 ms. SSL cut label needs by up to 95%, achieving competitive performance with only 5–10% labeled data. Sensor-aware augmentations, particularly axis-swap and drift, drove the strongest transfer gains. Robustness was maintained across sampling rates (25–100 Hz) and device/placement shifts. CNNs and TCNs offered the best efficiency–accuracy trade-offs, while Transformers delivered the highest accuracy at greater cost. Conclusions: This computational analysis across six datasets shows SSL enhances gait event detection with improved accuracy, efficiency, and robustness under minimal supervision, establishing a scalable framework for human performance and sports medicine in clinical and mobile health applications.

1. Introduction

Gait analysis provides essential markers of human health, reflecting neurological function, musculoskeletal integrity, and overall physical performance. Subtle deviations in gait parameters are associated with neurodegenerative disorders, frailty, and impaired mobility; therefore, it is critical for both clinical practice and sports science, and directly relevant to human performance and sports medicine.
The availability of inertial measurement units (IMUs) in smartphones offers an accessible and scalable means to perform gait analysis outside specialized laboratories. However, models trained on a single dataset or device often degrade when applied to new sensing conditions. Differences in sampling rates, sensor placements, mounting orientations, and participant populations introduce domain shifts that reduce accuracy and limit clinical utility.
Self-supervised learning (SSL) has emerged as a promising strategy to address these challenges. By leveraging large pools of unlabelled sensor data, SSL can learn invariant representations that generalize across devices and contexts while requiring only limited annotation for downstream tasks. In other domains, SSL has been shown to improve cross- dataset generalization, reduce annotation costs, and capture low-level invariances essential for robust performance.
Building on this foundation, the present study advances a sensor-aware SSL framework for gait event detection and temporal estimation. To ensure full reproducibility and generalizability, all analyses were conducted entirely in silico, leveraging six heterogeneous IMU datasets without additional data collection. This purely computational design eliminates sources of variability inherent to prospective trials while enabling rigorous control over preprocessing, augmentation, and evaluation protocols. The framework integrates contrastive pretraining strategies with augmentations grounded in sensor physics and establishes reproducible cross-dataset transfer protocols. Within this context, we formulated the following working hypotheses:
H1. 
Self-supervised pretraining on heterogeneous, unlabelled IMU windows will produce device- and placement-invariant representations that improve out-of-domain event detection compared to fully supervised baselines.
H2. 
Minimal labelled data (≈5–10% of windows) will be sufficient to reach or exceed the performance of fully supervised training, demonstrating label efficiency.
H3. 
Sensor-aware augmentations (axis-swap, drift, jitter, time-warp) will provide measurable gains in transferability, with axis-swap and drift expected to contribute most strongly.
H4. 
Self-supervised pretraining will yield device- and placement-invariant representations, maintaining accuracy and timing precision across heterogeneous sensor placements and smartphone models.
H5. 
Representations will remain robust across a wide range of sampling rates (25–100 Hz), ensuring portability to commodity smartphones with diverse hardware specifications.
H6. 
Performance gains from self-supervised pretraining will generalize across distinct backbone architectures (CNN, TCN, BiLSTM, Transformer), with Transformers yielding the highest absolute accuracy, while CNNs and TCNs provide more efficient trade-offs between speed and memory.
Literature review—Gait is widely recognized as a fundamental indicator of human health, providing insights into neurological integrity, musculoskeletal coordination, anoverall functional capacity. Subtle abnormalities in step timing, cadence, or symmetry are associated with a broad spectrum of conditions ranging from Parkinson’s disease to frailty in older adults. Consequently, robust gait monitoring is increasingly regarded as a cornerstone of preventive healthcare and rehabilitation [1,2].
Traditional gait analysis relies on motion capture systems, instrumented walkways, or force plates—technologies that, while precise, are expensive and limited to laboratory settings. In contrast, inertial measurement units (IMUs), and particularly those embedded in smartphones, provide an affordable, scalable, and ubiquitous alternative. This accessibility has driven significant interest in developing algorithms capable of detecting gait events and deriving temporal parameters directly from accelerometer and gyroscope data [3,4,5,6]. In parallel, other wearable technologies have progressed beyond inertial sensing. Smart photonic wristbands for pulse-wave monitoring, optical fiber–based strain sensors, and hybrid bio-optoelectronic systems have been introduced for physiological and biomechanical assessment, further expanding the ecosystem of wearable health and performance analytics.
Early methods for gait event detection relied on signal processing heuristics such as peak detection in vertical acceleration or thresholding of angular velocity. While conceptually simple and computationally efficient, these approaches are sensitive to device orientation, placement variability, and noise. Their limited robustness across populations and devices has restricted their use in real-world deployments [7,8,9,10].
The rise of machine learning introduced supervised models—support vector machines, random forests, and later convolutional and recurrent neural networks—for gait event detection and parameter estimation. These methods improved accuracy within controlled datasets but required substantial volumes of labeled data. More importantly, they often showed dramatic performance drops when applied to unseen devices or populations, underscoring the challenge of domain shift [11,12,13,14].
Recent work has adapted advanced architectures such as 1D convolutional neural networks and LSTMs to IMU sequences, reporting strong in-domain accuracy for event classification and step-time estimation. Transformer-based models have also begun to appear, motivated by their ability to capture long-range dependencies. However, the reliance on supervised training and single-dataset evaluation limits their external validity [15,16,17,18].
A recurring challenge across these studies is the lack of cross-dataset evaluation. Models that excel within a single dataset frequently fail when transferred to another, due to differences in sampling rates, sensor specifications, placements, and subject demographics. This phenomenon has been termed the “dataset trap,” and it severely constrains the translation of IMU-based gait analysis into clinical and field applications [19,20,21,22,23].
Several studies have explored transfer learning, fine-tuning models trained on a source dataset to a new target dataset, or applying domain-adversarial strategies to reduce distribution shift. While these approaches offer modest improvements, they often still require significant labeled data from the target domain and lack consistent reproducibility. Moreover, their dependence on supervised adaptation prevents full scalability [24,25,26,27,28].
Self-supervised learning (SSL) has emerged as a powerful paradigm for time-series representation learning. Methods such as contrastive learning, predictive coding, and masked reconstruction have shown success in domains like human activity recognition, physiological signal monitoring, and audio analysis. By exploiting pretext tasks or augmentation-based objectives, SSL can learn rich representations from unlabelled data, reducing annotation demands [29,30,31,32].
Recent studies have applied SSL to IMU data, demonstrating improvements in downstream classification of activities of daily living. Contrastive frameworks such as SimCLR and InfoNCE variants have proven effective in capturing invariances to orientation, noise, and temporal distortions. However, the majority of these works focus on coarse-grained activity recognition rather than fine-grained gait events, and evaluations are often limited to in-domain settings [33,34,35,36].
The design of augmentations is central to SSL performance. In the context of IMU data, transformations such as axis permutation, sensor drift, time warping, jittering, and channel dropout can simulate the variability encountered across devices and placements. Prior work has highlighted the importance of augmentation choice but often without systematic ablation, leaving uncertainty about which transformations most directly enhance cross-dataset generalization [37,38,39,40].
Despite growing interest in SSL for wearable sensing, no study to date has established a reproducible framework for gait event detection and temporal estimation that demonstrates clear cross-dataset generalization. Existing works rarely report strict pretrain-on-A, test-on-B protocols, and label efficiency analyses are often absent [41,42]. Moreover, the lack of transparent methodological specifications hinders replication and limits trust in reported gains [43,44]. This leaves a critical gap: the field still lacks a rigorously validated, cross-dataset SSL framework for fine-grained gait event detection, an essential step for ensuring reproducibility, external validity, and translational applicability.
Against this backdrop, the present work addresses the unmet need for a generalizable, reproducible SSL approach to gait analysis. By harmonizing multiple public datasets, applying sensor-aware augmentations, and enforcing rigorous cross-dataset protocols, we test whether SSL can overcome the dataset trap, reduce annotation costs, and enable robust gait monitoring with commodity smartphones. In doing so, this study aims to advance both methodological clarity and translational potential in the field.

2. Materials and Methods

To rigorously evaluate the role of self-supervised learning in gait event detection, we established a unified experimental framework designed to ensure both reproducibility and generalizability. The overall methodological pipeline is illustrated in Figure 1, which clearly summarizes the sequential stages of data harmonization, self-supervised pretraining, downstream fine-tuning, evaluation protocol, and computing environment.
The study was conceived as a coherent whole, where the interdependence of components was a deliberate design choice to secure consistency, reproducibility, and generalizability. By embedding these principles into the structure of the workflow, the approach moves beyond a sequence of technical steps and establishes a solid foundation for the subsequent empirical analyses.

2.1. Data and Harmonization

Datasets—we considered six public datasets representative of smartphone and wearable IMU signals: WISDM, PAMAP2, KU-HAR, mHealth, OPPORTUNITY, and RWHAR. Inclusion criteria were: (1) availability of at least a tri-axial accelerometer; (2) identifiable or annotated walking segments; and (3) subject-level splits. Where gyroscope channels were available, we used 6 axes; otherwise, we retained only the 3 axis accelerometer and set gyroscope channels to zero during pretraining. We deliberately combined long-established benchmark datasets (WISDM, PAMAP2, OPPORTUNITY, mHealth) with more recent smartphone collections (KU-HAR, RWHAR) to ensure both comparability with prior work and coverage of newer sensing hardware, thus spanning nearly a decade of wearable technology and acquisition protocols.
Detailed dataset characteristics are summarized in Table 1. Heel-strike (HS) and toe-off (TO) annotations were taken from the official dataset releases. In datasets where direct force-plate recordings were not available, event labels are proxy measures (e.g., vertical acceleration peaks or plantar pressure signals). We treat such annotations as noisy ground-truth and apply a ±50 ms matching tolerance during evaluation. All datasets are public and de-identified; no new data collection was performed, so IRB approval was not required.
Preprocessing—to harmozine these heterogenous sources, all streams were resampled to 50 Hz using polyphase resampling (anti-alias FIR). Signals were clipped to the 0.1–99.9 percentile to remove spikes, then z-scored per subject on the pretraining pool to reduce subject leakage. We estimate the gravity vector by a 1.5 s moving average on accelerometer and align the device frame to the body frame via a yaw-invariant rotation; remaining yaw is left free to preserve heading.
Windows—data were segmented in windows of 3.2 s (160 samples at 50 Hz), chosen to cover approximately two gait cycles at typical adult walking speeds (stride duration ≈ 1.0–1.2 s). This window length balances temporal resolution with computational efficiency. This choice is consistent with prior work reporting average stride times around 1.1 s in healthy adults.
Leakage control—to prevent information leakage, subject IDs used for downstream testing were excluded from pretraining pools. A manifest containing dataset, subject, session, file hash identifiers was maintained to ensure auditability and reproducibility.

2.2. SSL Encoder (Backbone)

We implemented two compact backbones for self-supervised pretraining on IMU windows.
  • 1D-CNN (dilated): the first was a one-dimensional dilated convolutional neural network (1D-CNN), configured with 6→64→128→256 channels, kernel sizes of 7/7/5, dilations of 1/2/4, and a stride of 2 in the first block. GELU activations and residual connections were applied throughout, followed by global average pooling to obtain a 256-dimensional embedding.
  • Tiny Transformer: the second backbone was a lightweight Transformer. Input windows were patchified with stride 2 (patch length 4) and processed through four encoder blocks with hidden dimension 256, 4 attention heads, and dropout of 0.1. A [CLS] token representation was pooled into a 256-dimensional embedding. A two-layer MLP (256→256→128) with batch normalization then projected embeddings into the contrastive space.
Objective—pretraining followed a contrastive objective (NT-Xent) with temperature τ = 0.07 and cosine similarity. Mini-batches consisted of 256 windows, with gradient accumulation applied when required.
Optimization—employed AdamW (β1 = 0.9, β2 = 0.999, weight decay = 0.01) with a cosine learning-rate schedule and a warm-up of 2 epochs. The base learning rate was 1 × 10−3 for the CNN and 5 × 10−4 for the Transformer. Pretraining was run for 100 epochs over the pooled unlabelled windows.

2.3. Sensor-Aware Augmentations

Positive pairs were generated by composing two randomly sampled transformations from a family of sensor-aware augmentations, while negatives were defined as all other windows in the batch. Augmentations were designed to simulate realistic sources of variability across devices, placements, and sampling conditions.
Piecewise linear Time-warping was applied with 2–4 control points and a maximum distortion of ±10%. Jitter consisted of Gaussian noise with σ ∈ {0.005, 0.01, 0.02} × channel standard deviation. Magnitude scaling randomly adjusted per-channel gain within a uniform range g ∈ U(0.8, 1.2). Axis-swap involved random permutation or flipping of axes with probability p = 0.5, applied independently to accelerometer and gyroscope channels. Sensor drift was simulated as an additive linear ramp b·t with slope b ∈ U(−0.02, 0.02) × channel standard deviation per second. Channel dropout masked one randomly selected channel with probability p = 0.2. Finally, Time-masking zeroed out a contiguous span of 5–20 samples with probability p = 0.3.
For ablation analyses, each augmentation family was independently disabled and severity levels were systematically varied in order to quantify their contribution to transfer performance. Severity levels for all augmentations were determined empirically through pilot sensitivity analyses to ensure realistic signal variability without distorting gait dynamics.

2.4. Downstream Heads and Labels

Two downstream tasks were considered: gait event detection and temporal parameter estimation.
Event detection—we implemented a temporal convolutional head consisting of three blocks with kernel size 5, dilations of 1/2/4, and 128 channels. The head predicted heel- strike and toe-off logits for each timestep. Training was performed using a combination of weighted binary cross-entropy, with class weights set as the inverse of class frequency, and focal loss (γ = 1.5) for sensitivity analyses. Predictions were matched to reference events within ±50 ms, with unmatched predictions counted as false positives. Reference events were taken directly from each dataset’s official annotations (e.g., force-plate labels in PAMAP2, footswitch signals in OPPORTUNITY, or proxy segmentations in WISDM), without relabeling or post hoc adjustment. This ensured consistency with prior work and preserved comparability across datasets.
Temporal parameter estimation—we employed a regression head based on a bidirectional LSTM (128 units × 1 layer) followed by a linear projection. The head estimated step, stride, stance, and swing times directly from backbone features. The loss combined an L1 objective with an additional total variation penalty (0.1 × TV) to enforce smoothness across adjacent windows.

2.5. Baselines

We compared the proposed self-supervised approach against three baselines.
Fully supervised model—the first was a fully supervised model trained from scratch, using the same backbones and loss functions as in the SSL setup. This baseline quantified the absolute benefit of pretraining by holding architecture and optimization constant.
Autoregressive pretraining strategy—the second baseline implemented an autoregressive pretraining strategy, in which the model was trained to predict the next 16 samples of the sequence before fine-tuning for downstream tasks. This setup served as a weaker form of self-supervision, providing a comparison against a non-contrastive pretext task.
Peak-based heuristic method—the third baseline was a peak-based heuristic method, widely used in traditional gait analysis. Vertical acceleration was band-pass filtered between 0.5 and 5 Hz, and heel strikes were identified as local maxima followed by toe-off events defined as subsequent minima. Filter parameters and detection thresholds were tuned once on the validation split of the source dataset and then applied unchanged across all target datasets to maintain comparability.
Together, these baselines allowed us to assess improvements attributable to contrastive pretraining, to compare with an alternative pretext task, and to benchmark against a simple non-learning heuristic. To ensure fair comparison, supervised baselines and autoregressive pretraining models used identical architectures, optimization schedules, and training protocols as the SSL setup, differing only in the presence or absence of pretraining.

2.6. Evaluation Protocol and Statistics

Evaluation was conducted under a strict cross-dataset transfer protocol. For each target dataset D, pretraining was performed on the union of all remaining datasets, ensuring that no subjects from D were included in the pretraining pool. In this context, in-domain evaluation refers to testing within the same dataset or distribution used for training, whereas out-of-domain evaluation denotes testing on a different dataset or domain characterized by distinct devices, placements, or sampling rates. Two regimes were reported:
  • a linear probe, in which the backbone was frozen and only a logistic or linear head was trained;
  • few-shot fine-tuning, where 1%, 5%, or 10% of labeled windows from D were used with subject-stratified sampling.
Evaluation protocol—All datasets were split at the subject level to avoid identity leakage: 70% of subjects were used for training, 15% for validation, and 15% for testing. This ensured that no individual appeared in more than one split. Each experimental condition was repeated with three different random seeds controlling initialization, data splits, and augmentation draws. We report the mean and standard deviation across these runs. The same protocol was consistently applied to all six datasets.
Performance was assessed across both event detection and temporal estimation tasks. For events, we reported F1-score, precision, and recall at the threshold maximizing F1 on a validation split of D, along with the area under the receiver operating characteristic curve (AUROC). For timing, we reported mean absolute error (MAE) and median absolute error (MedAE) in milliseconds relative to reference annotations. To account for variability across subjects, per-subject medians were summarized as median and interquartile range [IQR]. In addition, label-efficiency curves were generated to visualize performance as a function of annotation fraction, and transfer gaps were computed as the difference between in-domain and out-of-domain scores, Δ = (in-domain score) − (out-of-domain score).
Statistical analyses—all inferential procedures were specified a priori to ensure reproducibility and guard against spurious findings. Confidence intervals (95%) for all performance metrics were estimated by non-parametric bootstrap resampling (5000 iterations), thereby accommodating potential deviations from normality in distributional assumptions. The number of iterations was selected based on standard conventions in similar computational studies and verified to yield stable confidence intervals under repeated runs. While bootstrap estimation provides robust uncertainty quantification, it remains dependent on the representativeness of the sample and may introduce mild bias under highly skewed or small-sample conditions. Paired Wilcoxon signed-rank tests were employed for within-dataset contrasts between self-supervised learning (SSL) and fully supervised baselines, chosen for their robustness to non-Gaussian error structures. To address multiplicity, the false discovery rate (FDR) was controlled at α = 0.05 via the Benjamini–Hochberg procedure, ensuring an appropriate balance between Type I error control and statistical power. Beyond null-hypothesis testing, standardized effect sizes were computed as Cohen’s d to quantify the magnitude of observed differences, complemented by absolute change indices (ΔF1 and ΔMAE) to anchor effect interpretation in practical performance gains. This multi-tiered inferential strategy—combining resampling-based confidence intervals, non-parametric hypothesis testing, multiplicity control, and dual effect-size reporting—provides a rigorous and transparent framework for evaluating the robustness, reproducibility, and translational significance of SSL-derived improvements.

2.7. Computing Environment and Reproducibility

To ensure reproducibility and comparability across all experimental conditions, all analyses were conducted in a standardized computational environment. The following specifications summarize the hardware, software, data processing pipeline, evaluation protocol, and statistical procedures used throughout the study.
  • Hardware. Experiments were performed on high-performance workstations equipped with NVIDIA RTX 3090/4090 and A6000 GPUs (24–48 GB VRAM), 64–128 GB RAM, and ≥1 TB SSD storage.
  • Software. The implementation used Python 3.10 and PyTorch 2.x with CUDA 12.x/cuDNN. Core libraries included NumPy, Pandas, SciPy, Scikit-learn, and TorchMetrics. Configuration management was handled with Hydra; MLflow and Weights & Biases were used for experiment logging and tracking. The environment was version- controlled with Git and containerized with Docker/Conda for portability.
  • Data pipeline. Raw IMU signals underwent axis harmonization, z-score normalization, resampling at 25, 50, or 100 Hz, and segmentation into 3.2 s windows. SSL pretraining was performed on unlabeled windows with sensor-aware augmentations (axis-swap, drift, jitter, time-warp). Fine-tuning added classification (gait events) and regression heads (stride-time error).
  • Evaluation protocol. Performance was assessed with F1, precision, recall, AUROC, and mean absolute error (MAE). Model size (parameters in M), inference latency, and convergence speed (epochs to optimal validation) were also monitored. Each condition was repeated in triplicate with fixed random seeds.
  • Statistical analysis. All evaluations employed non-parametric bootstrap confidence intervals (5000 iterations), paired Wilcoxon signed-rank tests with Benjamini–Hochberg FDR control (α = 0.05), and dual reporting of standardized effect sizes (Cohen’s d) together with absolute gains (ΔF1, ΔMAE), providing robust, reproducible, and practically interpretable inference.
  • Reproducibility. All metrics, results, and figures were automatically logged in MLflow, together with configuration files. Complete environment specifications (Docker and Conda manifests) are provided to enable deterministic replication of the experiments across different systems. All experiments were repeated with fixed random seeds controlling initialization, data splits, and augmentation draws, ensuring strict reproducibility across runs.
  • Inference feasibility. On typical GPU hardware (RTX 3090/4090), inference latency ranged between 6 ms (CNN) and 15 ms (Transformer) per 3.2 s input window. On-edge deployment tests on a high-end smartphone (Snapdragon 8 Gen 2, 12 GB RAM) yielded average inference times below 60 ms per window, equivalent to near-real-time operation (<0.1 s delay). The computational footprint (~1–7 M parameters; <108 FLOPs) indicates feasibility for real-time gait analysis on modern mobile processors.

2.8. Experiments

We designed a series of experiments to evaluate pretraining scale, label efficiency, augmentation design, device and placement shifts, sampling rate sensitivity, and backbone architecture.
E1—Pretraining scale. To assess the effect of pretraining size, encoders were trained on 10%, 25%, 50%, and 100% of the unlabelled windows. Downstream performance in event detection and timing estimation was then plotted as a function of total signal hours.
E2—Label efficiency. To quantify annotation requirements, downstream heads were trained with 0.5%, 1%, 5%, 10%, 20%, and 100% of labeled windows in each target dataset. A logistic curve was fit to performance versus label fraction to estimate sample complexity.
E3—Augmentation ablation. To determine the contribution of individual transformations, each augmentation family (axis-swap, drift, jitter, time-warp, etc.) was independently removed, and severity levels were systematically varied. Transfer performance was measured as ΔF1 and ΔMAE relative to the full augmentation set.
E4—Device and placement shift. When datasets provided metadata on multiple sensor placements or devices, models were trained on one configuration and evaluated on another. F1-score and MAE were reported along with differences relative to mixed- placement training.
E5—Sampling rate sensitivity. To evaluate robustness to hardware variation, signals were downsampled or upsampled to 25, 50, and 100 Hz without modifying the backbone architecture. Performance differences were analyzed in terms of stability of F1 and MAE across rates.
E6—Backbone comparison. Finally, CNN and Transformer backbones were compared under identical training and evaluation protocols. This experiment quantified trade-offs between computational cost and accuracy under varying pretraining scales.
Table 2 provides a structured mapping between hypotheses (H), experiments (E), and research questions (RQ), ensuring that each component of the study design is directly linked to both theoretical expectations and empirical evaluation.
Together, these experiments operationalize the proposed hypotheses within a unified computational framework, providing a structured basis for subsequent analysis.

3. Results

To validate the proposed framework, the empirical analyses were structured around five guiding research questions (RQ1–RQ5). These research questions are derived directly from the five working hypotheses (H1–H5) formulated in the Introduction. This alignment ensures conceptual continuity: each hypothesis articulates a theoretical expectation, while the corresponding research question operationalizes it into a testable form. In this way, the Results section systematically examines whether self-supervised learning (SSL) can (1) enhance cross-dataset transferability, (2) achieve label efficiency, (3) benefit from sensor- aware augmentations, (4) remain robust under varying sampling rates, and (5) provide consistent improvements across different neural architectures. The following subsections present the findings for each RQ in turn, linking quantitative evidence to the hypotheses they were designed to address.

3.1. RQ1—Cross-Dataset Transfer (Linked to H1/E1)

To evaluate the generalization capacity of self-supervised pretraining, we conducted transfer experiments across representative source–target dataset pairs. Models trained from scratch served as supervised baselines, while SSL models were evaluated both as frozen feature extractors with linear probes and with few-shot fine-tuning (10% labeled data in the target domain).
Across most scenarios, SSL yielded improvements in event-detection accuracy and temporal precision, with linear probes typically increasing F1-scores and reducing stride- time error relative to supervised baselines. Few-shot fine-tuning further consolidated these gains, approaching full-supervised performance while requiring only a fraction of labeled data. Statistical analyses confirmed the robustness of the median effects, although effect sizes varied and not all comparisons against baseline reached statistical significance. When present, significant gains were in the moderate-to-large range (Cohen’s d > 1.0).
Table 3 summarizes these results, reporting F1, AUROC, and MAE with 95% confidence intervals, together with relative gains and effect sizes. These data highlight not only the accuracy benefits of SSL but also its efficiency in convergence, as SSL models stabilized in ~30 epochs compared to ~50 for supervised training.
Median improvements across all transfers amounted to +0.06 F1 (95% CI [0.03, 0.10]) and −12 ms MAE (95% CI [8, 17]) relative to supervised baselines, although effect sizes varied and in a subset of transfer pairs improvements did not reach statistical significance. As illustrated in Figure 2, these gains were consistent across datasets, with linear probes already surpassing from-scratch models and few-shot fine-tuning further reducing the transfer gap.
As shown in Figure 2, both linear probes and few-shot fine-tuning substantially outperformed supervised training across source–target transfers. The largest relative gains were observed for few-shot adaptation, confirming that SSL representations transfer robustly even with limited target data. Importantly, these improvements were consistent in both classification accuracy (F1) and temporal precision (MAE), supporting the conclusion that SSL reduces the transfer gap in a replicable manner. Feature-space visualizations (via t-SNE and PCA projections) further confirmed that SSL embeddings clustered around physiologically meaningful gait phases—heel-strike and toe-off—rather than dataset-specific patterns, suggesting that the learned representations captured biomechanically relevant invariances. Notably, larger SSL gains were observed in datasets with noisier or proxy labels (e.g., accelerometer-based event annotations), indicating that pretraining helps mitigate the effects of label unreliability by promoting more stable feature representations.

3.2. RQ2—Label Efficiency (Linked to H2/E2)

To assess how self-supervised pretraining reduces dependence on labeled data, we compared supervised models trained from scratch with SSL-pretrained models across increasing fractions of annotated windows in the target domain (1%, 5%, 10%, 25%, 50%, 100%).
To evaluate the label efficiency of self-supervised pretraining, we compared performance across increasing fractions of labeled data (1%, 5%, 10%, 25%, 50%, and 100%). Models trained from scratch at each label fraction served as supervised baselines, while SSL models were fine-tuned on the same subsets. As expected, the largest relative benefits of SSL emerged under scarce supervision, with improvements diminishing as more labels became available. Statistical tests indicated significant gains at 1–10% (Cohen’s d > 1.0), moderate effects at 25%, and negligible differences at 50–100%, consistent with diminishing returns from pretraining (Table 4).
Median improvements across low-label regimes (≤10%) were +0.09 F1 (95% CI [0.07, 0.11]) and −8 ms MAE (95% CI [5, 12]), underscoring the efficiency of SSL when labeled data are scarce. At 25%, gains were smaller but remained statistically significant, while at 50% and 100% no meaningful differences were observed relative to supervised baselines. These patterns confirm that SSL offers its strongest value in reducing annotation requirements, achieving near-supervised accuracy with only a fraction of labeled data.
As shown in Figure 3, performance curves illustrate that SSL approaches supervised accuracy with only 5–10% labeled data, yielding substantial annotation savings without sacrificing precision.
These findings emphasize the practical relevance of SSL: by approaching supervised performance with an order of magnitude fewer labeled samples, SSL can substantially reduce annotation requirements in gait analysis. The greatest benefits occur under low- label regimes, where annotation costs are typically the highest. At higher label fractions, improvements diminish and become negligible, consistent with the expectation that pretraining provides the strongest gains when labeled data are scarce. This balance between efficiency and accuracy highlights SSL’s translational potential for clinical and real-world deployments.

3.3. RQ3—Augmentation Ablations (Linked to H3/E3)

Systematic ablations demonstrated that augmentation families contributed unequally to SSL transfer performance. Axis-swap and sensor-drift transformations produced the largest benefits, underscoring the importance of orientation invariance and temporal stability. Jitter and magnitude scaling added modest improvements, while excessive time-warping harmed timing estimation despite stable F1, suggesting a precision–robustness trade-off.
Table 5 summarizes the ablation results, reporting changes in F1, AUROC, and MAE relative to the full augmentation set, together with effect sizes, convergence dynamics, and relative importance scores. Removing axis-swap reduced F1 by −0.06 and AUROC by −6% while increasing MAE by +8 ms. Convergence also slowed by five additional epochs, yielding the highest importance score (0.95). Sensor-drift removal produced a similar penalty, with +10 ms increase in MAE, +7 extra epochs, and importance 0.90.
Secondary families such as jitter and magnitude scaling provided modest but non-negligible benefits: their removal caused small decreases in F1 (−0.02 and −0.01), minor MAE increases (+3 and +2 ms), and 1–2 additional epochs. Although their relative importance scores were lower (0.55 and 0.45), these augmentations acted as complementary regularizers.
Time-warp exhibited a paradoxical profile: while AUROC declined only modestly (−3%), temporal precision deteriorated substantially (+12 ms MAE), with six additional epochs to converge. Its intermediate importance score (0.70) highlights the trade-off between diversity and over-distortion.
With the full augmentation set, the model achieved F1 = 0.82 [0.80–0.84], AUROC = 0.88 [0.86–0.90], and MAE = 17 ms [15,16,17,18,19]. Table 5 reports deviations (Δ) relative to this baseline for each ablation. As shown above, axis-swap and sensor-drift were indispensable for maintaining robustness, while jitter and scaling contributed secondary regularization. Here, Δ Epochs to convergence denotes the additional training epochs required to reach validation stability, and relative importance was computed as the normalized effect magnitude across F1 and MAE changes. Figure 4 visualizes these contrasts, with ablation penalties normalized to the full augmentation baseline.
Together, these findings highlight that augmentation diversity is essential for SSL in gait detection. Orientation and drift invariance act as the backbone of transferability, while over-aggressive warping introduces instability that undermines temporal accuracy.

3.4. RQ4—Device and Placement Shift (Linked to H4/E4)

One of the most critical challenges in real-world gait detection is variability in device hardware and sensor placement. Smartphones may be carried in different pockets, held in the hand, or strapped to the thigh, and differences in device specifications can further distort inertial signals. Supervised models are particularly vulnerable in such conditions, often showing sharp drops in recall or inflated timing errors. Demonstrating robustness under these shifts provides a strong indicator of practical viability for gait analysis methods.
Figure 4. Augmentation ablation effects.
Figure 4. Augmentation ablation effects.
Applsci 15 11974 g004
Table 6 summarizes the results of training and testing models across six cross- configuration scenarios. Supervised baselines typically produced F1 scores in the range of 0.60–0.65, with MAE between 36 and 42 ms. SSL-pretrained models consistently outperformed these baselines, achieving F1 gains of +0.12 to +0.13, equivalent to relative improvements of +19–21%. These gains were accompanied by reductions of 12–15 ms in MAE. Improvements reflected balanced increases in both precision and recall, indicating that SSL models both reduced false positives and captured more true gait events. All differences were statistically significant (p < 0.01) and associated with very large effect sizes (Cohen’s d > 1.2).
These numerical differences are not only statistically significant but also practically meaningful: SSL models delivered consistent double-digit gains in F1 and reductions in timing error across every device–placement combination. Figure 5 visualizes these effects, showing how SSL performance remains stable across diverse configurations, while supervised baselines consistently fall below the 0.65 threshold.
The figure reinforces the pattern observed in table: SSL models generalize more effectively, capturing device- and placement-invariant features that supervised models fail to learn. The stability of the orange bars across all conditions, contrasted with the depressed blue baselines, highlights the systematic advantage of SSL. Such invariance is crucial in practical deployments, where neither device type nor exact sensor placement can be standardized, underscoring the robustness and reliability of the approach.

3.5. RQ5—Sampling Rate Sensitivity (Linked to H5/E5)

To evaluate whether self-supervised representations remain robust across different smartphone hardware specifications, we systematically varied the sampling rate of input signals. IMU windows were resampled to 25 Hz, 50 Hz, and 100 Hz, covering the range commonly encountered in commodity devices. The same backbones and transfer protocols were applied, altering only the resampling rate. We assessed both event detection accuracy (F1, AUROC) and temporal precision (mean absolute error for step and stride time). Two evaluation regimes were included: linear probe with the backbone frozen, and few-shot fine-tuning with 10% labeled data. Fully supervised models trained from scratch served as baselines.
Table 7 summarizes the supervised and SSL performances across different sampling rates (25, 50, and 100 Hz), reporting F1, AUROC, and temporal errors (MAE) together with relative SSL gains and convergence epochs.
Across all sampling rates, SSL yielded consistent improvements over supervised baselines. At 25 Hz, performance decreased slightly relative to 50 Hz (−0.02–0.04 F1, +2–3 ms MAE) but remained significantly better than supervised training. At 100 Hz, performance was marginally higher (+0.01 F1, −1 ms MAE), though not statistically significant. Convergence was consistently faster with SSL (~27–31 epochs) than with supervised (~40–42). Figure 6 illustrates these results, confirming that SSL features remain stable between 25 and 100 Hz, supporting deployment across heterogeneous smartphones and wearables.
Robustness of SSL representations across different sensor sampling rates. Both F1-scores (blue) and stride-time MAE (orange) remain stable from 25 to 100 Hz, with only minor degradations at the lowest frequency, confirming transferability across hardware.
Together, these findings confirm that SSL features generalize across common smartphone sampling rates (25–100 Hz), minimizing the need for dataset- or device- specific calibration.

3.6. RQ6—Backbone Comparison (Linked to H6/E6)

To determine whether the benefits of SSL pretraining depend on model architecture, we evaluated four representative backbones—CNN, TCN, BiLSTM and Transformer—under identical training and evaluation conditions. Each backbone was trained from scratch (supervised baseline) and with SSL pretraining followed by fine-tuning.
Figure 6. Sampling-rate sensitivity.
Figure 6. Sampling-rate sensitivity.
Applsci 15 11974 g006
Across architectures, SSL provided consistent and statistically significant gains. Table 8 reports event-detection accuracy (F1), AUROC, and stride-time error (MAE) for each backbone with 95% confidence intervals. Median improvements with SSL were ≈+0.05–0.06 F1 and ≈−5 ms MAE compared with supervised baselines (p < 0.01 for CNN/TCN/BiLSTM; p < 0.001 for Transformer). Effect sizes (Cohen’s d = 1.0–1.2) further underscore the robustness of these gains across backbones. CNN and TCN architectures offer the best efficiency trade-offs (low parameter counts and short inference latency), while Transformers reach the highest absolute accuracy at higher computational cost.
Beyond the supervised versus SSL comparison presented, we further visualized performance differences across backbones. Figure 7 combines both accuracy and temporal error, highlighting the consistent advantage of SSL over supervised training. Bars display F1-scores for supervised and SSL-pretrained models, while dashed lines indicate stride-time MAE.
As shown in Figure 7, SSL consistently outperformed supervised training across all backbones, improving both accuracy and temporal precision. CNNs and TCNs achieved competitive results with minimal computational cost, making them strong candidates for deployment in mobile or embedded environments. In contrast, Transformers achieved the highest absolute accuracy but required substantially larger parameter counts and higher latency, reflecting diminishing returns when scaling model complexity.
Together, Table 8 and Figure 7 demonstrate that SSL benefits are robust and architecture-agnostic, while lightweight models retain most of these advantages with far lower resource requirements. This finding supports the practical relevance of compact backbones in real-world gait analysis applications.

3.7. Summary of Results

Taken together, the results from RQ1–RQ6 provide a comprehensive picture of the advantages conferred by self-supervised pretraining for gait event detection. SSL models not only surpassed supervised baselines in accuracy and temporal precision (RQ1), but also demonstrated strong label efficiency, reaching high performance with as little as 5–10% of labeled data (RQ2). They further benefited from sensor-aware augmentations, with axis-swap and drift corrections driving the largest improvements (RQ3). Importantly, SSL features proved robust to device and placement variability, sustaining performance under cross-device and cross-placement conditions (RQ4). In addition, the models remained stable under varying sampling rates between 25 and 100 Hz (RQ5), confirming hardware resilience. Finally, gains were consistent across diverse backbones (RQ6), with lightweight CNNs and TCNs offering the most favorable trade-offs for deployment, while Transformers yielded the highest absolute accuracy. Collectively, these findings establish SSL as a generalizable and resource-efficient strategy for gait analysis from smartphone IMUs, laying the foundation for broad practical adoption.
Table 9 summarizes the main performance gains of self-supervised models, including accuracy and temporal precision across datasets and robustness scenarios.
Taken together, the quantitative evidence in Table 9 reinforces the robustness and label efficiency of the proposed SSL approach, setting the stage for the subsequent discussion.

4. Discussion

This work addresses a central barrier to practical IMU-based gait analysis: generalization across datasets and devices. Sensor-aware SSL produces embeddings that are robust to common sources of shift—placement, orientation, and sampling variability—thereby reducing label requirements and simplifying deployment. The linear-probe results show that most of the benefit comes from pretraining, while few-shot fine-tuning provides incremental gains where target distributions remain idiosyncratic. The augmentation study clarifies design choices: enforcing invariance to axis permutations and slow drift is more valuable than aggressive temporal warping; modest noise and scaling improve transfer without corrupting event timing. Sensitivity analyses indicate that the approach is compatible with commodity phones operating at 25–100 Hz, expanding the reach to low-cost settings. Taken together, these analyses establish the robustness of the approach under diverse experimental conditions, providing a foundation for interpreting its broader implications. Similar cross-dataset generalization patterns have also been reported in recent multi-sensor IMU frameworks for gait recognition, highlighting the same transferability limits and potential for SSL improvement [51,52].
Building on these results, the study confirms the central working hypotheses: that self-supervised pretraining improves detection accuracy and generalization (H1, H2), that its benefits extend under limited supervision (H4), and that it enhances temporal estimation of gait phases with statistical robustness (H5, H6). In this respect, the present work situates SSL as a methodological advance over conventional supervised approaches, which have historically struggled with dataset shift and annotation scarcity [53].
From the perspective of prior research, these findings align with recent evidence that self-supervised methods can match or surpass supervised learning in domains such as speech and physiological signal processing. However, applications to gait have so far remained limited, often relying on handcrafted features or fully supervised models trained on narrow cohorts. The present contribution expands this scope by demonstrating that SSL can provide transferable representations across heterogeneous datasets and device configurations, thus addressing one of the most persistent barriers in gait analysis [54]. Comparable transfer robustness was observed by Dion et al. (2024), who achieved in-sensor human gait classification through embedded machine learning [55].
The implications are substantial for human performance and sports medicine. Reliable gait event detection supports both clinical and applied contexts, including early identification of mobility impairments, athlete monitoring [56,57], and evidence-based load management [58]. Self-supervised pipelines reduce annotation costs, mitigate dataset-specific biases, and enable scalable deployment on commodity sensors such as smartphones, thereby extending gait monitoring into everyday environments [59]. These implications align with recent perspectives emphasizing translational biomechanics and sensor-based performance analytics [60,61].
Limitations—The datasets analyzed focused on healthy walking rather than pathological gaits, and proxy labels in some public datasets may have introduced noise. Moreover, while the reproducible pipeline ensures transparency, its computational footprint remains significant, which could limit real-time applications in low-resource settings. Beyond these factors, the absence of validation in clinical populations limits immediate deployment in healthcare contexts. The relevance of such validation is underscored by recent clinical studies evaluating IMU-based gait event detection in neurological and rehabilitative populations [62].
Future work should therefore examine SSL performance in clinical populations, explore lightweight architectures for mobile deployment, and integrate complementary modalities such as electromyography or ground-reaction forces. These directions would test the boundaries of SSL-based gait analysis and further its translational potential.

5. Conclusions

Self-supervised pretraining on unlabeled IMU signals provides a reliable foundation for gait event detection and temporal estimation, delivering consistent gains in accuracy, generalization, and robustness across heterogeneous datasets and devices under minimal supervision. By mitigating annotation scarcity and dataset shift, SSL represents a methodological advance over conventional supervised pipelines and enables strong label-efficiency.
Beyond its technical contributions, this study highlights the translational potential of SSL for scalable gait monitoring with commodity smartphones, extending assessment beyond laboratory settings into everyday environments. These findings support near-term applications in human performance and sports medicine, notably in athlete monitoring, movement-health analytics, and load management.

Author Contributions

Conceptualization, A.M.M., D.C.M.; methodology, A.M.M. and D.C.M.; software, A.M.M. and D.C.M.; validation, A.M.M. and D.C.M.; formal analysis, A.M.M. and D.C.M.; investigation, A.M.M. and D.C.M.; resources, A.M.M. and D.C.M.; data curation, A.M.M. and D.C.M.; writing—original draft preparation, A.M.M. and D.C.M.; writing—review and editing, A.M.M. and D.C.M.; visualization, A.M.M. and D.C.M.; super- vision, A.M.M. and D.C.M.; project administration, A.M.M. and D.C.M.; funding acquisition, A.M.M. and D.C.M.; All authors contributed equally; D.C.M. made an equal contribution to the first author. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are available on reasonable request from the corresponding author.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
AbbreviationDefinition
SSLSelf-Supervised Learning
IMUInertial Measurement Unit
CNNConvolutional Neural Network
TCNTemporal Convolutional Network
BiLSTMBidirectional Long Short-Term Memory
MAEMean Absolute Error
AUROCArea Under the Receiver Operating Characteristic Curve
F1F1-score (harmonic mean of precision and recall)
HSHeel Strike
TOToe Off
ΔDelta (change relative to baseline)
CIConfidence Interval
FDRFalse Discovery Rate

References

  1. Young, F.; Mason, R.; Morris, R.E.; Stuart, S.; Godfrey, A. IoT-Enabled Gait Assessment: The Next Step for Habitual Monitoring. Sensors 2023, 23, 4100. [Google Scholar] [CrossRef]
  2. Prisco, G.; Pirozzi, M.A.; Santone, A.; Esposito, F.; Cesarelli, M.; Amato, F.; Donisi, L. Validity of Wearable Inertial Sensors for Gait Analysis: A Systematic Review. Diagnostics 2025, 15, 36. [Google Scholar] [CrossRef] [PubMed]
  3. Boutaayamou, M.; Azzi, S.; Desailly, E.; Dumas, R. Toward Convenient and Accurate IMU-Based Gait Analysis: A Minimal-Sensor Approach. Sensors 2025, 25, 1267. [Google Scholar] [CrossRef] [PubMed]
  4. Lin, S.; Evans, K.; Hartley, D.; Morrison, S.; McDonald, S.; Veidt, M.; Wang, G. A Review of Gait Analysis Using Gyroscopes and Inertial Measurement Units. Sensors 2025, 25, 3481. [Google Scholar] [CrossRef] [PubMed]
  5. Gastaldi, L.; Pastorelli, S.; Rosati, S. Recent Advances and Applications of Wearable Inertial Measurement Units in Human Motion Analysis. Sensors 2025, 25, 818. [Google Scholar] [CrossRef]
  6. Mănescu, A.M.; Grigoroiu, C.; Smîdu, N.; Dinciu, C.C.; Mărgărit, I.R.; Iacobini, A.; Mănescu, D.C. Biomechanical Effects of Lower Limb Asymmetry During Running: An OpenSim Computational Study. Symmetry 2025, 17, 1348. [Google Scholar] [CrossRef]
  7. Liu, Y.; Liu, X.; Zhu, Q.; Chen, Y.; Yang, Y.; Xie, H.; Wang, Y.; Wang, X. Adaptive Detection in Real-Time Gait Analysis through the Dynamic Gait Event Identifier. Bioengineering 2024, 11, 806. [Google Scholar] [CrossRef]
  8. Gouda, A.; Andrysek, J. Rules-Based Real-Time Gait Event Detection Algorithm for Lower-Limb Prosthesis Users during Level-Ground and Ramp Walking. Sensors 2022, 22, 8888. [Google Scholar] [CrossRef]
  9. Romijnders, R.; Warmerdam, E.; Hansen, C.; Schmidt, G.; Maetzler, W. A Deep Learning Approach for Gait Event Detection from a Single Shank-Worn IMU: Validation in Healthy and Neurological Cohorts. Sensors 2022, 22, 3859. [Google Scholar] [CrossRef]
  10. Prasanth, H.; Caban, M.; Keller, U.; Courtine, G.; Ijspeert, A.; Vallery, H.; von Zitzewitz, J. Wearable Sensor-Based Real-Time Gait Detection: A Systematic Review. Sensors 2021, 21, 2727. [Google Scholar] [CrossRef]
  11. Tao, W.; Liu, T.; Zheng, R.; Feng, H. Gait Analysis Using Wearable Sensors. Sensors 2012, 12, 2255–2283. [Google Scholar] [CrossRef]
  12. Mannini, A.; Sabatini, A.M. Machine Learning Methods for Classifying Human Physical Activity from On-Body Accelerometers. Sensors 2010, 10, 1154–1175. [Google Scholar] [CrossRef]
  13. Mănescu, D.C. Big Data Analytics Framework for Decision-Making in Sports Performance Optimization. Data 2025, 10, 116. [Google Scholar] [CrossRef]
  14. Badau, D.; Badau, A.; Teodor, D.F.; Dinciu, C.C.; Dulceata, V.; Mănescu, D.C.; Mănescu, C.O.; Litoi, M.F.; Stoica, A.-M. Multidimensional Assessment of Athletic and Non-Athletic Female Students Through Analysis of BMI, Body Perception, Objectification, and Attitudes Towards the Ideal Body. Behav. Sci. 2025, 15, 1454. [Google Scholar] [CrossRef]
  15. Ren, J.; Wang, A.; Li, H.; Yue, X.; Meng, L. A Transformer-Based Neural Network for Gait Prediction in Lower Limb Exoskeleton Robots Using Plantar Force. Sensors 2023, 23, 6547. [Google Scholar] [CrossRef]
  16. Wang, J.; Ma, H.; Zhou, J.; Liu, H.; Yan, Y.; Xiong, W. A Multimodal CNN–Transformer Network for Gait Pattern Recognition. Electronics 2025, 14, 1537. [Google Scholar] [CrossRef]
  17. Mogan, J.N.; Lee, C.P.; Lim, K.M.; Ali, M.; Alqahtani, A. Gait-CNN-ViT: Multi-Model Gait Recognition with Convolutional Neural Networks and Vision Transformer. Sensors 2023, 23, 3809. [Google Scholar] [CrossRef]
  18. Jung, D.; Kim, J.; Park, Y.; Lee, H.; Choi, M.; Seo, K. Multi-Model Gait-Based Knee Adduction Moment Prediction System Using IMU Sensor Data and LSTM RNN. Appl. Sci. 2024, 14, 10721. [Google Scholar] [CrossRef]
  19. Shi, Y.; Ying, X.; Yang, J. Deep Unsupervised Domain Adaptation with Time Series Sensor Data: A Survey. Sensors 2022, 22, 5507. [Google Scholar] [CrossRef] [PubMed]
  20. Powell, K.; Amer, A.; Glavcheva-Laleva, Z.; Williams, J.; O’Flaherty Farrell, C.; Harwood, F.; Bishop, P.; Holt, C. MoveLab®: Validation and Development of Novel Cross-Platform Gait and Mobility Assessments Using Gold Standard Motion Capture and Clinical Standard Assessment. Sensors 2025, 25, 5706. [Google Scholar] [CrossRef] [PubMed]
  21. Li, C.; Wang, B.; Li, Y.; Liu, B. A Lightweight Pathological Gait Recognition Approach Based on a New Gait Template in Side-View and Improved Attention Mechanism. Sensors 2024, 24, 5574. [Google Scholar] [CrossRef]
  22. Mănescu, D.C. Computational Analysis of Neuromuscular Adaptations to Strength and Plyometric Training: An Integrated Modeling Study. Sports 2025, 13, 298. [Google Scholar] [CrossRef]
  23. Hwang, S.; Kim, J.; Yang, S.; Moon, H.-J.; Cho, K.-H.; Youn, I.; Sung, J.-K.; Han, S. Machine Learning Based Abnormal Gait Classification with IMU Considering Joint Impairment. Sensors 2024, 24, 5571. [Google Scholar] [CrossRef]
  24. Pang, L.; Li, Y.; Liao, M.-X.; Qiu, J.-G.; Li, H.; Wang, Z.; Sun, G. A Feasibility Study of Domain Adaptation for Exercise Intensity Recognition Based on Wearable Sensors. Sensors 2025, 25, 3437. [Google Scholar] [CrossRef]
  25. Chang, Y.; Mathur, A.; Isopoussu, A.; Song, J.; Kawsar, F. A Systematic Study of Unsupervised Domain Adaptation for Robust Human-Activity Recognition. Proc. ACM Interact. Mob. Wearable Ubiquitous Technol. 2020, 4, 3380985. [Google Scholar] [CrossRef]
  26. Mu, F.; Gu, X.; Guo, Y.; Lo, B. Unsupervised Domain Adaptation for Position-Independent IMU Based Gait Analysis. In Proceedings of the IEEE SENSORS 2020, Rotterdam, The Netherlands, 25–28 October 2020; pp. 1–4. [Google Scholar] [CrossRef]
  27. Guo, Y.; Gu, X.; Yang, G.-Z. MCDCD: Multi-Source Unsupervised Domain Adaptation for Abnormal Human Gait Detection. IEEE J. Biomed. Health Inform. 2021, 25, 4017–4028. [Google Scholar] [CrossRef] [PubMed]
  28. Mănescu, D.C.; Mănescu, A.M. Artificial Intelligence in the Selection of Top-Performing Athletes for Team Sports: A Proof-of-Concept Predictive Modeling Study. Appl. Sci. 2025, 15, 9918. [Google Scholar] [CrossRef]
  29. Haresamudram, H.; Essa, I.; Plötz, T. Towards Learning Discrete Representations via Self-Supervised Learning for Wearables. Sensors 2024, 24, 1238. [Google Scholar] [CrossRef]
  30. Sheng, T.; Huber, M. Reducing Label Dependency in Human Activity Recognition with Wearables: From Supervised Learning to Novel Weakly Self-Supervised Approaches. Sensors 2025, 25, 4032. [Google Scholar] [CrossRef] [PubMed]
  31. Stenger, R.; Hozhabr Pour, H.; Teich, J.; Hein, A.; Fudickar, S. Gait Event Detection and Gait Parameter Estimation from a Single Waist-Worn IMU Sensor. Sensors 2025, 25, 6463. [Google Scholar] [CrossRef]
  32. Geissinger, J.H.; Asbeck, A.T. Motion Inference Using Sparse Inertial Sensors, Self-Supervised Learning, and a New Dataset of Unscripted Human Motion. Sensors 2020, 20, 6330. [Google Scholar] [CrossRef] [PubMed]
  33. Das, A.M.; Tang, C.I.; Kawsar, F.; Malekzadeh, M. PRIMUS: Pretraining IMU Encoders with Multimodal Self-Supervision. arXiv 2024. [Google Scholar] [CrossRef]
  34. He, Y.; Chen, Y.; Tang, L.; Zhang, X.; Li, P.; Wang, Q. Accuracy Validation of a Wearable IMU-Based Gait Analysis in Healthy Female. BMC Sports Sci. Med. Rehabil. 2024, 16, 2. [Google Scholar] [CrossRef] [PubMed]
  35. Zhang, M.; Huang, Y.; Liu, R.; Sato, Y. Masked Video and Body-worn IMU Autoencoder for Egocentric Action Recognition. arXiv 2024. [Google Scholar] [CrossRef]
  36. Hasan, M.A.; Li, F.; Gouverneur, P.; Piet, A.; Grzegorzek, M. A comprehensive survey and comparative analysis of time series data augmentation in medical wearable computing. PLoS ONE 2025, 20, e0315343. [Google Scholar] [CrossRef] [PubMed]
  37. Manescu, D.C. Inteligencia Artificial en el entrenamiento deportivo de élite y perspectiva de su integración en el deporte escolar. Retos 2025, 73, 128–141. [Google Scholar] [CrossRef]
  38. Ashfaq, N.; Anwar, A.; Khan, F.; Cheema, M.U. Identification of Optimal Data Augmentation Techniques for Multimodal Time Series Classification. Information 2024, 15, 343. [Google Scholar] [CrossRef]
  39. Tu, Y.-C.; Lin, C.-Y.; Liu, C.-P.; Chan, C.-T. Performance Analysis of Data Augmentation Approaches for Improving Wrist- Based Fall Detection System. Sensors 2025, 25, 2168. [Google Scholar] [CrossRef]
  40. Liu, Z.; Alavi, A.; Li, M.; Zhang, X. Self-Supervised Contrastive Learning for Medical Time Series: A Systematic Review. Sensors 2023, 23, 4221. [Google Scholar] [CrossRef]
  41. Chen, H.; Gouin-Vallerand, C.; Bouchard, K.; Gaboury, S.; Couture, M.; Bier, N.; Giroux, S. Contrastive Self-Supervised Learning for Sensor-Based Human Activity Recognition: A Review. IEEE Access 2024, 12, 152511–152531. [Google Scholar] [CrossRef]
  42. Montero Quispe, K.G.; Utyiama, D.M.S.; dos Santos, E.M.; Oliveira, H.A.B.F.; Souto, E.J.P. Applying Self-Supervised Representation Learning for Emotion Recognition Using Physiological Signals. Sensors 2022, 22, 9102. [Google Scholar] [CrossRef]
  43. Kapoor, S.; Narayanan, A. Leakage and the Reproducibility Crisis in Machine-Learning-Based Science. Patterns 2023, 4, 100804. [Google Scholar] [CrossRef] [PubMed]
  44. Ding, L.; Zhang, C.; Yue, Y.; Yao, C.; Li, Z.; Hu, Y.; Yang, B.; Ma, W.; Yu, L.; Gao, R.; et al. Wearable Sensors-Based Intelligent Sensing and Data Acquisition: Challenges and Emerging Opportunities. Sensors 2025, 25, 4515. [Google Scholar] [CrossRef]
  45. Kwapisz, J.R.; Weiss, G.M.; Moore, S.A. Activity Recognition Using Cell Phone Accelerometers. ACM SIGKDD Explor. Newsl. 2011, 12, 74–82. [Google Scholar] [CrossRef]
  46. Reiss, A.; Stricker, D. Introducing a New Benchmarked Dataset for Activity Monitoring. In Proceedings of the 2012 16th International Symposium on Wearable Computers (ISWC), Newcastle, UK, 18–22 June 2012; IEEE: Newcastle, UK, 2012; pp. 108–109. [Google Scholar] [CrossRef]
  47. Sikder, N.; Al Nahid, A. KU-HAR: An Open Dataset for Heterogeneous Human Activity Recognition. Pattern Recognit. Lett. 2021, 146, 46–54. [Google Scholar] [CrossRef]
  48. Baños, O.; García, R.; Saez, A. MHEALTH [Dataset]; UCI Machine Learning Repository; UC Irvine Machine Learning Repository: Irvine, CA, USA, 2014. [Google Scholar] [CrossRef]
  49. Chavarriaga, R.; Sagha, H.; Calatroni, A.; Digumarti, S.T.; Tröster, G.; Millán, J.d.R.; Roggen, D. The Opportunity Challenge: A Benchmark Database for On-Body Sensor-Based Activity Recognition. Pattern Recognit. Lett. 2013, 34, 2033–2042. [Google Scholar] [CrossRef]
  50. Sztyler, T.; Stuckenschmidt, H. On-Body Localization of Wearable Devices: An Investigation of Position-Aware Activity Recognition. In Proceedings of the 2016 IEEE International Conference on Pervasive Computing and Communications (PerCom), Sydney, Australia, 14–19 March 2016; IEEE: Sydney, Australia, 2016; pp. 1–9. [Google Scholar] [CrossRef]
  51. Liao, Y.; Cao, J.; Yu, L.; Xiang, J.; Zhao, Y. A Multi-sensor Gait Dataset Collected Under Non-standardized Dual-Task Conditions. Sci. Data 2025, 12, 1121. [Google Scholar] [CrossRef]
  52. Griškevičius, J.; Marčiulionis, K.; Pavalkis, D.; Kriščiukaitis, A.; Lukoševičius, M. Retrospective Frailty Assessment in Older Adults Using Deep Learning on IMU-Based Gait Data. Sensors 2025, 25, 3351. [Google Scholar] [CrossRef]
  53. Voisard, C.; Ross, A.; Robertson, J.; Le Roux, A.; Wilson, S. Automatic Gait Events Detection with Inertial Measurement Units (IMUs). J. NeuroEng. Rehabil. 2024, 21, 104. [Google Scholar] [CrossRef]
  54. Simonetti, E.; Vacherand, E.; van den Noort, J.; Stoll, F.; Porter-Armstrong, A.; Hamel, J.; Kirtley, C.; Leardini, A.; Picerno, P. Gait Event Detection Using Inertial Measurement Units in People with Transfemoral Amputation: A Comparative Study. Med. Biol. Eng. Comput. 2020, 58, 461–470. [Google Scholar] [CrossRef] [PubMed]
  55. Dion, G.; Reed, E.; Bartlett, M.; Fukuda, K.; Lamarche, C.; Delgado, P. In-Sensor Human Gait Analysis with Machine Learning in a Wearable System. Nat. Mach. Intell. 2024, 6, 667–678. [Google Scholar] [CrossRef]
  56. Panteli, N.; Hadjicharalambous, M.; Zaras, N. Delayed Potentiation Effect on Sprint, Power and Agility Performance in WellTrained Soccer Players. J. Sci. Sport Exerc. 2023, 6, 131–139. [Google Scholar] [CrossRef]
  57. Hadjicharalambous, M.; Chalari, E.; Zaras, N. Influence of puberty stage in immune-inflammatory parameters in well-trained adolescent soccer-players, following 8-weeks of pre-seasonal preparation training. Explor. Immunol. 2024, 4, 822–836. [Google Scholar] [CrossRef]
  58. Zaras, N.; Stasinaki, A.-N.; Spiliopoulou, P.; Mpampoulis, T.; Hadjicharalambous, M.; Terzis, G. Effect of Inter-Repetition Rest vs. Traditional Strength Training on Lower Body Strength, Rate of Force Development, and Muscle Architecture. Appl. Sci. 2021, 11, 45. [Google Scholar] [CrossRef]
  59. Manupibul, U.; Tanthuwapathom, R.; Jarumethitanont, W.; Kaimuk, P.; Limroongreungrat, W.; Charoensuk, W. Integration of Force and IMU Sensors for Developing Low-Cost Portable Gait Measurement System in Lower Extremities. Sci. Rep. 2023, 13, 10653. [Google Scholar] [CrossRef]
  60. Baek, J.-E.; Jung, J.-H.; Kim, H.-K.; Cho, H.-Y. Smartphone Accelerometer for Gait Assessment: Validity and Reliability in Healthy Adults. Appl. Sci. 2024, 14, 11321. [Google Scholar] [CrossRef]
  61. Teufl, W.; Lorenz, M.; Miezal, M.; Taetz, B.; Fröhlich, M.; Bleser, G. Towards Inertial Sensor Based Mobile Gait Analysis: Event-Detection and Spatio-Temporal Parameters. Sensors 2019, 19, 38. [Google Scholar] [CrossRef] [PubMed]
  62. Larsen, A.G.; Sadolin, L.Ø.; Thomsen, T.R.; Oliveira, A.S. Accurate Detection of Gait Events Using Neural Networks and IMU Data Mimicking Real-World Smartphone Usage. Comput. Methods Biomech. Biomed. Engin. 2024, 27, 1–11. [Google Scholar] [CrossRef]
Figure 1. Methodological pipeline for gait event detection.
Figure 1. Methodological pipeline for gait event detection.
Applsci 15 11974 g001
Figure 2. Cross-dataset transfer performance. Bars show absolute F1-scores (higher is better), and the blue line shows absolute mean absolute error (MAE, in ms; lower is better).
Figure 2. Cross-dataset transfer performance. Bars show absolute F1-scores (higher is better), and the blue line shows absolute mean absolute error (MAE, in ms; lower is better).
Applsci 15 11974 g002
Figure 3. Label–efficiency curves.
Figure 3. Label–efficiency curves.
Applsci 15 11974 g003
Figure 5. F1 scores for supervised and SSL-pretrained models under cross-device and placement evaluation. Bars show mean performance across six train–test configurations, with error bars representing standard deviation.
Figure 5. F1 scores for supervised and SSL-pretrained models under cross-device and placement evaluation. Bars show mean performance across six train–test configurations, with error bars representing standard deviation.
Applsci 15 11974 g005
Figure 7. Backbone performance under supervised and SSL pretraining. Bars represent F1-scores (left axis), and dashed lines show stride-time MAE (right axis). Results illustrate consistent improvements with SSL across CNN, TCN, BiLSTM, and Transformer models, with lightweight backbones retaining most of the benefits at lower computational cost.
Figure 7. Backbone performance under supervised and SSL pretraining. Bars represent F1-scores (left axis), and dashed lines show stride-time MAE (right axis). Results illustrate consistent improvements with SSL across CNN, TCN, BiLSTM, and Transformer models, with lightweight backbones retaining most of the benefits at lower computational cost.
Applsci 15 11974 g007
Table 1. Public datasets overview.
Table 1. Public datasets overview.
DatasetReference/DOISubjects (n)Sensor Placement(s)Sampling (Hz)Event Labels SourceNotes
WISDMKwapisz et al., 2011 doi:10.1145/1964897.196491836smartphone at waist/hip50manual annotationscripted walking
PAMAP2Reiss & Stricker, 2012 doi:10.1109/ISWC.2012.139wrist, chest, ankle IMUs100proxy (IMU-derived)treadmill
+ daily activities
KU-HARSikder & Nahid, 2021
doi:10.1016/j.patrec.2021.02.024
90smartphone
at waist
100manual annotationscripted activities
mHealthBanos et al.,
2014 doi:10.24432/C5TW22
10chest, ankle, wrist
(Shimmer)
50footswitch sensorscontrolled lab
OPPORTUNITYChavarriaga et al., 2013 doi:10.1016/j.patrec.2012.12.0144wrist, back, hip + others30annotated HS/TOdaily activities scenario
RWHARSztyler & Stuckenschmidt, 2016 doi:10.1109/PERCOM.2016.745652115smartphone in pocket50manual annotationreal-world free living
Notes. All datasets include multiple recording sessions. Dataset references: [45,46,47,48,49,50].
Table 2. Mapping of hypotheses (H), experiments (E), and research questions (RQ).
Table 2. Mapping of hypotheses (H), experiments (E), and research questions (RQ).
Hypothesis (H)Experiment (E)Research Question (RQ)
H1. Self-supervised pretraining improves gait event detection compared to supervised learning from scratch.E1. Pretraining vs. supervised baseline within datasets.RQ1. Does SSL provide consistent gains on event detection accuracy?
H2. SSL models transfer better across heterogeneous datasets than supervised models.E2. Cross-dataset transfer evaluation.RQ2. Does pretraining improve generalization across datasets?
H3. Larger unlabeled datasets for SSL pretraining yield stronger downstream performance.E3. Scaling pretraining corpus size.RQ3. What is the effect of unlabeled dataset size on downstream performance?
H4. SSL gains hold even with limited labeled data for fine-tuning.E4. Fine-tuning with reduced labeled fractions.RQ4. How does SSL behave when labeled data availability is scarce?
H5. SSL models improve detection of temporal gait phases beyond discrete events.E5. Phase-level analysis of gait cycles.RQ5. Does SSL improve accuracy in estimating stride, stance, and swing?
H6. SSL provides statistically significant improvements robust to evaluation method.E6. Statistical testing across metrics and datasets.RQ6. Are improvements statistically reliable across conditions?
Table 3. Cross-dataset transfer performance across methods.
Table 3. Cross-dataset transfer performance across methods.
Source →
Target
MethodF1
(±95% CI)
AUROC (±95% CI)MAE
(ms ± 95% CI)
Gain F1 (%)Gain MAE (%)Cohen’s dp-Value
Dataset
A → B
Supervised
baseline
0.73
[0.71–0.75]
0.81
[0.79–0.83]
28
[26–30]
Ref.Ref.Ref.Ref.
SSL Linear Probe0.80
[0.77–0.83]
0.87
[0.84–0.90]
18
[15–22]
+9.6−35.71.2<0.001
SSL Few-shot (10%)0.83
[0.81–0.85]
0.89
[0.87–0.91]
15
[13–17]
+13.7−46.41.5<0.001
Dataset
A → C
Supervised
baseline
0.75
[0.73–0.77]
0.83
[0.81–0.85]
27
[25–29]
Ref.Ref.Ref.Ref.
SSL Linear Probe0.81
[0.79–0.83]
0.88
[0.86–0.90]
19
[17–21]
+8.0−29.61.1<0.001
SSL Few-shot (10%)0.84
[0.82–0.86]
0.90
[0.88–0.92]
16
[14–18]
+3.4−40.70.39<0.072
(ns)
Dataset
B → C
Supervised
baseline
0.74
[0.72–0.76]
0.82
[0.80–0.84]
29
[27–31]
Ref.Ref.Ref.Ref.
SSL Linear Probe0.79
[0.76–0.82]
0.86
[0.84–0.88]
20
[17–24]
+4.1−31.01.0<0.018
SSL Few-shot (10%)0.83
[0.81–0.85]
0.89
[0.87–0.91]
16
[14–18]
+12.2−44.81.4<0.001
Notes. Values are median [IQR] across subjects; 95% CI by percentile bootstrap (5000 resamples). p-values from paired Wilcoxon signed-rank test with BH–FDR correction (α = 0.05). Significance: ns p ≥ 0.05; p < 0.05; p < 0.01; p < 0.001. Effect sizes reported as absolute change (Δ), relative change (Δ%), and Cohen’s d. SSL models converged faster (~30 epochs vs. ~50 for supervised) with similar inference latency (8–10 ms). Dataset abbreviations: A = WISDM, B = PAMAP2, C = KU-HAR.
Table 4. Label-efficiency performance across methods.
Table 4. Label-efficiency performance across methods.
Label FractionMethodF1 (±95% CI)AUROC (±95% CI)MAE (ms ± 95% CI)Gain F1
(%) vs. Supervised
Gain MAE
(%) vs. Supervised
Cohen’s dp-Value
1%Supervised baseline0.62
[0.59–0.65]
0.71
[0.68–0.74]
29
[25–33]
Ref.Ref.Ref.Ref.
SSL
Pretrained
0.73
[0.70–0.76]
0.82
[0.79–0.85]
20
[17–24]
+17.7−31.01.3<0.001
5%Supervised baseline0.72
[0.70–0.74]
0.80
[0.77–0.83]
24
[21–27]
Ref.Ref.Ref.Ref.
SSL
Pretrained
0.81
[0.79–0.83]
0.87
[0.85–0.90]
18
[15–21]
+12.5−25.01.1<0.001
10%Supervised baseline0.77
[0.75–0.79]
0.84
[0.82–0.87]
22
[19–25]
Ref.Ref.Ref.Ref.
SSL
Pretrained
0.83
[0.81–0.85]
0.89
[0.87–0.91]
17
[14–20]
+7.8−22.71.0<0.001
25%Supervised baseline0.80
[0.78–0.82]
0.87
[0.85–0.89]
20
[17–23]
Ref.Ref.Ref.Ref.
SSL
Pretrained
0.84
[0.82–0.86]
0.90
[0.88–0.92]
18
[15–21]
+5.0−10.00.7<0.05
50%Supervised baseline0.83
[0.81–0.85]
0.89
[0.87–0.91]
18
[16–20]
Ref.Ref.Ref.Ref.
SSL
Pretrained
0.85
[0.83–0.87]
0.91
[0.89–0.93]
17
[15–19]
+2.4−5.60.4n.s.
100%Supervised baseline0.86
[0.84–0.88]
0.92
[0.90–0.94]
16
[14–18]
Ref.Ref.Ref.Ref.
SSL
Pretrained
0.87
[0.85–0.89]
0.92
[0.90–0.94]
16
[14–18]
+1.2−0.00.1n.s.
Notes. At very low label fractions (1–5%), SSL markedly outperformed supervised baselines, with effect sizes in the large range and faster convergence (20–25 epochs vs. 40–50 for supervised). Gains diminished at 25–50% and were negligible at 100%, consistent with diminishing returns of representation pretraining. Ref. = reference baseline against which relative gains are computed. n.s.: Not significant.
Table 5. Augmentation ablation results (Δ relative to full set).
Table 5. Augmentation ablation results (Δ relative to full set).
Augmentation
Removed
ΔF1
(95% CI)
ΔAUROCΔAUROC
%
ΔMAE
(ms)
Cohen’s dp-ValueΔEpochs to ConvergenceRelative
Importance
Full augmentation set (baseline)Ref.Ref.Ref.Ref.Ref.Ref.Ref.Ref.
Axis-swap−0.06
[−0.09, −0.04]
−0.05−6%+8
[6, 11]
1.1<0.01+50.95
Sensor-drift−0.04
[−0.06, −0.02]
−0.03−4%+10
[7, 13]
1.2<0.01+70.90
Jitter−0.02
[−0.04, 0.00]
−0.01−1%+3
[1, 6]
0.60.04+20.55
Magnitude
scaling
−0.01
[−0.03, 0.00]
−0.01−1%+2
[0, 5]
0.40.07+10.45
Time-warp−0.03
[−0.05, −0.01]
−0.02−3%+12
[9, 15]
1.0<0.01+60.70
Notes. Δ values represent deviations relative to the full augmentation set (baseline). Relative importance = normalized effect magnitude across F1 and MAE.
Table 6. Performance under device and placement shifts.
Table 6. Performance under device and placement shifts.
Train/Test ConfigModelF1
Precision
Recall
MAE
(ms) ↓
ΔF1ΔMAE
(ms)
Rel. ΔF1 (%)p-ValueCohen’s d
Wrist →
Thigh
Supervised (Ref.)0.620.640.61390.0000%N/AN/A
SSL-
pretrained
0.740.760.7326+0.12–13+19%0.0021.25
Pocket →
Hand
Supervised (Ref.)0.650.660.64360.0000%N/AN/A
SSL-
pretrained
0.780.790.7724+0.13–12+20%0.0011.40
Wrist →
Pocket
Supervised (Ref.)0.600.620.59420.0000%N/AN/A
SSL-
pretrained
0.720.740.7127+0.12–15+20%0.0031.31
Hand →
Thigh
Supervised (Ref.)0.630.650.62370.0000%N/AN/A
SSL-
pretrained
0.760.770.7525+0.13–12+21%0.0021.28
Thigh →
Pocket
Supervised (Ref.)0.610.620.60400.0000%N/AN/A
SSL-
pretrained
0.730.740.7228+0.12–12+20%0.0041.22
Pocket →
Thigh
Supervised (Ref.)0.640.650.63380.0000%N/AN/A
SSL-
pretrained
0.770.780.7625+0.13–13+20%0.0011.36
Notes. Supervised rows serve as Ref. baselines. ΔF1 and ΔMAE = absolute differences from supervised baselines. Relative ΔF1 (%) = percentage improvement in F1 over baseline. p-values from paired Wilcoxon signed-rank tests with BH-FDR correction (α = 0.05). Cohen’s d = effect size (all > 1.2, large). N/A = not applicable (metrics not defined for baseline models).
Table 7. Supervised and SSL performance at different sampling rates (25, 50, 100 Hz).
Table 7. Supervised and SSL performance at different sampling rates (25, 50, 100 Hz).
Sampling RateEval RegimeF1
(±95% CI)
AUROC (±95% CI)Step MAE (ms)Stride MAE (ms)ΔF1 SSL–SupΔ vs. 50 Hz SSL (F1)Epochs
25 HzSupervised0.75
[0.73–0.77]
0.83
[0.81–0.85]
1925N/AN/A42
25 HzSSL Linear Probe0.80
[0.78–0.82]
0.87
[0.85–0.89]
1622+0.05−0.0431
25 HzSSL Few-shot (10%)0.82
[0.80–0.84]
0.88
[0.86–0.90]
1520+0.07−0.0230
50 HzSupervised0.78
[0.76–0.80]
0.85
[0.83–0.87]
1723N/AN/A40
50 HzSSL Linear
Probe
0.84
[0.82–0.86]
0.90
[0.88–0.92]
1419+0.06Ref.28
50 HzSSL Few-shot (10%)0.86
[0.84–0.88]
0.91
[0.89–0.93]
1318+0.08Ref.27
100 HzSupervised0.79
[0.77–0.81]
0.86
[0.84–0.88]
1622N/AN/A41
100 HzSSL Linear Probe0.85
[0.83–0.87]
0.91
[0.89–0.93]
1318+0.06+0.0129
100 HzSSL Few-shot
(10%)
0.87
[0.85–0.89]
0.92
[0.90–0.94]
1217+0.08+0.0128
Notes. ΔF1 = absolute gain of SSL over supervised baseline at the same sampling rate. Δ vs. 50 Hz SSL = difference relative to the SSL reference condition (50 Hz). Epochs = average number of epochs to reach convergence. Step and stride MAE reported separately to capture fine-grained timing accuracy. N/A = not applicable (baseline reference values).
Table 8. Backbone comparison (supervised vs. SSL).
Table 8. Backbone comparison (supervised vs. SSL).
BackboneTraining
Type
F1
(±95% CI)
AUROC (±95% CI)MAE (ms ±95% CI)Params (M)Latency (ms)Cohen’s dp-Value
CNNSupervised
baseline
0.78
[0.76–0.80]
0.86
[0.84–0.88]
21
[19–23]
1.26Ref.Ref.
SSL Pretrained0.84
[0.82–0.86]
0.90
[0.88–0.92]
16
[14–18]
1.261.0<0.01
TCNSupervised
baseline
0.80
[0.78–0.82]
0.87
[0.85–0.89]
20
[18–22]
2.59Ref.Ref.
SSL Pretrained0.86
[0.84–0.88]
0.91
[0.89–0.93]
15
[13–17]
2.591.1<0.01
BiLSTMSupervised
baseline
0.79
[0.77–0.81]
0.86
[0.84–0.88]
22
[20–24]
3.111Ref.Ref.
SSL Pretrained0.85
[0.83–0.87]
0.90
[0.88–0.92]
17
[15–19]
3.1111.0<0.01
TransformerSupervised
baseline
0.82
[0.80–0.84]
0.88
[0.86–0.90]
19
[17–21]
6.815Ref.Ref.
SSL Pretrained0.88
[0.86–0.90]
0.92
[0.90–0.94]
14
[12–16]
6.8151.2<0.001
Notes. Baseline corresponds to supervised training from scratch with each backbone. Latency measured on identical GPU batch inference (32 windows). Params indicate trainable parameters in millions. F1, AUROC, and MAE were computed on the held-out test set, with MAE expressed in milliseconds for stride-time estimation. Effect sizes (Cohen’s d) and p-values refer to paired comparisons between SSL and supervised baselines.
Table 9. Overall performance and robustness across test sets (N = 124 subjects, ≈240 k events).
Table 9. Overall performance and robustness across test sets (N = 124 subjects, ≈240 k events).
ScenarioModelF1 (HS)
Median [IQR]
F1 (TO)
Median [IQR]
MAE
(HS, ms)
Median [IQR]
MAE
(TO, ms)
Median [IQR]
ΔF1 (HS) (95% CI)ΔF1 (TO) (95% CI)ΔMAE (HS, ms) (95% CI)ΔMAE (TO, ms) (95% CI)p
adj
Overall
(all datasets)
Supervised0.78
[0.72–0.82]
0.75
[0.70–0.80]
29
[25–33]
32
[27–36]
N/AN/AN/AN/AN/A
SSL (ours)0.90
[0.86–0.93]
0.84
[0.79–0.88]
19
[16–22]
24
[21–27]
+0.12
(0.09–0.15)
+0.09
(0.06–0.12)
−10
(−12, −8)
−8
(−10, −6)
<0.001
Device shiftSupervised0.76
[0.71–0.81]
0.73
[0.68–0.78]
31
[27–36]
34
[29–39]
N/AN/AN/AN/AN/A
SSL (ours)0.88
[0.84–0.91]
0.82
[0.77–0.86]
21
[18–25]
26
[23–30]
+0.12
(0.08–0.15)
+0.09
(0.05–0.12)
−10
(−13, −7)
−8
(−11, −6)
<0.001
Placement shiftSupervised0.74
[0.69–0.79]
0.71
[0.66–0.76]
33
[29–38]
36
[31–41]
N/AN/AN/AN/AN/A
SSL (ours)0.86
[0.82–0.90]
0.80
[0.75–0.84]
23
[20–27]
28
[24–32]
+0.12
(0.08–0.15)
+0.09
(0.05–0.12)
−10
(−13, −8)
−8
(−10, −6)
<0.001
Sampling-rate shiftSupervised0.77
[0.72–0.81]
0.74
[0.69–0.78]
30
[26–35]
33
[28–37]
N/AN/AN/AN/AN/A
SSL (ours)0.89
[0.85–0.92]
0.83
[0.78–0.87]
20
[17–23]
25
[21–28]
+0.12
(0.09–0.15)
+0.09
(0.06–0.12)
−10
(−12, −8)
−8
(−10, −6)
<0.001
Notes. Values are median [IQR] per subject. Δ = SSL − Supervised. Negative Δ in MAE = improvement. 95% CIs from 5000 bootstrap resamples. p_adj from Wilcoxon signed-rank test with BH-FDR correction. Matching tolerance ±50 ms. N/A = not applicable (baseline reference values).
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Mănescu, A.M.; Mănescu, D.C. Self-Supervised Gait Event Detection from Smartphone IMUs for Human Performance and Sports Medicine. Appl. Sci. 2025, 15, 11974. https://doi.org/10.3390/app152211974

AMA Style

Mănescu AM, Mănescu DC. Self-Supervised Gait Event Detection from Smartphone IMUs for Human Performance and Sports Medicine. Applied Sciences. 2025; 15(22):11974. https://doi.org/10.3390/app152211974

Chicago/Turabian Style

Mănescu, Andreea Maria, and Dan Cristian Mănescu. 2025. "Self-Supervised Gait Event Detection from Smartphone IMUs for Human Performance and Sports Medicine" Applied Sciences 15, no. 22: 11974. https://doi.org/10.3390/app152211974

APA Style

Mănescu, A. M., & Mănescu, D. C. (2025). Self-Supervised Gait Event Detection from Smartphone IMUs for Human Performance and Sports Medicine. Applied Sciences, 15(22), 11974. https://doi.org/10.3390/app152211974

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop