Next Article in Journal
Data-Dependent Weighted E-Value Aggregation for Fusion Learning
Previous Article in Journal
On an Inverse First-Passage Problem for Jump-Diffusion Processes
Previous Article in Special Issue
KAN-Former: 4D Trajectory Prediction for UAVs Based on Cross-Dimensional Attention and KAN Decomposition
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
This is an early access version, the complete PDF, HTML, and XML versions will be available soon.
Article

A Biomechanics-Guided and Time–Frequency Collaborative Deep Learning Framework for Parkinsonian Gait Severity Assessment

1
Department of Neurologic Surgery, Mayo Clinic Florida, Jacksonville, FL 32224, USA
2
Department of Neurosurgery, The 904th Hospital of the Joint Logistics Support Force of PLA, Wuxi 214044, China
3
College of Artificial Intelligence and Automation, Hohai University, Changzhou 213200, China
*
Author to whom correspondence should be addressed.
Mathematics 2026, 14(1), 89; https://doi.org/10.3390/math14010089 (registering DOI)
Submission received: 17 November 2025 / Revised: 14 December 2025 / Accepted: 18 December 2025 / Published: 26 December 2025

Abstract

Parkinson’s Disease (PD) is a neurodegenerative disorder in which gait abnormalities serve as key indicators of motor impairment and disease progression. Although wearable sensor-based gait analysis has advanced, existing methods still face challenges in modeling multi-sensor spatial relationships, extracting adaptive multi-scale temporal features, and effectively integrating time–frequency information. To address these issues, this paper proposes a multi-sensor gait neural network that integrates biomechanical priors with time–frequency collaborative learning for the automatic assessment of PD gait severity. The framework consists of three core modules: (1) BGS-GAT (Biomechanics-Guided Graph Attention Network), which constructs a sensor graph based on plantar anatomy and explicitly models inter-regional force dependencies via graph attention; (2) AMS-Inception1D (Adaptive Multi-Scale Inception-1D), which employs dilated convolutions and channel attention to extract multi-scale temporal features adaptively; and (3) TF-Branch (Time–Frequency Branch), which applies Real-valued Fast Fourier Transform (RFFT) and frequency-domain convolution to capture rhythmic and high-frequency components, enabling complementary time–frequency representation. Experiments on the PhysioNet multi-channel foot pressure dataset demonstrate that the proposed model achieves 0.930 in accuracy and 0.925 in F1-score for four-class severity classification, outperforming state-of-the-art deep learning models.
Keywords: PD; gait analysis; biomechanical prior; graph attention network; time–frequency collaborative modeling PD; gait analysis; biomechanical prior; graph attention network; time–frequency collaborative modeling

Share and Cite

MDPI and ACS Style

Lin, W.; Zhou, T.; Yang, Q. A Biomechanics-Guided and Time–Frequency Collaborative Deep Learning Framework for Parkinsonian Gait Severity Assessment. Mathematics 2026, 14, 89. https://doi.org/10.3390/math14010089

AMA Style

Lin W, Zhou T, Yang Q. A Biomechanics-Guided and Time–Frequency Collaborative Deep Learning Framework for Parkinsonian Gait Severity Assessment. Mathematics. 2026; 14(1):89. https://doi.org/10.3390/math14010089

Chicago/Turabian Style

Lin, Wei, Tianqi Zhou, and Qiwen Yang. 2026. "A Biomechanics-Guided and Time–Frequency Collaborative Deep Learning Framework for Parkinsonian Gait Severity Assessment" Mathematics 14, no. 1: 89. https://doi.org/10.3390/math14010089

APA Style

Lin, W., Zhou, T., & Yang, Q. (2026). A Biomechanics-Guided and Time–Frequency Collaborative Deep Learning Framework for Parkinsonian Gait Severity Assessment. Mathematics, 14(1), 89. https://doi.org/10.3390/math14010089

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Article metric data becomes available approximately 24 hours after publication online.
Back to TopTop