Next Article in Journal
Two-Way Communication Digital Power Controllers for Wireless Retinal Prosthesis Systems
Next Article in Special Issue
The Middleware for an Exoskeleton Assisting Upper Limb Movement
Previous Article in Journal
Design of a Slab Tamm Plasmon Resonator Coupled to a Multistrip Array Waveguide for the Mid Infrared
Previous Article in Special Issue
The Effect of a Horse-Riding Simulator with Virtual Reality on Gross Motor Function and Body Composition of Children with Cerebral Palsy: Preliminary Study
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Performance of Deep Learning Models in Forecasting Gait Trajectories of Children with Neurological Disorders

by
Rania Kolaghassi
*,
Mohamad Kenan Al-Hares
,
Gianluca Marcelli
and
Konstantinos Sirlantzis
School of Engineering, University of Kent, Canterbury CT2 7NT, UK
*
Author to whom correspondence should be addressed.
Sensors 2022, 22(8), 2969; https://doi.org/10.3390/s22082969
Submission received: 25 March 2022 / Revised: 6 April 2022 / Accepted: 11 April 2022 / Published: 13 April 2022
(This article belongs to the Special Issue Rehabilitation Robots and Sensors)

Abstract

:
Forecasted gait trajectories of children could be used as feedforward input to control lower limb robotic devices, such as exoskeletons and actuated orthotic devices (e.g., Powered Ankle Foot Orthosis—PAFO). Several studies have forecasted healthy gait trajectories, but, to the best of our knowledge, none have forecasted gait trajectories of children with pathological gait yet. These exhibit higher inter- and intra-subject variability compared to typically developing gait of healthy subjects. Pathological trajectories represent the typical gait patterns that rehabilitative exoskeletons and actuated orthoses would target. In this study, we implemented two deep learning models, a Long-Term Short Memory (LSTM) and a Convolutional Neural Network (CNN), to forecast hip, knee, and ankle trajectories in terms of corresponding Euler angles in the pitch, roll, and yaw form for children with neurological disorders, up to 200 ms in the future. The deep learning models implemented in our study are trained on data (available online) from children with neurological disorders collected by Gillette Children’s Speciality Healthcare over the years 1994–2017. The children’s ages range from 4 to 19 years old and the majority of them had cerebral palsy (73%), while the rest were a combination of neurological, developmental, orthopaedic, and genetic disorders (27%). Data were recorded with a motion capture system (VICON) with a sampling frequency of 120 Hz while walking for 15 m. We investigated a total of 35 combinations of input and output time-frames, with window sizes for input vectors ranging from 50–1000 ms, and output vectors from 8.33–200 ms. Results show that LSTMs outperform CNNs, and the gap in performance becomes greater the larger the input and output window sizes are. The maximum difference between the Mean Absolute Errors (MAEs) of the CNN and LSTM networks was 0.91 degrees. Results also show that the input size has no significant influence on mean prediction errors when the output window is 50 ms or smaller. For output window sizes greater than 50 ms, the larger the input window, the lower the error. Overall, we obtained MAEs ranging from 0.095–2.531 degrees for the LSTM network, and from 0.129–2.840 degrees for the CNN. This study establishes the feasibility of forecasting pathological gait trajectories of children which could be integrated with exoskeleton control systems and experimentally explores the characteristics of such intelligent systems under varying input and output window time-frames.

1. Introduction

Assistive devices such as wheelchairs and exoskeletons are designed to facilitate or enhance movement [1]. For people with impaired motor or nervous systems which affect their movement, these devices increase mobility and independence and allow for better interaction with the environment [1].
Exoskeletons are external load-bearing robotic suits made of sensors, actuators, and controllers to increase the strength and endurance of the user [2,3]. In the 1960s, after earlier conceptualisations, the first exoskeleton ’Hardiman’ was built by General Electric [2]. Since then, exoskeletons have witnessed continuous advancements in their structure and control [3].
Exoskeletons can be divided into two broad categories: passive and active. Passive exoskeletons are purely mechanical devices that provide torque to joints when contractible materials such as springs and dampeners expand, releasing the stored energy [4]. On the other hand, active exoskeletons contain actuators that provide torque to joints using an external power supply [4].
There are exoskeletons designed for the upper limbs, lower limbs, or the full body [5]. Lower limb exoskeletons are often used for three primary applications [6]. Firstly, they are used for the rehabilitation of the gait of patients with mobility disorders, the benefits of which extend to the therapists, relieving them from the strain of fully handling the patients themselves [6]; examples include Robotic Orthosis Lokomat by Hocoma [7], Ekso GT exoskeleton by Ekso Bionics [8], and ALEX by the University of Delware [9]. Secondly, exoskeletons can be used by patients with more severe mobility disorders that are partially or completely paralysed, outside rehabilitation centres [6]; they can assist with their daily locomotion, allowing them to stand/walk for longer periods. This has reportedly led to desirable physiological and psycho-somatic outcomes, including but not limited to improved spasticity, quality of life, bowel function, etc. [10]; the Vanderbilt exoskeleton is one example [11]. Thirdly, exoskeletons are used to augment the abilities of healthy users, allowing heavy loads to be carried with more ease or a movement to be performed with lower strain [6]; examples include Berkeley Lower Extremity Exoskeleton (BLEEX) [12] and Hybrid Assistive Limb (HAL) exoskeleton [13].
Powered exoskeletons need a control strategy to guide the exoskeleton’s interaction with the user. The control strategy coordinates the exoskeleton’s movement with the user’s body when full support is provided, while it synchronises the exoskeleton’s movement with the user’s body when partial support is provided [14]. The control strategy often consists of a three-level hierarchy: high, mid, and low levels [14]. The high level of control is responsible for intention detection, predicting what state the user would like to be in and the overall behaviour of the exoskeleton [14,15,16]. This includes estimating desired torque from EMG signals [17], classifying locomotion modes (e.g., standing up, sitting down, walking up/down a staircase, etc.) [18] and their transitions, as well as environment classification (predicting the user’s interaction with the surrounding environment [19]). The mid-level control is responsible for selecting one of the continuous states of the exoskeleton and switching between them [14,15]. Baud et al. [14], in their review on control strategies for lower limb exoskeletons, divide this level into two sub-levels: detection/synchronisation and action. Detection/synchronisation involves identifying the state of the user, such as the phase of gait, while action is responsible for computing the output of the identified state. Lastly, the low-level control directly controls the actuators and is responsible for implementing the control strategy to supply torque to the joints. It tracks the reference input and ensures stability [15,20,21].
The timing and magnitude of the supportive torque provided depend on the control strategy implemented, which is highly influenced by the exoskeleton’s application. In rehabilitation applications, an exoskeleton may follow a predefined trajectory. This is referred to as trajectory tracking and the trajectory tracked can be the trajectory of a healthy user [6]. Another control strategy used in exoskeletons for rehabilitation is assist as needed, where the amount of support provided by the exoskeletons is variable and may change through the course of rehabilitation. Impedance control is an example of an assist as needed strategy, where the assistance provided depends on the effort of the patient [6]. In locomotion assistance applications, trajectory tracking is also most commonly used. The trajectory tracked could be a predefined trajectory for a healthy user, or in the case of hemiparetic patients, the trajectory used for the pathological limb can be the trajectory of the healthy limb (also known as complementary limb motion estimation [22]). As for strength augmentation applications, hybrid and force control are commonly used [6].
Gait trajectory prediction can be integrated into the control strategy of exoskeletons [23]. Several researchers investigated the use of deep learning for gait trajectory prediction. Liu et al. [24] developed a deep spatio-temporal model that consists of Long Short-Term Memory (LSTM) units to predict two time-steps in the future, with predictions averaged to smooth fluctuations. Zaroug et al. [25] implemented an auto-encoder LSTM for predicting linear acceleration and angular velocity trajectories. They experimented with varying lengths of input time-steps, between five and 40 steps, to predict five or 10 steps in the future (equivalent to 30 ms or 60 ms). An LSTM with a weighted discount loss function was proposed by Su et al. [26] for the prediction of the angular velocities of thigh, shank, and foot segments. They used 10 or 30 time-steps as input to predict five or 10 steps in the future, corresponding to 100 ms and 200 ms, respectively. Hernandez et al. [27] used a hybrid Convolutional Neural Network (CNN) and LSTM neural network, DeepConvLSTM, to forecast kinematic trajectories with an average MAE of 3.6 , while Jia et al. [28] implemented a deep neural network with LSTM units and a feature fusion layer that combines kinematic (i.e., joint angles) and physiological (i.e., EMG) data for trajectory prediction. Zarough et al. [29] also compared between vanilla, stacked, bidirectional, and autoencoder LSTMs while Zhu et al. [30] used attention-based CNN-LSTM, predicting trajectories 60 ms in the future.
Values of kinematic parameters within a gait cycle vary more significantly between different individuals (inter-subject) compared to the values of kinematic parameters of the same individual (intra-subject) [31]. Intra-subject trajectory prediction models (models tested on data from the same individuals used for training the models) have been shown to be more accurate in their predictions than inter-subject models (models tested on data from individuals who were not used for training the models) [26]. However, the variability of pathological gait compared to healthy gait is even greater. For example, children with spastic cerebral palsy were found to have higher within-day and between-day variability compared to healthy children, possibly attributed to the spasticity that limits the range of motion of their joints [32,33].
Given that existing studies have used models trained on healthy gait trajectories only, the ability of deep learning models to forecast pathological trajectories with greater heterogeneity and variability is yet to be evaluated. The main contribution of this study is to investigate, for the first time, the performance of deep learning networks, specifically the Long Short-Term Memory (LSTM) neural network and Convolutional Neural Network (CNN), in forecasting pathological gait trajectories of children with neurological disorders. We conduct a comparison between the two networks. Furthermore, we investigate the influence of the length of the input and output windows on prediction accuracy and provide technical recommendations.

2. Materials and Methods

2.1. Data

The deep learning models implemented in our study are trained on data from children with neurological disorders. The data used are available online and were collected by Gillette Children’s Speciality Healthcare over the years 1994–2017 [34]. They have been previously used for the development of a deep learning model to automatically detect gait events, specifically foot contact and foot off events [35]. The children of the dataset ranged from 4 to 19 years of age. The majority of them had cerebral palsy (73%), while the rest had a combination of neurological, developmental, orthopaedic, and genetic disorders (27%). Children were recorded with a motion capture system (VICON) with a sampling frequency of 120 Hz while walking for 15 m. The data consist of a 99-dimensional vector with kinematics and marker positions. The data were divided into a training and testing set. The statistical distribution of the features of the children in the training set is as follows: age (11.4 ± 6.2 years), weight (35.7 ± 17.7 kg), height (135.7 ± 21.6 cm), leg length (70.3 ± 14.0 cm), and walking speed (0.84 ± 0.28 m/s). The statistical distribution of the children in the testing set is as follows: age (11.0 ± 4.5 years), weight (35.9 ± 16.7 kg), height (135.6 ± 21.4 cm), leg length (70.6 ± 12.8 cm), and walking speed (0.85 ± 0.29 m/s) [35]. In our study, only kinematics were used for trajectory forecasting; these include angles of the hip, knee, and ankle in the yaw, pitch, and roll dimensions. The kinematics represent Euler angles, calculated using the plug-in-gait mechanical model [35].

2.2. Data Processing

Euler angles of the hip, knee, and ankle in yaw, pitch, and roll dimensions were available for both legs, however, only one leg was used for training the models. The pre-processing of data involved trimming the leading and the trailing zeros, and the removal of trails with spurious data (which were assumed to be Euler angles greater or less than a 90 cut-off).
Since deep learning models are required to be trained with fixed-length sequences, fixed-length inputs and their corresponding target trajectories were generated using the sliding window method, illustrated in Figure 1. The sliding window method involves generating a shorter sequence x i n with k time-steps, where k is the size of the input, i.e., the number of time-steps utilised by the model to make predictions, x i n = { x 1 , x 2 , , x k } . Another shorter sequence y o u t with z time-steps is created, were z is the size of the output, i.e., the number of time-steps that will be predicted by the model, y o u t = { x k + 1 , x k + 2 , , x k + z } . Each input window has a corresponding output window (target label to train the models), and the two windows represent one training sample. The stride, which is the distance between the beginning of one sample and the beginning of the next sample, is set to 5 time-steps. Ten samples are generated from each trial.
The models were trained using several sizes of input and output windows. The input window sizes for the LSTM were 50, 100, 200, 400, 600, 800, and 1000 ms. Given that the data were captured at a sampling frequency of 120Hz, those durations correspond to 6, 12, 24, 48, 72, 96, and 120 time-steps. Output window sizes for the LSTM were 8.33, 25, 50, 100, and 200 ms which correspond to 1, 3, 6, 12, and 24 time-steps. Every combination of input and output window sizes was used, yielding a total of 35 combinations. As for the CNN, we only focused on using 6 and 120 input time-steps (the smallest and largest input sizes we used with the LSTM) to predict 1, 3, 6, 12, and 24 output time-steps. We used these time ranges following values proposed in the literature by researchers that forecasted healthy gait. Zaroug et al. experimented with a wide range of input windows from 5–40 steps to forecast 5 or 10 steps in the future, which correspond to 30 and 60 ms respectively [25]. Output windows ranging from 50–60 ms were common too [25,28,30]. A few researchers have also predicted output windows of 100 ms or larger [26,29].
For n training samples, f features (hip, knee, and ankle angles in the yaw, pitch, and roll directions), l i n input window size and l o u t output window size, an input matrix X and target output matrix Y are created, where X R n × l i n × f and Y R n × l o u t × f . The matrices are then normalised such that X [ 0 , 1 ] and Y [ 0 , 1 ] . The aim is to build a model g ( ) that maps X to Y ^ , i.e., Y ^ = g ( X ) , where Y ^ is a close approximation to true value Y.

2.3. Long Short-Term Memory (LSTM) Architecture

The Long Short-Term Memory (LSTM) neural network is commonly used for time-series applications including forecasting future trajectories. It is a type of gated Recurrent Neural Network (RNN), which solved the issue of vanishing and exploding gradients during training and is capable of learning long-term dependencies [36]. Each LSTM cell contains three gates: input, output, and forget gate. The equations of these gates, as explained in Goodfellow et al. [36], are reported below.
The forget gate unit f i ( t ) equation for time-step t, cell i, input vector x ( t ) , hidden layer vector h ( t ) , biases b f , recurrent weights W f , and input weights U f is:
f i ( t ) = σ b i f + j U i , j f x j ( t ) + j W i , j f h j ( t 1 ) .
The internal state of an LSTM cell, s i ( t ) , is updated depending on the value of the forget gate unit f i ( t ) , and its equation for biases b, input weights U, and recurrent weights W is:
s i ( t ) = f i ( t ) s i ( t 1 ) + g i ( t ) σ b i + j U i , j x j ( t ) + j W i , j h j ( t 1 ) .
The external input gate unit, g i ( t ) , is calculated as:
g i ( t ) = σ b i g + j U i , j g x j ( t ) + j W i , j g h j ( t 1 ) .
The output of the LSTM cell h i ( t ) is calculated as:
h i ( t ) = t a n h s i ( t ) q i ( t ) .
For biases b o , input weights U o , and recurrent weights W o , the value of the output gate unit is:
q i ( t ) = σ b i o , + j U i , j o x j ( t ) + j W i , j o h j ( t 1 ) .
The LSTM network implemented in this study contained 4 layers of LSTM units, with 128 units per layer. The hidden state of the final layer is used as input to a fully connected layer which is then reshaped to obtain the desired output. The overall architecture is illustrated in Figure 2.

2.4. Convolutional Neural Network (CNN) Architecture

In this study, a Convolutional Neural Network (CNN) is used to map input trajectories X to forecasted predictions Y ^ . The convolution operation is formed between a sequence and a kernel, whose weights are tuned during the learning process. Equation (6) is used to calculate the output of the convolution operation S, adapted from the format included in Goodfellow et al. [36] to work with a 1D time-series I and kernel K.
S ( i ) = ( I K ) ( i ) = m I ( m ) K ( i m ) .
The CNN architecture consists of several 1D convolutions and pooling layers, followed by a dense fully connected layer. The number of kernels in each convolution layer, as well as the size of the pooling layers, is illustrated in Figure 3. Dilated convolutions were tried, but they did not improve performance, therefore they were not included.

2.5. Baseline Methods

The performance of the deep learning models was benchmarked against a simpler machine learning model, a Fully Connected Network (FCN), and two non-intelligent models, which we refer to as Naïve Method 1 and Naïve Method 2. The FCN contains three hidden layers and 200 nodes per layer. As for the non-intelligent models, the first naïve method uses the final time-step in the input window as the predicted value for all output time-steps. The second naïve method uses the mean value of the input as the predicted value for all output time-steps.

2.6. Details of Network Implementation

The objective of the deep learning models is to find a mapping between the input trajectories X and forecasted trajectories Y ^ , such that the error between predicted trajectories Y ^ and true trajectories Y is minimised. The loss function that has been used to optimise the deep learning models is the Mean Squared Error (MSE).
The number of trials used from the dataset was 16,782. Each trial was used to extract 10 samples using the sliding window method, described in Section 2.2, resulting in a total of 167,820 samples. The samples were split into training (70%), validation (20%), and testing (10%) sets. Both the CNN and LSTM are trained in mini-batches, where the size of each batch is 32 samples, with the Adam optimiser used for learning.
To select the hyper-parameters and architecture for the CNN, LSTM, and FCN models (described in Section 2.3, Section 2.4 and Section 2.5), a hyper-parameter search was performed. Hyper-parameter optimisation involved defining a search space, where windows of values of some parameters of the model were chosen (e.g., learning rate, number of layers of LSTM, number of hidden units, and size of kernels). The objective is to choose the value of parameters that optimises the performance of the networks. The tree-structured Parzen estimator algorithm, which is a type of Bayesian hyper-parameter sampler [37], was used to select the optimal parameters (from a defined search space) that minimise the validation loss. The search space for the hyper-parameters and the corresponding selected values are included in Table 1. The parameters were optimised for predictions with an input window size of 72 time-steps, and an output window size of 12 time-steps. The number of epochs was also a parameter that was included in the hyper-parameter optimisation, but the values were fine-tuned manually afterwards. The numbers of epochs for training the LSTM, CNN, and FCN were 60, 150, and 180, respectively. Once the optimal architecture and parameters of the networks were selected, the training and validation sets were combined to train these networks. The total numbers of trainable parameters of the optimised LSTM, CNN, and FCN are 495,320, 4,666,888, and 380,216, respectively. The final performance of the networks was evaluated on the testing set.
The framework described has been implemented using Python, with the following libraries: Pytorch, Numpy, Matplotlib, SciPy, and Scikit-learn. Optuna was used for the hyper-parameter search [37]. An Nvidia Geforce RTX 2070 GPU was also used for computation.

2.7. Evaluation Metrics

Several metrics have been used to evaluate the performance of the models. To measure how close the predicted trajectories Y ^ are to the observed trajectories Y we have calculated the mean square error MSE, mean absolute error MAE, and Pearson correlation coefficient P. These were calculated after the de-normalisation (i.e., re-scaling to the original ranges) of the predicted trajectories. Given that n is the number of testing samples, f is the number of features, and l o u t is the output window size, the equations are as follows:
Mean squared error:
M S E = 1 n × f × l o u t i = 1 n j = 1 f k = 1 l o u t y i , j , k y ^ i , j , k 2 .
Mean squared error standard deviation:
σ M S E = 1 n × f × l o u t i = 1 n j = 1 f k = 1 l o u t y i , j , k y ^ i , j , k 2 M S E .
Mean absolute error:
M A E = 1 n × f × l o u t i = 1 n j = 1 f k = 1 l o u t | y i , j , k y ^ i , j , k | .
Mean absolute error standard deviation:
σ M A E = 1 n × f × l o u t i = 1 n j = 1 f k = 1 l o u t | y i , j , k y ^ i , j , k | M A E .
Pearson correlation coefficient:
P = 1 f j = 1 f c o v ( y j , y ^ j ) s t d ( y j ) × s t d ( y ^ j ) .
These metrics are used to evaluate and compare the performance of the networks we implemented, with results presented in Section 3.

3. Results

3.1. LSTM Network Performance for Varying Input and Output Window Sizes

The LSTM model has been trained using 35 combinations of input and output window sizes. The input window sizes were 6, 12, 24, 48, 72, 96, and 120 time-steps, which correspond to 50, 100, 200, 400, 600, 800, and 1000 ms, respectively, while the output window sizes were 1, 3, 6, 12, and 24 time-steps, and correspond to 8.33, 25, 50, 100, and 200 ms. The performance of the models (MAE, MSE, and Pearson correlation coefficient) are reported in Table 2.
Results in Table 2 show that the smaller the size of the output window, the lower the error of predictions. LSTMs predicting one output time-step had the lowest mean errors, while LSTMs predicting 24 time-steps had the highest mean errors, as expected. On the other hand, the size of the input window did not significantly influence the mean losses when predicting short output windows, specifically six output time-steps or below (see Figure 4a–c). The size of the input window had an influence when predicting larger output windows, i.e., 12 and 24 time-steps, with larger input sizes resulting in lower mean errors (see Figure 4d,e). For 12 and 24 time-step output windows, using 120 time-steps as input window size, which is the largest input size we tried, led to the lowest mean absolute errors (see Figure 4d,e).

3.2. Performance of the CNN and Comparisons with LSTM Network

We trained the CNN and compared its performance with the LSTM network. We only focused on using six and 120 input time-steps (the smallest and largest input sizes we used with the LSTM) to predict 1, 3, 6, 12, and 24 output time-steps. The performances of both models (MAE, MSE, and Pearson correlation coefficient) are reported in Table 3.
As shown in Table 3, the CNN errors increase when increasing the number of predicted time-steps, as observed with LSTM. In Figure 5, we compare the performance of the CNN and LSTM and it is noticeable that their mean errors are very similar when predicting small output windows, such as one and three time-steps. However, the difference in the MAE for CNN and LSTM widens for larger output windows including 6, 12, and 24 future time-steps, with LSTM outperforming CNN. Interestingly, the size of this difference depends on the input window size: when six time-steps are used as input, the CNN MAE is larger than the LSTM one; when 120 time-steps are used as input, the difference in their MAEs is even larger.

3.3. Benchmarking Performance of Deep Learning Models

We benchmark the performance of our two deep learning models to a simpler machine learning architecture, the Fully Connected Network (FCN), and two naïve/non-intelligent methods. The first naïve method uses the final time-step in the input window as the predicted value for all output time-steps. The second naïve method uses the mean value of the input as the predicted value for all output time-steps. Table 4 shows the MAE for the deep learning and benchmark models. Our deep learning models outperformed all naïve methods and the majority of the FCN predictions. In two cases, the FCN obtained better results: the first was when the input and output window sizes were six and 24 time-steps, respectively, where the FCN had lower MAE compared to the LSTM and CNN; the other case was when the input and output sizes were 120 and 24 time-steps, respectively, with FCN performing better than CNN, but not better than LSTM.
Naïve Method 1 resulted in lower MAEs compared to Naïve Method 2, therefore, pairwise t-tests were conducted between the Naïve Method 1 and all other intelligent methods (LSTM, CNN, and FCN). This was done to determine whether the mean absolute errors were significantly different (with p < 0.05); the results in Table 4 confirm that intelligent models perform better than non-intelligent models. Furthermore, pairwise t-tests were conducted between the LSTM and all the other models; the differences in the MAEs were found to be statistically significant, as shown in Table 4.

3.4. Accuracy of the Models across the Different Time-Steps

In the previous sections, we were reporting the mean error across all features and time-steps. Here, we calculate the MAE, for each time-step for a given output window, separately (see Figure 6); the MAE is calculated using an adapted form of Equation (9) (see Section 2.7), where the summation over k = 1: l o u t is not performed. As expected, results show that predictions further in the future deviate more from the actual values, and this deviation is more pronounced after around the 3rd time-step, as shown in Figure 7. Figure 7 also shows that the LSTM MAE increases with the increase in the size of the output window.

3.5. Performance of the Models for Each Joint

We investigated whether errors for a particular joint were higher than others. The MAE results for each of the hip, knee, and ankle joints are presented in Figure 8. The MAE for each joint represents the combined errors for the angles predicted in the pitch, roll, and yaw dimensions. They were calculated using an adapted form of Equation (9) (see Section 2.7), where the numbers of features, f, in the summation over j = 1:f are reduced to the pitch, yaw, and roll angles for a single joint, rather than for all joints. The results do not show any particular trend.

4. Discussion

In this paper, we implement deep learning models, specifically LSTM and CNN, to forecast trajectories of children with pathological gait. To the best of our knowledge, this is the first time trajectories of pathological gait of children, which exhibit larger inter- and intra-subject variability compared to the trajectories of typically developing children, are predicted using deep learning models.
The advantage of deep learning models is that they make predictions based on current input data, but also utilise knowledge of learned representations of gait trajectories acquired during a prior learning stage from numerous gait sequences. We used LSTM and CNN to forecast hip, knee, and ankle trajectories based on varying input and output window sizes. Input window sizes for the LSTM were 50, 100, 200, 400, 600, 800, and 1000 ms (for data captured at a sampling frequency of 120Hz, these durations correspond to 6, 12, 24, 48, 72, 96, and 120 time-steps). Input window sizes for the CNN were 50 and 1000 ms (corresponding to six and 120 time-steps). The reason for using input window sizes up to 1000 ms is that the average length of a gait cycle for a typically developing school-aged child is 980–990 ms [38]. This means we trained deep learning models to make predictions based on data from approximately one full gait cycle, or lower. Output window sizes for the LSTM and CNN were 8.33, 25, 50, 100, and 200 ms (corresponding to 1, 3, 6, 12, and 24 time-steps); this means that we have tried to forecast up to 20% of the cycle. We used these time ranges following values proposed in the literature by researchers that forecasted healthy gait (refer to Section 2.2 for details).
The LSTM (a type of gated recurrent network) was selected for forecasting gait trajectories as it has been commonly and successfully used with sequential data [36]. The LSTM has the advantage of taking into account the order of values in an input sequence. It has the ability to learn long-term dependencies [36]. The LSTM network was compared to a CNN, which is mostly used for computer vision problems with 2D grid-like topology inputs using 2D convolutions, but it is increasingly used with time-series sequences using 1D convolutions [39]. Therefore, we implemented a CNN to evaluate if it shows promising performance in the task of forecasting trajectories for children with neurological disorders.
Our results show that the LSTM’s performance is better than the CNN. However, the difference between the MAE of the two networks was larger with larger input and output window sizes. The performance gap, measured in MAE, was highest with 120 time-steps as input, and 24 time-steps as output, and was 0.91 degrees. There was one case where the MAE for the CNN was higher than the LSTM, which used the smallest combination of input and output windows (six and one time-steps, respectively). For this case, the difference in MAE between the CNN and LSTM was small, 0.014 degrees, and the CNN had a higher standard deviation. Our results are different from those of Moreira et al. [40], who found that the CNN was more robust for ankle joint torque estimation based on kinematics, speed, and anthropometry. Our results are also different from those of Molinaro et al. [41], in which a temporal convolution network (with dilated convolution layers) outperformed an LSTM implementation. It must be stressed that Molinaro et al. considered joint moments rather than joint angles.
We also compared the influence of the size of the input and output windows on predictions. The size of the input window did not have a significant influence on the accuracy of the LSTM network when predicting small output window sizes (including one, three, and six time-steps). However, for predicting longer output window sizes (including 12 and 24 time-steps), larger input windows resulted in lower errors. For both 12 and 24 time-steps, the lowest error was achieved using 120 input time-steps. This is different from what was reported by [25], who found that after increasing input size beyond 30 time-steps, the mean errors for predicting five time-steps in the future increased (10 time-steps correspond to 60 ms in that paper).
There were cases in our results in which the difference between the actual and predicted trajectories was large compared to the mean absolute error. We could not tell whether this was due to the model’s lack of generalisability for certain types of pathological gait patterns, or due to an underlying issue with the data for those samples, such as sensor or marker errors. This is because the subjects were anonymised and the dataset used didn ot contain supplementary information/videos for each trial. This is one of the limitations to our study. Another limitation to this study, also caused by the anonymisation of the dataset, was the inability to test whether there is a significant difference in performance between individualised models (models that are subject-specific and need to be trained on data from the user of the exoskeleton) and generalised models (models that are subject-independent and make predictions without the need to be trained on data from a specific user). This is an important point to consider when designing exoskeleton control strategies.

5. Conclusions

To conclude, we have used two deep learning models, LSTM and CNN, to forecast the trajectories of children with neurological disorders. The results show that our deep learning models outperform the three baseline methods we implemented in our study (with the LSTM being the top performer), with only two exceptions in which the FCN was better (see Section 3.5 for details). We also experimented with varying input and output windows to quantify how the performance is affected by the amount of input data and the length of the future horizon. A potential application of our approach is the control of lower limb robotics, whereby forecasted trajectories of the models can be used as a proxy for the user’s intentions. These intentions can be integrated into the control hierarchy of exoskeletons, specifically into the high-level control responsible for detecting the user’s intention and passing it on to the mid and low levels to generate appropriate movement commands. For real-time systems, there is always a trade-off between performance and speed. Input windows therefore need to be large enough to achieve acceptable errors, but not too large to slow the system down. As future work, we should evaluate the performance of the models on data collected using wearable sensors (e.g., IMUs and foot pressure sensors) rather than motion capture systems. We should also evaluate the difference between the performance of individualised and generalised models. Furthermore, we believe that the forecasted trajectories need to be paired with a corrective algorithm, unique to every gait sequence. In this scenario, the user intention (coming from a trajectory forecasting model) is adjusted by a corrective algorithm that produces the ’desired trajectory’ used by the mid- and low-level controllers of the exoskeleton.

Author Contributions

Conceptualisation, R.K., M.K.A.-H., K.S.; methodology, R.K., M.K.A.-H., K.S.; software, R.K., M.K.A.-H.; validation, R.K., M.K.A.-H., G.M., K.S.; formal analysis, R.K.; investigation, R.K.; resources, K.S.; data curation, R.K.; writing—original draft preparation, R.K.; writing—review and editing, R.K., M.K.A.-H., G.M., K.S.; visualisation, R.K.; supervision, M.K.A.-H., G.M., K.S.; project administration, M.K.A.-H., K.S.; funding acquisition, K.S. All authors have read and agreed to the published version of the manuscript.

Funding

This study is part of the MOTION project, funded by the Interreg 2 Seas programme 2014-2020 and, co-funded by the European Regional Development Fund under subsidy contract 2S05-038. (project links: www.motion-interreg.eu and www.interreg2seas.eu/en/MOTION) (accessed on 1 October 2021).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The dataset used in this study has been collected by Gillette Speciality Healthcare, and publicly shared at https://simtk.org/frs/?group_id=1946 (accessed on 1 October 2021).

Acknowledgments

We would like to thank Marco Santopietro for his technical suggestions.

Conflicts of Interest

The authors declare no conflict of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript, or in the decision to publish the results.

Abbreviations

The following abbreviations are used in this manuscript:
LSTMLong Short-Term Memory
CNNConvolutional Neural Network
FCNFully Connected Network
MAEMean Absolute Error
MSEMean Squared Error

References

  1. Webster, J.B. Introduction. In Atlas of Orthoses and Assistive Devices, 5th ed.; Webster, J.B., Murphy, D.P., Eds.; Elsevier: Philadelphia, PA, USA, 2019; p. 376. [Google Scholar] [CrossRef]
  2. Makinson, B.J. Research and Development Prototype for Machine Augmentation of Human Strength and Endurance Hardiman I Project; General Electric: Schenectady, NY, USA, 1971. [Google Scholar]
  3. Kumar, V.; Hote, Y.V.; Jain, S. Review of Exoskeleton: History, Design and Control. In Proceedings of the 2019 3rd International Conference on Recent Developments in Control, Automation Power Engineering (RDCAPE), Noida, India, 10–11 October 2019; pp. 677–682. [Google Scholar] [CrossRef]
  4. Bosch, T.; van Eck, J.; Knitel, K.; de Looze, M. The effects of a passive exoskeleton on muscle activity, discomfort and endurance time in forward bending work. Appl. Ergon. 2016, 54, 212–217. [Google Scholar] [CrossRef]
  5. Bai, S.; Virk, G.S.; Sugar, T.G. (Eds.) Wearable Exoskeleton Systems: Design, Control and Applications; Control, Robotics & Sensors; Institution of Engineering and Technology: London, UK, 2018. [Google Scholar]
  6. Chen, B.; Ma, H.; Qin, L.Y.; Gao, F.; Chan, K.M.; Law, S.W.; Qin, L.; Liao, W.H. Recent developments and challenges of lower extremity exoskeletons. J. Orthop. Transl. 2016, 5, 26–37. [Google Scholar] [CrossRef] [Green Version]
  7. Riener, R.; Lunenburger, L.; Jezernik, S.; Anderschitz, M.; Colombo, G.; Dietz, V. Patient-cooperative strategies for robot-aided treadmill training: First experimental results. IEEE Trans. Neural Syst. Rehabil. Eng. 2005, 13, 380–394. [Google Scholar] [CrossRef] [PubMed]
  8. Høyer, E.; Opheim, A.; Jørgensen, V. Implementing the exoskeleton Ekso GTTM for gait rehabilitation in a stroke unit—Feasibility, functional benefits and patient experiences. Disabil. Rehabil. Assist. Technol. 2020, 1–7. [Google Scholar] [CrossRef] [PubMed]
  9. Winfree, K.N.; Stegall, P.; Agrawal, S.K. Design of a minimally constraining, passively supported gait training exoskeleton: ALEX II. In Proceedings of the 2011 IEEE International Conference on Rehabilitation Robotics, Zurich, Switzerland, 29 June–1 July 2011; pp. 1–6. [Google Scholar] [CrossRef]
  10. Gorgey, A.S. Robotic exoskeletons: The current pros and cons. World J. Orthop. 2018, 9, 112–119. [Google Scholar] [CrossRef] [PubMed]
  11. Quintero, H.A.; Farris, R.J.; Goldfarb, M. A Method for the Autonomous Control of Lower Limb Exoskeletons for Persons with Paraplegia. J. Med. Devices Trans. ASME 2012, 6, 041003. [Google Scholar] [CrossRef]
  12. Zoss, A.; Kazerooni, H.; Chu, A. Biomechanical design of the Berkeley lower extremity exoskeleton (BLEEX). IEEE/ASME Trans. Mechatron. 2006, 11, 128–138. [Google Scholar] [CrossRef]
  13. Sankai, Y. HAL: Hybrid Assistive Limb Based on Cybernics. In Robotics Research; Kaneko, M., Nakamura, Y., Eds.; Springer: Berlin/Heidelberg, Germany, 2011; pp. 25–34. [Google Scholar]
  14. Baud, R.; Manzoori, A.R.; Ijspeert, A.; Bouri, M. Review of control strategies for lower-limb exoskeletons to assist gait. J. NeuroEng. Rehabil. 2021, 18, 119. [Google Scholar] [CrossRef]
  15. Al-Shuka, H.F.N.; Song, R.; Ding, C. On high-level control of power-augmentation lower extremity exoskeletons: Human walking intention. In Proceedings of the 2018 Tenth International Conference on Advanced Computational Intelligence (ICACI), Xiamen, China, 29–31 March 2018; pp. 169–174. [Google Scholar] [CrossRef]
  16. Kolaghassi, R.; Al-Hares, M.K.; Sirlantzis, K. Systematic Review of Intelligent Algorithms in Gait Analysis and Prediction for Lower Limb Robotic Systems. IEEE Access 2021, 9, 113788–113812. [Google Scholar] [CrossRef]
  17. Xiong, B.; Zeng, N.; Li, H.; Yang, Y.; Li, Y.; Huang, M.; Shi, W.; Du, M.; Zhang, Y. Intelligent Prediction of Human Lower Extremity Joint Moment: An Artificial Neural Network Approach. IEEE Access 2019, 7, 29973–29980. [Google Scholar] [CrossRef]
  18. Song, J.; Zhu, A.; Tu, Y.; Wang, Y.; Arif, M.A.; Shen, H.; Shen, Z.; Zhang, X.; Cao, G. Human Body Mixed Motion Pattern Recognition Method Based on Multi-Source Feature Parameter Fusion. Sensors 2020, 20, 537. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  19. Laschowski, B.; McNally, W.; Wong, A.; McPhee, J. Environment Classification for Robotic Leg Prostheses and Exoskeletons Using Deep Convolutional Neural Networks. Front. Neurorobot. 2022, 15, 730965. [Google Scholar] [CrossRef] [PubMed]
  20. Al-Shuka, H.F.N.; Song, R. On low-level control strategies of lower extremity exoskeletons with power augmentation. In Proceedings of the 2018 Tenth International Conference on Advanced Computational Intelligence (ICACI), Xiamen, China, 29–31 March 2018; pp. 63–68. [Google Scholar] [CrossRef]
  21. Vantilt, J.; Tanghe, K.; Afschrift, M.; Bruijnes, A.K.; Junius, K.; Geeroms, J.; Aertbeliën, E.; De Groote, F.; Lefeber, D.; Jonkers, I.; et al. Model-based control for exoskeletons with series elastic actuators evaluated on sit-to-stand movements. J. NeuroEng. Rehabil. 2019, 16, 65. [Google Scholar] [CrossRef] [PubMed]
  22. Vallery, H.; Van Asseldonk, E.H.; Buss, M.; Van Der Kooij, H. Reference trajectory generation for rehabilitation robots: Complementary limb motion estimation. IEEE Trans. Neural Syst. Rehabil. Eng. 2009, 17, 23–30. [Google Scholar] [CrossRef] [Green Version]
  23. Tanghe, K.; De Groote, F.; Lefeber, D.; De Schutter, J.; Aertbeliën, E. Gait Trajectory and Event Prediction from State Estimation for Exoskeletons During Gait. IEEE Trans. Neural Syst. Rehabil. Eng. 2020, 28, 211–220. [Google Scholar] [CrossRef]
  24. Liu, D.X.; Wu, X.; Wang, C.; Chen, C. Gait trajectory prediction for lower-limb exoskeleton based on Deep Spatial-Temporal Model (DSTM). In Proceedings of the 2017 2nd International Conference on Advanced Robotics and Mechatronics (ICARM), Hefei and Tai’an, China, 27–31 August 2017; pp. 564–569. [Google Scholar] [CrossRef]
  25. Zaroug, A.; Lai, D.T.H.; Mudie, K.; Begg, R. Lower Limb Kinematics Trajectory Prediction Using Long Short-Term Memory Neural Networks. Front. Bioeng. Biotechnol. 2020, 8, 362. [Google Scholar] [CrossRef]
  26. Su, B.; Gutierrez-Farewik, E.M. Gait Trajectory and Gait Phase Prediction Based on an LSTM Network. Sensors 2020, 20, 7127. [Google Scholar] [CrossRef]
  27. Hernandez, V.; Dadkhah, D.; Babakeshizadeh, V.; Kulić, D. Lower body kinematics estimation from wearable sensors for walking and running: A deep learning approach. Gait Posture 2021, 83, 185–193. [Google Scholar] [CrossRef]
  28. Jia, L.; Ai, Q.; Meng, W.; Liu, Q.; Xie, S.Q. Individualized Gait Trajectory Prediction Based on Fusion LSTM Networks for Robotic Rehabilitation Training. In Proceedings of the 2021 IEEE/ASME International Conference on Advanced Intelligent Mechatronics (AIM), Delft, The Netherlands, 12–16 July 2021; pp. 988–993. [Google Scholar] [CrossRef]
  29. Zaroug, A.; Garofolini, A.; Lai, D.T.H.; Mudie, K.; Begg, R. Prediction of gait trajectories based on the Long Short Term Memory neural networks. PLoS ONE 2021, 16, e0255597. [Google Scholar] [CrossRef]
  30. Zhu, C.; Liu, Q.; Meng, W.; Ai, Q.; Xie, S.Q. An Attention-Based CNN-LSTM Model with Limb Synergy for Joint Angles Prediction. In Proceedings of the 2021 IEEE/ASME International Conference on Advanced Intelligent Mechatronics (AIM), Delft, The Netherlands, 12–16 July 2021; pp. 747–752. [Google Scholar] [CrossRef]
  31. Kirtley, C. Clinical Gait Analysis: Theory and Practice; Churchill Livingstone Elsevier: Edinburgh, UK, 2006. [Google Scholar]
  32. Steinwender, G.; Saraph, V.; Scheiber, S.; Zwick, E.B.; Uitz, C.; Hackl, K. Intrasubject repeatability of gait analysis data in normal and spastic children. Clin. Biomech. 2000, 15, 134–139. [Google Scholar] [CrossRef]
  33. Moon, Y.; Sung, J.; An, R.; Hernandez, M.E.; Sosnoff, J.J. Gait variability in people with neurological disorders: A systematic review and meta-analysis. Hum. Mov. Sci. 2016, 47, 197–208. [Google Scholar] [CrossRef] [PubMed]
  34. Automatic Real-Time Gait Event Detection in Children Using Deep Neural Networks. Available online: https://simtk.org/frs/?group_id=1946 (accessed on 1 October 2021).
  35. Kidzinski, L.; Delp, S.; Schwartz, M. Automatic real-time gait event detection in children using deep neural networks. PLoS ONE 2019, 14, e0211466. [Google Scholar] [CrossRef] [Green Version]
  36. Goodfellow, I.; Bengio, Y.; Courville, A. Deep Learning; MIT Press: Cambridge, MA, USA, 2016; Available online: http://www.deeplearningbook.org (accessed on 2 February 2022).
  37. Akiba, T.; Sano, S.; Yanase, T.; Ohta, T.; Koyama, M. Optuna: A Next-generation Hyperparameter Optimization Framework. In Proceedings of the 25rd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, Anchorage, AK, USA, 4–8 August 2019. [Google Scholar]
  38. Gieysztor, E.; Kowal, M.; Paprocka-Borowicz, M. Gait Parameters in Healthy Preschool and School Children Assessed Using Wireless Inertial Sensor. Sensors 2021, 21, 6423. [Google Scholar] [CrossRef] [PubMed]
  39. Kiranyaz, S.; Avci, O.; Abdeljaber, O.; Ince, T.; Gabbouj, M.; Inman, D.J. 1D convolutional neural networks and applications: A survey. Mech. Syst. Signal Process. 2021, 151, 107398. [Google Scholar] [CrossRef]
  40. Moreira, L.; Figueiredo, J.; Vilas-Boas, J.P.; Santos, C.P. Kinematics, Speed, and Anthropometry-Based Ankle Joint Torque Estimation: A Deep Learning Regression Approach. Machines 2021, 9, 154. [Google Scholar] [CrossRef]
  41. Molinaro, D.D.; Kang, I.; Camargo, J.; Gombolay, M.C.; Young, A.J. Subject-Independent, Biological Hip Moment Estimation During Multimodal Overground Ambulation Using Deep Learning. IEEE Trans. Med Robot. Bionics 2022, 4, 219–229. [Google Scholar] [CrossRef]
Figure 1. Illustration of the sliding window method. Continuous gait trajectories are used to generate input and output windows for training the model using the sliding window method. Each pair of input and output windows forms one sample. Several samples can be generated from one continuous gait trajectory by sliding the window for a specified distance, also known as stride length.
Figure 1. Illustration of the sliding window method. Continuous gait trajectories are used to generate input and output windows for training the model using the sliding window method. Each pair of input and output windows forms one sample. Several samples can be generated from one continuous gait trajectory by sliding the window for a specified distance, also known as stride length.
Sensors 22 02969 g001
Figure 2. Architecture of the LSTM network used in this study.
Figure 2. Architecture of the LSTM network used in this study.
Sensors 22 02969 g002
Figure 3. Architecture of the convolutional neural network used in this study.
Figure 3. Architecture of the convolutional neural network used in this study.
Sensors 22 02969 g003
Figure 4. Performance of LSTM, measured by MAE, for varying input and output window sizes. Input sizes range from 6 to 120 time-steps, corresponding to 200 to 1000 ms. Output sizes range from 1 to 24 time-steps, corresponding to 8.33 to 200 ms. Sub-figures (ae) correspond to networks with 1, 3, 6, 12, and 24 output time-steps, respectively.
Figure 4. Performance of LSTM, measured by MAE, for varying input and output window sizes. Input sizes range from 6 to 120 time-steps, corresponding to 200 to 1000 ms. Output sizes range from 1 to 24 time-steps, corresponding to 8.33 to 200 ms. Sub-figures (ae) correspond to networks with 1, 3, 6, 12, and 24 output time-steps, respectively.
Sensors 22 02969 g004
Figure 5. Comparison of the performance of the CNN and LSTM network, measured by MAE, for varying output window sizes. In (a), the input size is fixed at 6 time-steps, while in (b) the input is fixed at 120 time-steps.
Figure 5. Comparison of the performance of the CNN and LSTM network, measured by MAE, for varying output window sizes. In (a), the input size is fixed at 6 time-steps, while in (b) the input is fixed at 120 time-steps.
Sensors 22 02969 g005
Figure 6. Mean absolute errors for each individual time-step predicted by the LSTM and CNN for a given output window. Input window size is fixed at 120 time-steps. Sub-figures (ad) correspond to networks with 3, 6, 12, and 24 output time-steps, respectively.
Figure 6. Mean absolute errors for each individual time-step predicted by the LSTM and CNN for a given output window. Input window size is fixed at 120 time-steps. Sub-figures (ad) correspond to networks with 3, 6, 12, and 24 output time-steps, respectively.
Sensors 22 02969 g006
Figure 7. Mean absolute errors for each individual time-step predicted by the LSTM networks with 3, 6, 12, and 24 output window sizes. Input window size is fixed at 120 time-steps.
Figure 7. Mean absolute errors for each individual time-step predicted by the LSTM networks with 3, 6, 12, and 24 output window sizes. Input window size is fixed at 120 time-steps.
Sensors 22 02969 g007
Figure 8. MAEs for each of the hip, knee, and ankle joints for the CNN and LSTM network with varying output sizes. Input window size is fixed at 120 time-steps and the MAE for each joint represents the combined MAE for the yaw, pitch, and roll dimensions. Sub-figures (ae) correspond to networks with 1, 3, 6, 12, and 24 output time-steps, respectively.
Figure 8. MAEs for each of the hip, knee, and ankle joints for the CNN and LSTM network with varying output sizes. Input window size is fixed at 120 time-steps and the MAE for each joint represents the combined MAE for the yaw, pitch, and roll dimensions. Sub-figures (ae) correspond to networks with 1, 3, 6, 12, and 24 output time-steps, respectively.
Sensors 22 02969 g008
Table 1. Hyper-parameter optimisation for the LSTM network, CNN, and FCN.
Table 1. Hyper-parameter optimisation for the LSTM network, CNN, and FCN.
Hyper-ParameterSearch
Space
Selected
Value
LSTMlearning rate[0.1, 0.01, 0.001, 0.0001, 0.00001]0.001
number of LSTM layers[1, 2, 3, 4]4
number of LSTM hidden units[16, 32, 64, 100, 128]128
CNNlearning rate[0.1, 0.01, 0.001, 0.0001, 0.00001]0.0001
conv1D layer 1 channels[16, 32, 48]32
conv1D layer 2 channels[32, 48, 64]48
conv1D layer 3 channels[64, 128, 256]256
conv1D layer 4 channels[128, 256, 512]512
kernel size for layers 1, 2[1, 2, 3, 4, 5, 6, 7]7
kernel size for layers 3, 4[1, 2, 3, 4, 5, 6, 7]7
padding[0, 1, 2, 3, 4, 5]4
conv1D stride[1, 2, 3, 4, 5]1
dilation[1, 2, 4]1
FCNlearning rate[0.1, 0.01, 0.001, 0.0001, 0.00001]0.001
hidden layers[1, 2, 3, 4, 6, 8, 10]3
nodes per layer[10, 20, 40, 60, 100, 140, 160, 200]200
Table 2. Performance of LSTM in forecasting gait trajectories for varying input and output window sizes.
Table 2. Performance of LSTM in forecasting gait trajectories for varying input and output window sizes.
Input
Window
Size (ms)
Input
Time-
Steps
Output
Window
Size (ms)
Output
Time-
Steps
MSE
(Degrees)
MSE
std
(Degrees)
MAE
(Degrees)
MAE
std
(Degrees)
Mean
Pearson
Correlation
Coefficient
5068.3310.0340.0650.1430.1151.000
100128.3310.0770.1300.2140.1771.000
200248.3310.0270.0550.1260.1051.000
400488.3310.0190.1610.0950.0991.000
600728.3310.0300.2660.1260.1191.000
800968.3310.0200.1250.1070.0921.000
10001208.3310.0220.3180.1090.0981.000
5062530.0790.5260.1750.2201.000
100122530.0770.4740.1870.2061.000
200242530.0790.7930.1760.2181.000
400482530.0803.0680.1690.2271.000
600722530.0922.5970.1730.2501.000
800962530.1041.2790.2000.2521.000
10001202530.1172.0280.2230.2611.000
5065060.6144.1150.4610.6330.998
100125060.4223.9560.3650.5370.998
200245060.4164.4160.3500.5410.998
400485060.3812.7730.3560.5050.998
600725060.4267.6530.3520.5500.998
800965060.3633.3770.3320.5020.998
10001205060.4055.4140.3590.5260.998
506100123.54815.7491.1041.5260.984
10012100123.31015.5451.0581.4800.985
20024100123.20014.7901.0081.4780.985
40048100123.27930.5671.0071.5050.985
60072100122.72314.9930.9271.3650.987
80096100122.52413.3660.9061.3050.988
1000120100122.1579.9400.8471.2000.990
5062002416.98157.0842.5313.2520.928
100122002415.02549.9372.3833.0570.935
200242002415.11455.8362.3573.0910.934
400482002414.05850.3332.2822.9750.937
600722002412.61748.4342.1582.8210.945
800962002410.38938.3721.9732.5490.955
1000120200248.97136.8621.8282.3730.961
Bold entries denote the lowest MSE and MAE values for a given output window size.
Table 3. Performance of CNN in forecasting gait trajectories for varying input and output window sizes.
Table 3. Performance of CNN in forecasting gait trajectories for varying input and output window sizes.
Input
Window
Size (ms)
Input
Time-
Steps
Output
Window
Size (ms)
Output
Time-
Steps
MSE
(Degrees)
MSE
std
(Degrees)
MAE
(Degrees)
MAE
std
(Degrees)
Mean
Pearson
Correlation
Coefficient
5068.3310.0690.9600.1290.2291.000
5062530.1841.6990.2340.3600.999
5065060.8914.6850.5520.7660.996
506100125.26520.2771.3581.8500.977
5062002420.43765.5962.8403.5170.913
10001208.3310.0610.7090.1380.2031.000
10001202530.2162.0200.2590.3860.999
10001205060.9944.9660.5760.8140.995
1000120100125.49618.7241.4401.8500.975
10001202002418.00754.4532.7383.2420.926
Table 4. Benchmarking performance of deep learning models.
Table 4. Benchmarking performance of deep learning models.
Input
Window
Size (ms)
Input
Time-
Steps
Output
Window
Size (ms)
Output
Time-
Steps
LSTM
MAE
(Degrees)
CNN
MAE
(Degrees)
FCN
MAE
(Degrees)
Naïve
Method 1 a
MAE
(Degrees)
Naïve
Method 2 b
MAE
(Degrees)
5068.3310.143 *0.129 * , 0.195 * , 0.449 1.513 * ,
5062530.175 *0.234 * , 0.294 * , 0.888 1.916 * ,
5065060.461 *0.552 * , 0.568 * , 1.517 2.486 * ,
506100121.104 *1.358 * , 1.336 * , 2.640 3.494 * ,
506200242.531 *2.840 * , 2.489 * , 4.417 5.096 * ,
10001208.3310.109 *0.138 * , 0.369 * , 0.448 6.090 * ,
10001202530.223 *0.259 * , 0.594 * , 0.888 6.121 * ,
10001205060.359 *0.576 * , 0.902 * , 1.520 6.167 * ,
1000120100120.847 *1.440 * , 1.454 * , 2.651 6.254 * ,
1000120200241.828 *2.738 * , 2.320 * , 4.448 6.385 * ,
a Naïve Method 1: all output time-steps are predicted to be the value of the last input time-step. b Naïve Method 2: all output time-steps are predicted to be the mean value of the input time-steps. * and represent statistical significance compared to Naïve Method 1 and LSTM, respectively. Significance is based on pairwise t-tests (p < 0.05). Bold entries denote the lowest MAE value for a given input and output window size.
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Kolaghassi, R.; Al-Hares, M.K.; Marcelli, G.; Sirlantzis, K. Performance of Deep Learning Models in Forecasting Gait Trajectories of Children with Neurological Disorders. Sensors 2022, 22, 2969. https://doi.org/10.3390/s22082969

AMA Style

Kolaghassi R, Al-Hares MK, Marcelli G, Sirlantzis K. Performance of Deep Learning Models in Forecasting Gait Trajectories of Children with Neurological Disorders. Sensors. 2022; 22(8):2969. https://doi.org/10.3390/s22082969

Chicago/Turabian Style

Kolaghassi, Rania, Mohamad Kenan Al-Hares, Gianluca Marcelli, and Konstantinos Sirlantzis. 2022. "Performance of Deep Learning Models in Forecasting Gait Trajectories of Children with Neurological Disorders" Sensors 22, no. 8: 2969. https://doi.org/10.3390/s22082969

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop