Author Contributions
Conceptualization, A.N.A., G.E., K.S.v.S., M.P. and B.K.; Methodology, A.N.A., G.E., K.S.v.S., M.P. and B.K.; Software, A.N.A. and K.S.v.S.; Validation, A.N.A.; Formal Analysis, A.N.A., G.E. and B.K.; Investigation, A.N.A., G.E., K.S.v.S. and M.P.; Resources, K.S.v.S., M.P. and B.K.; Data Curation, K.S.v.S. and M.P.; Writing-Original Draft Preparation, A.N.A., G.E., K.S.v.S., M.P. and B.K.; Writing-Review & Editing, A.N.A., G.E., K.S.v.S., M.P. and B.K.; Visualization, A.N.A. and G.E.; Supervision, G.E., K.S.v.S., M.P. and B.K.; Project Administration, A.N.A., K.S.v.S., M.P. and B.K.; Funding Acquisition, K.S.v.S., M.P. and B.K.
Figure 1.
An example of two locomotion samples. (a) A typical walking sample; (b) A walking sample interwoven with a turning activity at 7 s. accAP: acceleration in the anteroposterior direction; accML: acceleration in the mediolateral direction; accVT: acceleration in the vertical direction.
Figure 1.
An example of two locomotion samples. (a) A typical walking sample; (b) A walking sample interwoven with a turning activity at 7 s. accAP: acceleration in the anteroposterior direction; accML: acceleration in the mediolateral direction; accVT: acceleration in the vertical direction.
Figure 2.
A three-layer neural network with three input neurons, two hidden layers of four neurons each, and one output layer.
Figure 2.
A three-layer neural network with three input neurons, two hidden layers of four neurons each, and one output layer.
Figure 3.
(a) A cyclic connection of a recurrent neural network (RNN) folded and unfolded. (b) An long short-term memory (LSTM) memory block consisting of one cell at time t and the three gates (it, ot, and ft) which control the activation of the cell ct and its output ht.
Figure 3.
(a) A cyclic connection of a recurrent neural network (RNN) folded and unfolded. (b) An long short-term memory (LSTM) memory block consisting of one cell at time t and the three gates (it, ot, and ft) which control the activation of the cell ct and its output ht.
Figure 4.
Example boxplots of the normalized predicted values (
) for multiple 10-s sequences, grouped by subject. Subjects 1 and 4 were non-fallers and the other two were fallers. The final prediction per subject was given by the median of the predictions, as per van Schooten et al. [
21]. The green line inside the box represents the median, the box represents the range of first and third quartile and circles represent outliers.
Figure 4.
Example boxplots of the normalized predicted values (
) for multiple 10-s sequences, grouped by subject. Subjects 1 and 4 were non-fallers and the other two were fallers. The final prediction per subject was given by the median of the predictions, as per van Schooten et al. [
21]. The green line inside the box represents the median, the box represents the range of first and third quartile and circles represent outliers.
Figure 5.
A typical loss versus epoch graph during the training of a deep neural network (DNN). The data has been split at (a) subject level or (b) sample level. Loss is the training loss and val_loss is the validation loss. The gap between the training and validation loss indicates the amount of over-fitting.
Figure 5.
A typical loss versus epoch graph during the training of a deep neural network (DNN). The data has been split at (a) subject level or (b) sample level. Loss is the training loss and val_loss is the validation loss. The gap between the training and validation loss indicates the amount of over-fitting.
Figure 6.
Examples of the receiver operating characteristic (ROC) curves (solid red lines) and their corresponding area under the curve (AUC) values obtained using a ConvLSTM model. The dashed blue line represents the ROC for chance. The dataset was split at (a) the subject level and (b) the sample level.
Figure 6.
Examples of the receiver operating characteristic (ROC) curves (solid red lines) and their corresponding area under the curve (AUC) values obtained using a ConvLSTM model. The dashed blue line represents the ROC for chance. The dataset was split at (a) the subject level and (b) the sample level.
Figure 7.
A boxplot of the AUCs of different DNN architectures. For the LSTM architecture, at least two LSTM layers were involved, while for the ConvLSTM architecture, only one LSTM layer was involved. The dataset was split at (a) the subject level and (b) the sample level.
Figure 7.
A boxplot of the AUCs of different DNN architectures. For the LSTM architecture, at least two LSTM layers were involved, while for the ConvLSTM architecture, only one LSTM layer was involved. The dataset was split at (a) the subject level and (b) the sample level.
Figure 8.
A sample of obtained ROCs for multi-task deep learning (MTDL) with fall status as the main task and subject identity as the auxiliary task. For the auxiliary task, the ROCs were computed using one-versus-all. The corresponding average AUC is reported. For (a,b), the main and auxiliary losses were given the same weight (1:1); for (c,d), the main loss function was given higher weight (104:1) than the auxiliary loss function. The dashed blue lines in (a,c) represent the chance ROC.
Figure 8.
A sample of obtained ROCs for multi-task deep learning (MTDL) with fall status as the main task and subject identity as the auxiliary task. For the auxiliary task, the ROCs were computed using one-versus-all. The corresponding average AUC is reported. For (a,b), the main and auxiliary losses were given the same weight (1:1); for (c,d), the main loss function was given higher weight (104:1) than the auxiliary loss function. The dashed blue lines in (a,c) represent the chance ROC.
Figure 9.
(a) An example of a 10-s data sample included in the training and testing set and (b) an example of a data sample excluded in Experiment 5 due to the low dominant frequencies in the VT-axis. In the bottom-right corners, histograms of the VT-frequencies up to 3 Hz are depicted. Both examples were included in the first 4 experiments.
Figure 9.
(a) An example of a 10-s data sample included in the training and testing set and (b) an example of a data sample excluded in Experiment 5 due to the low dominant frequencies in the VT-axis. In the bottom-right corners, histograms of the VT-frequencies up to 3 Hz are depicted. Both examples were included in the first 4 experiments.
Table 1.
Descriptive statistics of the population.
Table 1.
Descriptive statistics of the population.
| Male (%) | Age (Years) | Weight (kg) | Height (cm) |
---|
Mean | 74.1 | 75.3 | 49.2 | 170.6 |
Standard deviation | - | 6.8 | 13.3 | 8.8 |
25% Quantile | - | 70.0 | 64.0 | 165.0 |
75% Quantile | - | 80.0 | 81.8 | 176.0 |
Table 2.
Average AUC (standard deviation) and corresponding average training time per neural network (NN) architecture type for a subset of the data. The difference in training time between the ways of splitting the data is due to the slower convergence when splitting at the sample level.
Table 2.
Average AUC (standard deviation) and corresponding average training time per neural network (NN) architecture type for a subset of the data. The difference in training time between the ways of splitting the data is due to the slower convergence when splitting at the sample level.
| Subject Level | Sample Level |
---|
| AUC | Time (h) | AUC | Time (h) |
---|
CNN | 0.52 (0.07) | 6 | 0.74 (0.07) | 7 |
LSTM | 0.61 (0.10) | 160 | 0.91 (0.06) | 180 |
ConvLSTM | 0.60 (0.09) | 35 | 0.90 (0.05) | 40 |
Table 3.
The ConvLSTM architecture (ConvLSTM is our proposed model that combines convolutional and recurrent models). To keep the architecture clear, we omitted the input layer (layer 00) and the dropout layers (the even layer indices) applied after each convolutional neural network (CNN) layer. N was set to 128.
Table 3.
The ConvLSTM architecture (ConvLSTM is our proposed model that combines convolutional and recurrent models). To keep the architecture clear, we omitted the input layer (layer 00) and the dropout layers (the even layer indices) applied after each convolutional neural network (CNN) layer. N was set to 128.
Layer Index | 01 | 03 | 05 | 07 | 09 | 11 | 12 |
---|
type of filter | CNN | CNN | CNN | CNN | CNN | LSTM | Dense |
number of filters | N | N | N | | | N | 2 |
Table 4.
Average AUC, training duration, and number of folds obtained when applying the ConvLSTM model to different dataset sizes. The dataset was cut into three subsets at the subject level.
Table 4.
Average AUC, training duration, and number of folds obtained when applying the ConvLSTM model to different dataset sizes. The dataset was cut into three subsets at the subject level.
| Dataset Size in Minutes |
---|
| 10 | 30 | 60 | 120 | Complete Dataset |
---|
Average AUC | 0.61 | 0.63 | 0.65 | 0.65 | 0.65 |
Training duration (h) | 35 | 90 | 150 | 250 | 350 |
Number of folds | 10 | 10 | 10 | 2 | 1 |
Table 5.
Average AUC and the corresponding standard deviation when splitting at sample or subject levels.
Table 5.
Average AUC and the corresponding standard deviation when splitting at sample or subject levels.
| AUC |
---|
| Average | Standard Deviation |
---|
subject level | 0.65 | 0.09 |
sample level | 0.94 | 0.07 |
Table 6.
Average AUC and the corresponding standard deviation of the main task (fall status), obtained when the ConvLSTM is applied to the test set. The p-value was obtained using the z-test to test the difference in the performance to the base model.
Table 6.
Average AUC and the corresponding standard deviation of the main task (fall status), obtained when the ConvLSTM is applied to the test set. The p-value was obtained using the z-test to test the difference in the performance to the base model.
Characteristic | AUC Main Task (std dev) | p-Value Diff to Base Model |
---|
Experiment 4 | Experiment 5 | Experiment 4 | Experiment 5 |
---|
Gender | 0.70 (0.06) | 0.75 (0.05) | 0.070 | <0.001 |
Age | 0.70 (0.05) | 0.74 (0.05) | 0.082 | <0.001 |
Weight | 0.68 (0.05) | 0.72 (0.05) | 0.306 | 0.005 |
Height | 0.63 (0.06) | 0.65 (0.06) | 0.987 | 0.897 |